Abstract
BEAM enables dynamic expert selection in Mixture-of-Experts models through trainable binary masks, achieving significant computational savings while maintaining high performance.
Mixture-of-Experts (MoE) architectures enhance the efficiency of large language models by activating only a subset of experts per token. However, standard MoE employs a fixed Top-K routing strategy, leading to redundant computation and suboptimal inference latency. Existing acceleration methods either require costly retraining with architectural changes or suffer from severe performance drop at high sparsity due to train-inference mismatch. To address these limitations, we propose BEAM (Binary Expert Activation Masking), a novel method that learns token-adaptive expert selection via trainable binary masks. With a straight-through estimator and an auxiliary regularization loss, BEAM induces dynamic expert sparsity through end-to-end training while maintaining model capability. We further implement an efficient custom CUDA kernel for BEAM, ensuring seamless integration with the vLLM inference framework. Experiments show that BEAM retains over 98\% of the original model's performance while reducing MoE layer FLOPs by up to 85\%, achieving up to 2.5times faster decoding and 1.4times higher throughput, demonstrating its effectiveness as a practical, plug-and-play solution for efficient MoE inference.
Community
Interesting work on practical MoE inference acceleration. BEAM introduces trainable binary expert masks for token-adaptive routing, which helps avoid the redundant computation of fixed Top-K routing while staying more deployment-friendly than methods that require major architectural changes or expensive retraining. The results are also amazing: over 98% performance retention with up to 85% MoE FLOPs reduction, 2.5× faster decoding, and 1.4× higher throughput. Overall, this is a simple and practical plug-and-play solution for making MoE models more efficient in real-world inference.
Get this paper in your agent:
hf papers read 2605.14438 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper