LISTEN TO THIS ARTICLE
MoE's Dirty Secret Is Load Balancing
Every frontier lab now ships a sparse Mixture-of-Experts model. Google's Switch Transformer started the trend. DeepSeek-V3 proved it could scale. Mistral's Mixtral made it accessible. But here's the number that should bother you: in a typical 8-expert MoE layer, two or three experts handle over 60% of all tokens while the rest sit nearly idle. You're paying for eight experts and getting maybe three.
That's the dirty secret at the heart of the MoE efficiency story. The architecture promises to decouple total parameter count from per-token compute, letting you build massive models that only activate a fraction of their weights on each forward pass. In theory, you get the knowledge capacity of a dense giant with the inference cost of something much smaller. In practice, the routing mechanisms that decide which expert handles which token are broken in ways that compound as you scale. And the fixes being proposed in early 2026 tell us a lot about where MoE scaling laws actually hit their limits.
The Efficiency Promise vs. the Routing Reality
MoE's pitch is elegant: instead of forcing every parameter to process every token, you train a gating network to route each token to the top-k experts best suited for it. A model with 400 billion total parameters might only activate 50 billion per token. Training costs scale with total parameters, but inference costs scale with the active subset. That's the theoretical efficiency frontier everyone's chasing.
The problem is that gating networks develop preferences. Some experts get really good at common patterns early in training, so the router sends them more tokens, so they get even better, so the router sends them even more. It's like a restaurant where two chefs end up cooking every dish while six others stand around watching. You hired eight chefs. You're feeding eight chefs. But your kitchen's throughput is bottlenecked by two.
This isn't a minor inconvenience. Load imbalance creates GPU utilization nightmares. When one expert is slammed and another is idle, the hardware running the idle expert is burning watts doing nothing useful. At the scale frontier labs operate, that translates directly into millions of dollars in wasted compute per training run.

Replicate-and-Quantize: A Duct-Tape Fix That Works
A February 2026 paper from Liu et al. proposes what might be the most pragmatic approach I've seen to this problem. Their "Replicate-and-Quantize" strategy takes the overloaded popular experts, duplicates them across hardware, and then quantizes the underutilized experts to free up the memory budget for those replicas. It's a post-training intervention, meaning you don't need to retrain the model. Plug and play.
The logic is blunt: if expert 2 handles 3x more tokens than expert 7, give expert 2 three copies spread across devices and compress expert 7 down to 4-bit precision since it barely fires anyway. The total memory footprint stays roughly constant, but the actual throughput on real workloads improves because you're no longer waiting on a single overloaded expert to churn through its queue.
I find this approach honest in a way that a lot of MoE research isn't. It doesn't pretend to solve the routing problem. It just acknowledges that routing is broken and works around it at the systems level. That's engineering, not science. But it's the kind of engineering that actually ships.
The Optimizer Angle Nobody Expected
While systems-level fixes address deployment, Shaier's "Excitation" paper from the same month attacks the problem from the optimizer side. The core idea: standard optimizers like Adam treat all parameters equally, but in a sparse MoE, experts activate at wildly different rates across batches. The optimizer has no mechanism to distinguish a high-consensus expert that's doing real work from one that's barely firing and accumulating stale gradients.
Excitation introduces a competitive update dynamic that modulates parameter updates based on batch-level expert utilization. Highly-utilized experts get amplified updates that reinforce their specialization, while low-utilization experts get selectively suppressed. Think of it as sharpening the routing signal: instead of spreading gradient energy evenly across all experts, Excitation concentrates it on the experts the router has already identified as most useful. The framework explicitly tested the opposite approach, boosting underutilized experts, and found it significantly underperformed, confirming that stable sparse training requires reinforcing high-consensus routing paths rather than propping up neglected ones.
The results show faster convergence and better final performance, with the model developing cleaner expert specialization instead of the "structural confusion" that standard optimizers produce in deep MoE architectures. That matters. Sharper routing means less wasted compute on experts that aren't contributing, and more effective use of the experts that are.
MoE Is Leaking Into Everything
What's striking about the February 2026 MoE literature isn't just the LLM papers. MoE routing is showing up in domains that have nothing to do with language modeling. TiMi applies multimodal MoE to time series forecasting, using expert specialization to handle the alignment problem between textual causal signals and numerical data. Dai et al. use entropy-triggered MoE routing in graph-based recommendation systems, letting different experts handle different modality combinations based on how uncertain the router is about a given input.
This proliferation matters because it stress-tests the architecture in ways that pure language modeling doesn't. Time series data has different distributional properties than text. Recommendation graphs have sparse, power-law connectivity patterns. If MoE routing struggles with load balance on relatively uniform text token distributions, these harder distributions expose the cracks even faster. The entropy-triggered routing in the recommendation paper is an admission that static top-k routing can't handle real-world input diversity. The same tension between static routing and dynamic workloads shows up in multi-agent orchestration, as we explored in Why Most Agent Orchestration Frameworks Are Wrong.
I've now read half a dozen MoE papers from February alone, and every single one includes some novel routing fix. That's not a sign of a mature architecture. That's a sign of an architecture with a fundamental design flaw that everyone's patching differently.

The Scaling Law Question
The original Chinchilla scaling laws told us how to balance parameters and data for dense models. MoE models break those laws because total parameters and active parameters diverge. Google Brain's work on Switch Transformer scaling suggested that MoE models benefit from scaling total expert count even when per-token compute stays fixed, but with diminishing returns and increasing routing instability.
Here's what the headlines miss: the MoE efficiency frontier isn't a clean curve. It's jagged. You get big wins going from dense to 8-expert MoE. You get diminishing wins going from 8 to 64 experts. And somewhere past 128 experts, the routing problem becomes so severe that you're spending more engineering effort on load balancing than you'd save by just training a bigger dense model. The math doesn't lie.
The 2026 papers suggest we're in the "diminishing returns" zone for naive expert scaling and entering a phase where architectural innovations around routing, optimization, and deployment infrastructure determine whether MoE continues to define the efficiency frontier or gets overtaken by other sparse computation approaches. As we covered in LLM-Powered Swarms and the 300x Overhead Nobody Wants to Talk About, compute efficiency isn't just an academic concern. It directly determines what's deployable.
What This Actually Changes
MoE architectures aren't going away. They're too useful. Every frontier model released in the last year has been either explicitly MoE or uses MoE-like conditional computation. But the naive version of the story, where you just add more experts and watch performance climb, is dead.
The real efficiency frontier in 2026 isn't about how many experts you can cram into a model. It's about utilization rate: what fraction of your total parameters are doing useful work on a given input. The Replicate-and-Quantize approach gets you there through brute-force systems engineering. Excitation gets you there through smarter optimization. Entropy-triggered routing gets you there through input-adaptive gating. All three are admissions that the default MoE recipe is wasteful.
My position: the next generation of frontier models won't be defined by "we trained a bigger MoE." They'll be defined by "we got 80% expert utilization instead of 40%." That's where the scaling law curve bends next. Not more parameters. Better use of the parameters you already have. The labs that figure out routing are the ones that'll define what "efficient" means through 2027.
Sources
Research Papers:
- A Replicate-and-Quantize Strategy for Plug-and-Play Load Balancing of Sparse Mixture-of-Experts LLMs, Zijie Liu, Jie Peng, Jinhao Duan et al. (2026)
- Excitation: Momentum For Experts, Sagi Shaier (2026)
- TiMi: Empower Time Series Transformers with Multimodal Mixture of Experts, Jiafeng Lin, Yuxuan Wang, Huakun Luo et al. (2026)
- Modality-Guided Mixture of Graph Experts with Entropy-Triggered Routing for Multimodal Recommendation, Ji Dai, Quan Fang, Dengsheng Cai (2026)
Industry / Case Studies:
- State of AI Report 2025, Air Street Capital
- Hugging Face Blog: Mixture of Experts Explained, Hugging Face
- Google DeepMind Research Blog, Google DeepMind
Related Swarm Signal Coverage: