LISTEN TO THIS ARTICLE
The Inference Budget Just Got Interesting
OpenAI's o1 model uses 300x more inference compute than GPT-4 to solve hard math problems. That's not a bug. It's a design choice. But here's what the headlines miss: we're scaling compute at test time without any principled way to decide when to stop.
A new paper from researchers at UW-Madison reframes the entire debate. They found that instruction-tuned models generating longer reasoning chains don't actually improve accuracy beyond a certain point, they just burn tokens. Meanwhile, Large Reasoning Models trained with reinforcement learning show the opposite pattern: more compute reliably translates to better performance, but only up to a task-specific threshold. The gap between these two approaches tells us something important about where the field is headed.
The Problem With Just Adding More Tokens
The dominant paradigm right now is simple: give the model more time to think, get better answers. Chain-of-thought prompting. Self-consistency sampling. Tree search over reasoning paths. All of these techniques scale inference compute by generating more tokens.
But Hu et al.'s resource rationality framework exposes a problem. They tested instruction-tuned models on GSM8K, a grade-school math benchmark, and found accuracy plateauing after about 200 tokens of reasoning. Beyond that, performance didn't improve. The model just kept elaborating without getting any closer to the right answer.
This is the inference equivalent of hiring more people to dig the same hole faster. At some point, you're just adding bodies without adding progress.
The part that actually worries me is the efficiency gap. Instruction-tuned models generated an average of 450 tokens per problem. LRMs trained with RL generated 180 tokens for comparable accuracy. That's a 2.5x difference in compute cost for the same result. In production, where you're paying per token, that gap becomes a serious line item.

What Large Reasoning Models Actually Do Differently
LRMs like DeepSeek-R1 and OpenAI's o1 don't just generate longer chains of thought. They're trained via reinforcement learning to discover reasoning strategies that maximize accuracy under compute constraints. The difference is subtle but critical: they're optimizing for the right answer, not for plausible-sounding explanations.
Hu's team found that LRMs show adaptive resource rationality, they automatically allocate more compute to harder problems. On GSM8K, easy problems got solved in 50-100 tokens. Hard problems required 300-400 tokens. The model learned to detect difficulty and adjust its search accordingly.
This isn't just nice to have. It's the difference between scaling inference compute profitably and burning budget on every query. If a system can't distinguish between problems that need deep search and problems that don't, you're overspending by default.
The Verification Bottleneck Nobody's Solving
Test-time compute scaling has a second problem that's getting worse: verification cost. When you generate 100 candidate reasoning paths via tree search, you need some way to pick the best one. Most systems use a reward model to score each path. That's expensive.
Qu's adaptive compute allocation paper quantifies this. In mathematical reasoning tasks, verification consumes 60-80% of total inference compute. The actual reasoning? That's the cheap part. Scoring all those candidate paths is where the budget goes.
The proposed solution is learned heuristics, train a small model to predict which reasoning paths are worth verifying before you run the expensive reward model. Qu's results show a 40% reduction in verification cost with minimal accuracy loss. But here's the catch: you're now maintaining two models, one of which exists solely to decide whether the first model's output is worth checking.
This is the architectural creep that happens when you optimize for the wrong metric. The research community is treating verification cost as a separate problem when it's actually a symptom of generating too many low-quality candidates in the first place. This connects directly to the broader observability challenges we've covered in production agent deployments, where verification overhead compounds with the need to trace multi-step reasoning paths.
Memory Injection as a Shortcut
Ding et al.'s MeKi paper takes a different approach entirely. Instead of scaling inference compute, they inject task-specific expert knowledge directly into the model's memory at test time. Think of it like giving the model a cheat sheet for specific problem types.
The efficiency gains are real. On edge devices where inference budget is constrained, MeKi achieves 85% of the performance of full test-time scaling at 10% of the compute cost. The trade-off is that you need to know what kind of problem you're solving before you start. It's not a general solution. It's a production engineering hack.
But that's exactly what production needs right now. Most deployed agent systems aren't solving novel problems. They're handling variations on known workflows. Customer service routing. Document classification. Code review. These are bounded domains where a memory-based approach actually works.
I've now read four papers this month on test-time compute scaling, and MeKi is the only one that acknowledges the edge device constraint. Everyone else is optimizing for cloud-scale workloads where you can throw unlimited compute at the problem. That's not where most AI gets deployed.

The Accordion Problem
Yang et al.'s Accordion-Thinking paper addresses a different inefficiency: readability. Long reasoning chains are hard for humans to parse. If you're generating 500 tokens of step-by-step reasoning, nobody's reading all of it. They skim. They skip to the answer.
Accordion-Thinking compresses reasoning chains by generating summaries at regular intervals. The model alternates between detailed reasoning and high-level summaries, creating a hierarchical structure that's easier to audit. The trade-off is added latency, you're now generating summaries on top of reasoning.
The efficiency question is whether this is worth it. If the goal is purely accuracy, probably not. If the goal is human oversight of model reasoning, maybe. But I'm skeptical that we're going to solve test-time efficiency by adding more generation steps.
What This Actually Changes
The research here splits into two camps. One camp is trying to make existing test-time scaling approaches more efficient through better verification, memory injection, or compression. The other camp is questioning whether we should be scaling compute at test time at all.
The resource rationality framing from Hu et al. is the more interesting one. It suggests that instruction-tuned models are misaligned with efficient inference. They're trained to generate helpful explanations, not to find answers efficiently. LRMs trained with RL for accuracy are better at this, but they're also harder to train and less predictable in production.
The practical takeaway for anyone deploying these systems: don't assume more compute equals better performance. Test the relationship on your specific task. If accuracy plateaus after 200 tokens, cap your inference budget there. If you're hitting verification bottlenecks, consider whether you need tree search at all or if beam search with a smaller branching factor would work.
The bigger question is whether test-time compute scaling survives the next 18 months. If LRMs continue improving, we might see a shift back toward models that just solve problems correctly on the first pass instead of searching over 100 possible solutions. That would make all of this optimization work obsolete. This mirrors the pattern we've seen in tool use interfaces, where simpler approaches often win over complex orchestration when the underlying models get better.
But right now, the field is optimizing for the wrong thing. We're treating inference compute as infinite when it's actually the new bottleneck. The models that win in production won't be the ones that can search the deepest. They'll be the ones that know when to stop searching.
Sources
Research Papers:
- Are More Tokens Rational? Inference-Time Scaling in Language Models as Adaptive Resource Rationality, Zhimin Hu, Riya Roshan, Sashank Varma (2026)
- Adaptive Test-Time Compute Allocation via Learned Heuristics over Categorical Structure, Shuhui Qu (2026)
- MeKi: Memory-based Expert Knowledge Injection for Efficient LLM Scaling, Ning Ding, Fangcheng Liu, Kyungrae Kim et al. (2026)
- Accordion-Thinking: Self-Regulated Step Summaries for Efficient and Readable LLM Reasoning, Zhicheng Yang, Zhijiang Guo, Yinya Huang et al. (2026)
Related Swarm Signal Coverage: