LISTEN TO THIS ARTICLE
The Control Interface Problem in Physical AI
NVIDIA just released a video foundation model that can simulate physical worlds with startling accuracy. A team from the University of Illinois Urbana-Champaign and Hanyang University built an AI agent that controls a nuclear reactor simulator. Another group demonstrated vision-language-action models that let robots learn personalized behaviors from human feedback. The common thread? None of them solved the actual hard part.
The hard part isn't world simulation. It's not vision-language integration. It's the control interface, the moment where a model's understanding must translate into physical action that doesn't break things, hurt people, or waste massive amounts of money. I've read eight papers this month on physical AI, and exactly one of them acknowledges this problem explicitly.
Why World Models Don't Mean Physical Intelligence
NVIDIA's Cosmos-Predict2.5 can generate realistic video predictions of physical environments. It unifies Text2World, Image2World, and Video2World generation in a single flow-based architecture, trained on 200 million curated video clips. The model produces coherent multi-second predictions of how objects move, collide, and interact. It's technically impressive.
The gap between prediction and control is where things fall apart. A model that predicts a robot arm will collide with a workpiece isn't the same as a model that prevents the collision. The difference isn't academic, it's the difference between a simulation tool and a deployment-ready system.
The UIUC team's work on nuclear reactor control makes this explicit. Their domain-specific foundation model for reactor control operates in a simulator environment where the penalty for failure is restarting the simulation. The paper highlights a core limitation: "even frontier vision-language models achieve only 50-53% accuracy on basic quantitative physics tasks," performing as approximate guessers that preserve semantic plausibility while violating physical constraints.
The implication is clear: most physical AI research tests models in video games and simulated warehouses where failure costs nothing. Real physical systems have failure modes that matter. A humanoid robot in a manufacturing line can't crash and restart. The control interface has to work the first time.
The distinction between predicting outcomes and controlling systems exposes a deeper architectural problem. World models excel at forward simulation, they take a state and predict what happens next. Control systems require inverse reasoning: given a desired outcome, what actions produce it? Current foundation models approach this through trial and error in simulation, which works until the sim-to-real gap kills performance. The control interface needs closed-loop feedback at speeds measured in milliseconds, not the inference latencies typical of large models.
The VLA Scaling Story Nobody Wants to Hear
Vision-Language-Action models are pitched as the path to general-purpose robotics. The idea is seductive: train a large multimodal model on millions of robot demonstrations, let it learn a general policy for manipulation, then fine-tune for specific tasks. Scaling up training data should produce better performance, just like it did for language models.
The data from on-device VLA deployment tells a different story. Hardware constraints limit the practical size of deployable VLAs to models that fit inside the compute envelope of edge devices. A paper on hardware co-design scaling laws for on-device LLMs demonstrates that roofline modeling, the relationship between memory bandwidth, compute throughput, and model size, creates hard limits on what architectures can actually run in physical systems.
Those limits matter more than most research admits. A warehouse robot can't depend on cloud inference with 200ms round-trip latency. Manufacturing robots need sub-10ms response times. The VLA models that achieve impressive results in lab demos often can't run fast enough to control real hardware at the speeds required for useful work.
The scaling assumption breaks down further when you examine what VLAs actually learn. Personalized agent training from human feedback requires substantial per-user interaction data to achieve meaningful alignment. One study evaluated 40 users across 30 scenarios per phase, treating live interaction as the primary learning signal rather than static datasets. The model that works well on generic pick-and-place tasks doesn't automatically adapt to the specific way a particular factory floor wants objects handled.
The economics of VLA deployment reveal another constraint that research papers ignore: the cost of failure during learning. Every failed grasp attempt in a real factory costs time, potentially damages equipment, and might require human intervention to reset. Language models can generate thousands of bad outputs during training without consequence. Physical systems pay for every mistake in real dollars. This asymmetry between digital and physical learning costs changes the optimization problem. The VLA that needs 10,000 attempts to learn a new task might be acceptable in simulation but economically unviable in production.
Geospatial Reasoning Is Worse Than You Think
GPSBench evaluated whether large language models understand GPS coordinates. The results are bad enough to worry about any physical AI system that operates in the real world.
Current frontier models fail basic geospatial reasoning tasks. Given two GPS coordinates, models regularly misidentify which city they correspond to, calculate incorrect distances between points, and fail to infer obvious spatial relationships. Even the best-performing model, Gemini-2.5-Pro, manages only 23% accuracy on coordinate-to-city mapping. Exact city identification collapses to between 1% and 23% across all 14 models tested, despite country-level identification reaching 59-97%.
This isn't a minor limitation. Any embodied agent that operates outdoors, delivery robots, agricultural drones, autonomous vehicles, needs reliable geospatial reasoning. The current generation of foundation models can't provide it. They can generate plausible-sounding descriptions of routes and locations. They can't consistently map between coordinates and physical space.
The failure mode is consistent across model architectures and scale. Scaling up parameters doesn't fix the underlying problem. These models don't learn stable representations of geographic space from text training data. They learn lexical patterns that work often enough to pass casual evaluation but fail under systematic testing.
The geospatial reasoning problem compounds when you consider how outdoor robots actually need to use location data. It's not enough to know that two coordinates are in different cities. The system needs to understand elevation changes, terrain types, local regulations, weather patterns, and dozens of other spatially-indexed factors. Foundation models trained on text and images don't build these spatial indexes because they're not present in the training data in a form that supports geometric reasoning. The models memorize facts about places without learning the geometric relationships between them.
Domain-Specific Models Are Winning Quietly
While general-purpose physical AI gets the headlines, domain-specific foundation models are shipping. The nuclear reactor control work from UIUC demonstrates why. Instead of training a generalist agent that could theoretically operate any industrial process control system, the team built a specialized 360-million-parameter model trained exclusively on nuclear reactor simulation data.
The specialized model achieves control performance that general agents can't match. It understands the specific physics of reactor dynamics, the constraints of the control system, and the safety boundaries that matter for this particular application. It doesn't waste capacity learning to manipulate warehouse boxes or work through office environments.
This pattern repeats across manufacturing AI applications. Human-AI co-embodied systems for scientific experimentation and manufacturing use task-specific models rather than general VLAs. The models learn narrow skills: aligning microscope samples, calibrating measurement equipment, adjusting manufacturing parameters. They don't try to learn general manipulation policies.
The economic argument for domain-specific models is straightforward. Training a specialized model costs less than training a generalist. Deployment requires fewer compute resources. Failure modes are easier to characterize and control because the operating domain is constrained. The return on investment shows up faster.
Domain specialization solves another problem that generalist approaches ignore: regulatory compliance. A nuclear reactor controller needs to prove safety properties before deployment. A general-purpose agent that could theoretically control any industrial system can't provide those guarantees because its behavior depends on training data that includes tasks unrelated to reactor control. The specialized model's narrow scope makes formal verification tractable. This matters more than benchmark performance when the failure modes include radiation release.
The Benchmark Problem Is Getting Worse
CostNav introduced a benchmark that measures physical AI agents on real-world economic cost. Instead of success rate on simulated tasks, it evaluates the actual dollar cost of completing objectives. Travel time, energy consumption, and resource usage all factor into the score.
Testing agents against CostNav reveals that current systems optimize for metrics that don't align with deployment economics. An agent that achieves 95% success rate but uses 3x more energy than necessary looks good on traditional benchmarks. It fails on cost-aware evaluation.
I've seen this movie before with language model benchmarks. Models optimize for whatever the benchmark measures, not what actually matters for production use. Physical AI is walking into the same trap. Success rate matters. So does energy efficiency, maintenance overhead, and failure recovery cost. The multimodal reasoning capabilities that make foundation models impressive in demos become liabilities when measured against deployment costs, they require compute resources that directly impact operating economics.
The benchmark that measures what matters for physical AI doesn't exist yet. CostNav is a start, but it only covers basic tasks. Manufacturing tasks need benchmarks that account for downtime cost, error recovery, and integration overhead. Service robot benchmarks need to measure user satisfaction, not just task completion. The disconnect between research metrics and deployment reality means that papers reporting impressive results often describe systems that can't economically deploy.
What Hardware Constraints Actually Mean
The roofline modeling paper lays out the math that most physical AI research ignores. The maximum throughput of an on-device VLA is bounded by:
Throughput = min(Compute Capacity, Memory Bandwidth × Operational Intensity)
Operational intensity is the ratio of compute operations to memory accesses. Transformer architectures have low operational intensity relative to the model sizes used in current VLAs. This creates a memory bandwidth bottleneck that limits practical throughput.
The implication: you can't just scale up model size and expect proportional improvements in on-device performance. Past a certain point, larger models run slower because they're memory-bound, not compute-bound. The sweet spot for deployable VLAs is smaller than the models that achieve top benchmark results.
This explains why production robotics systems still use classical control methods for most tasks. A PID controller running at 1kHz on a microcontroller outperforms a VLA model running at 10Hz on an edge GPU when the task allows it. The VLA only wins when the task requires the flexibility of learned policies.
The memory bandwidth problem gets worse with multimodal models. Vision encoders, language models, and action decoders all compete for the same memory interface. Batching helps in datacenter inference but doesn't work for real-time control where latency matters more than throughput. Edge devices can't afford the power budget for high-bandwidth memory interfaces. These constraints create a hard ceiling on model size that's well below what research papers assume. The 7B parameter VLA that works in the lab becomes a 1B parameter model in production, and that size reduction costs capability.
The Agentic Framework Trap
RoboGene proposes a diversity-driven agentic framework for generating robot training data. The system uses large language models to generate diverse task specifications, then generates synthetic training data for those tasks. The goal is to create broad coverage of possible manipulation scenarios without manually collecting real-world demonstrations.
The approach works in simulation. It doesn't address the distribution mismatch between synthetic training data and real-world deployment conditions. Sim-to-real transfer remains the hard part. Generating more diverse synthetic data doesn't close the reality gap.
The part that actually worries me is the implicit assumption that more data diversity automatically leads to better generalization. Language models show this isn't true. More training data helps until it doesn't. The relationship between data diversity and policy reliability isn't linear.
Physical AI research treats data scaling as the solution to generalization problems without testing whether that's true for embodied systems. The evidence from deployed systems suggests it's not. Domain-specific models with less diverse, higher-quality data outperform generalist models with massive synthetic datasets.
The agentic generation approach also amplifies a problem that benchmarks already struggle with: the gap between what models learn to do and what they can reliably do under deployment conditions. Synthetic data generated by LLMs inherits the statistical patterns of language model outputs. Those patterns don't match the physical constraints of real robots operating in real environments. The diversity-driven approach optimizes for coverage of the space of possible tasks, but physical systems care more about reliability on actual tasks than coverage of hypothetical scenarios.
Manufacturing Reality Doesn't Care About Your Foundation Model
The manufacturing AI research makes the clearest argument for practical constraints. Factory floors have narrow operating tolerances. A vision system that's 98% accurate might be impressive in a research paper. It's unusable in a production line where 2% defect rate costs millions in waste and rework.
This creates a different optimization target than research environments use. Manufacturing AI needs deterministic failure modes, not just high average performance. The system must be able to guarantee that certain classes of errors never occur, even if that means accepting lower performance on edge cases.
Current foundation models can't make those guarantees. They're trained to maximize average-case performance across diverse scenarios. Manufacturing needs worst-case guarantees in a narrow operating domain. The architecture mismatch is fundamental.
Human-AI co-embodied systems partially solve this by keeping humans in the loop for high-stakes decisions. The AI handles routine tasks and assists with complex operations, but a human operator maintains override authority. It's less autonomous than the vision for general-purpose robots, but it's what actually ships.
The co-embodied approach reveals something important about the control interface problem. The systems that work in manufacturing don't try to fully automate tasks. They augment human capability in specific ways that keep failure modes manageable. The robot doesn't replace the factory worker, it handles the physically demanding parts of assembly while the human manages quality control and handles exceptions. This division of labor acknowledges that current AI can't match human reliability on the tasks that matter most for production quality. Foundation models contribute vision and planning capabilities. Classical control systems handle the actual manipulation. Humans make the decisions that prevent expensive mistakes.
Simulation Fidelity Is the Wrong Target
Cosmos-Predict2.5's impressive world simulation capabilities suggest a path toward training physical AI systems in simulation before deployment. Generate enough realistic simulated environments, train policies that work in simulation, then transfer to real hardware.
The fidelity of the simulation isn't the limiting factor. The sim-to-real gap comes from unmodeled dynamics that don't show up in video predictions. Friction coefficients, material compliance, sensor noise, actuator backlash, these physical properties matter for control but aren't captured in visual predictions.
Better world models might help with high-level planning. They don't solve the control interface problem. The robot that can predict where an object will land after tossing it still needs a low-level controller that accounts for the actual dynamics of its actuators and sensors.
High-fidelity simulation creates another problem: computational cost. The realistic physics simulation required to capture contact dynamics, material properties, and sensor characteristics costs orders of magnitude more compute than video prediction. Training policies in high-fidelity simulation becomes prohibitively expensive at the scale required for foundation models. Researchers make trade-offs, simplifying physics to make training tractable, which brings back the sim-to-real gap they were trying to eliminate. The systems that work in deployment tend to use simple simulation for initial training, then rely heavily on real-world fine-tuning to bridge the gap. That real-world data collection is the expensive part that simulation was supposed to avoid.
Control Theory Meets Foundation Models
The integration of foundation models with classical control systems exposes fundamental mismatches in how these systems represent and reason about physical action. PID controllers optimize for stability and reference tracking using continuous feedback. Foundation models output discrete action sequences based on high-dimensional sensor observations. The interface between these paradigms requires translation layers that introduce latency and approximation error.
Current approaches treat foundation models as high-level planners that output waypoints or action primitives for classical controllers to execute. This works for tasks where planning and control can be cleanly separated. It fails for dynamic tasks where the boundary between planning and control blurs. Catching a thrown object requires prediction, planning, and control to happen simultaneously at timescales faster than foundation model inference.
The alternative approach, training end-to-end policies that directly map sensors to actuator commands, struggles with the sample efficiency problem. These policies need millions of training examples to learn what a PID controller captures in a few dozen parameters. The foundation model's advantage in generalization doesn't compensate for its disadvantage in sample efficiency for simple control tasks. This creates an awkward middle ground where neither pure learning nor pure control theory dominates, and hybrid systems accumulate the worst failure modes of both approaches.
What This Actually Changes
Physical AI is shipping, but not in the form most research assumes. Domain-specific models for constrained applications are demonstrating value. General-purpose humanoid robots remain stuck in demonstration videos and research labs, much like the robots with reasoning capabilities that show promise in controlled environments but struggle with real-world deployment.
The gap between research and deployment comes down to the control interface. Foundation models produce impressive predictions, language understanding, and visual reasoning. Translating that into reliable physical action at the speed and precision required for useful work remains unsolved.
This will change faster in domains where failure costs are low and operating envelopes are wide. Warehouse robots can afford occasional mistakes. Nuclear reactors can't. Manufacturing lines with tight tolerances can't. The economics of deployment favor specialized solutions over general-purpose agents.
The next 18 months will determine whether the VLA scaling hypothesis holds up under real deployment conditions. Current evidence suggests it won't, hardware constraints and control interface problems limit how far scaling can take us. The systems that ship will look more like the UIUC team's domain-specific reactor controller than like NVIDIA's general-purpose world simulator.
Don't bet on humanoid robots in manufacturing lines by 2027. Do bet on more specialized AI systems that handle narrow tasks reliably. The future of physical AI is less dramatic and more useful than the demos suggest.
Sources
Research Papers:
- World Simulation with Video Foundation Models for Physical AI, Arslan Ali et al., NVIDIA (2025)
- Agentic Physical AI toward a Domain-Specific Foundation Model for Nuclear Reactor Control, Yoonpyo Lee, Kazuma Kobayashi, Sai Puppala et al. (2025)
- Learning Personalized Agents from Human Feedback, Kaiqu Liang, Julia Kruk, Shengyi Qian et al. (2026)
- GPSBench: Do Large Language Models Understand GPS Coordinates?, Thinh Hung Truong, Jey Han Lau, Jianzhong Qi (2026)
- Hardware Co-Design Scaling Laws via Roofline Modelling for On-Device LLMs, Luoyang Sun, Jiwen Jiang, Yifeng Ding et al. (2026)
- Human-AI Co-Embodied Intelligence for Scientific Experimentation and Manufacturing, Xinyi Lin, Yuyang Zhang, Yuanhang Gan et al. (2025)
- RoboGene: Boosting VLA Pre-training via Diversity-Driven Agentic Framework for Real-World Task Generation, Yixue Zhang, Kun Wu, Zhi Gao et al. (2026)
- CostNav: A Navigation Benchmark for Cost-Aware Evaluation of Embodied Agents, Haebin Seong, Sungmin Kim, Yongjun Cho et al. (2025)
Related Swarm Signal Coverage: