▶️ LISTEN TO THIS ARTICLE
In January 2026, 250 engineers packed a side meeting at IETF 124 in Montreal to draft a charter for agent-to-agent communication standards. The Internet Engineering Task Force, the body that standardized HTTP and TCP/IP, now thinks AI agents need their own protocol layer. That should tell you something about how unfinished the current stack really is.
We've covered the protocol wars already: MCP vs. A2A vs. the alphabet soup. That story is about plumbing, about who wins the JSON-RPC turf war. There's a harder problem underneath. Agents can connect to tools and hand off tasks. What they can't do is actually talk to each other: negotiate terms, express uncertainty, resolve conflicting goals, or agree on what a word means.
Plumbing Is Not Language
MCP handles tool integration. A2A handles task delegation. Both pass JSON payloads between endpoints. This works fine when Agent A asks Agent B to "run this SQL query." It falls apart the moment the interaction requires judgment.
Picture two agents from different organizations trying to agree on a delivery timeline. Agent A says three days. Agent B says five. In a human negotiation, both sides would exchange reasoning, push back on assumptions, float compromises. In the current protocol stack, there's no standard way to express "I disagree and here's why." A2A's task model has states like "working," "completed," and "failed." It doesn't have "I think you're wrong."
This isn't new. FIPA tried to solve it in the late 1990s with a formal Agent Communication Language defining 22 speech acts: inform, request, propose, reject. Platforms like JADE implemented it. Researchers studied it for a decade. Practitioners mostly ignored it, because formal performatives crushed the flexibility real-world systems needed.
Twenty-Five Years Later, Same Gap
The irony is thick. We went from FIPA's rigid performatives to natural language prompts that agents throw at each other with zero formal structure. Neither extreme works.

A communication-centric survey from Yan et al. (2025) cataloged four systemic failures in LLM-based multi-agent systems: communication efficiency (agents waste tokens on redundant context), security gaps (no standard way to verify message integrity), inadequate benchmarking (we can't even measure whether agents communicated effectively), and scalability collapse (what works for three agents crumbles at thirty). Swapping MCP for A2A doesn't fix any of them. These are language-level problems, and agents don't have a language.
The AgenticPay benchmark from February 2026 makes this concrete. Researchers built a framework with over 110 multi-round negotiation tasks where buyer and seller agents had to reach deals through natural language. Even frontier models struggled: anchoring effects distorted offers, agents failed to make credible commitments, and multi-issue bargaining deadlocked when agents couldn't decompose proposals into tradeable components.
The Missing Layers
Fleming et al. proposed an answer in their Internet of Agents architecture (revised January 2026). They argue the stack is missing two entire layers. Layer 8, an Agent Communication Layer, would standardize message envelopes and speech-act performatives like REQUEST and INFORM. Layer 9, an Agent Semantic Layer, would handle semantic grounding: binding terms to shared definitions and disambiguating incoming prompts.
The proposal borrows from FIPA while trying to avoid FIPA's rigidity. The core insight: LLM context windows can't grow forever, so agents need to coordinate meaning at the protocol level rather than stuffing everything into a prompt. Current protocols handle transport. Nothing handles meaning.
The IETF agrees. Rosenberg's draft framework for AI agent protocols identifies agent discovery, credential management, and multimodal negotiation as areas needing standardization, all sitting above MCP/A2A. A separate draft proposes a dedicated Agent Context Protocol for sharing situational context, not just data payloads, but reasoning behind requests.
What This Actually Means
The multi-agent coordination community is building skyscrapers on two-story foundations. MCP and A2A solved real problems. But they solved the easy ones: how to call a tool, how to hand off a task. The hard ones, how agents express disagreement, negotiate under uncertainty, build shared understanding across organizational boundaries, remain untouched by any shipping protocol.

This matters more as deployments scale. Three agents in a pipeline can coordinate informally. Thirty agents across five organizations, each with different training data and different definitions of "good enough," can't. The coordination tax stops being about overhead. It becomes about mutual incomprehension.
My prediction: the protocol wars resolve within 18 months, probably with MCP absorbing A2A's task coordination features. The communication gap won't close that fast. It needs something genuinely new, a semantic layer that lets agents reason about each other's intent. FIPA tried it too early. The IETF is circling the problem now. Whoever cracks it will have built the thing that actually matters.
Sources
Research Papers:
- A Survey of AI Agent Protocols — Yang et al. (2025)
- Beyond Self-Talk: A Communication-Centric Survey of LLM-Based Multi-Agent Systems — Yan et al. (2025)
- A Layered Protocol Architecture for the Internet of Agents — Fleming et al. (2025)
- AgenticPay: A Multi-Agent LLM Negotiation System for Buyer-Seller Transactions — (2026)
- A Survey of Agent Interoperability Protocols — Ehtesham et al. (2025)
Industry / Standards:
- Agentic AI Communications: Identifying the Standards We Need — IETF (January 2026)
- Framework, Use Cases and Requirements for AI Agent Protocols — Rosenberg, IETF Draft
- Agent Context Protocol — Liu, IETF Draft
Related Swarm Signal Coverage: