🎧 LISTEN TO THIS ARTICLE
A January 2025 survey paper cataloged something the RAG community had been feeling for months: the old retrieve-once-generate-once pipeline was dead, and agents had killed it. The paper, from Singh et al. at Colorado State, didn't just name the shift. It mapped four distinct architectural patterns that are now reshaping how production systems handle knowledge retrieval. Those patterns matter because 80% of enterprise RAG projects still fail, and most of them fail for reasons that agentic approaches were specifically designed to fix.
What Broke With Traditional RAG
The standard RAG pipeline has a clean, appealing logic. User asks a question. System searches a vector database. Top results get stuffed into a prompt. Model generates an answer. Ship it.
The problem is that this pipeline makes a series of assumptions that collapse under real-world pressure. It assumes one retrieval pass is enough. It assumes the retrieved documents will be relevant. It assumes the model will actually use what it retrieves. Research keeps proving all three assumptions wrong.
The RAG-E framework found that generators ignore their own retriever's top-ranked documents in 47% to 67% of queries. That's not a minor accuracy dip. That's the generator looking at the evidence and deciding it knows better. We covered this in detail in The RAG Reliability Gap, and the numbers haven't improved since.
The deeper issue is architectural. Traditional RAG treats retrieval as a one-shot operation, something that happens before generation and then gets out of the way. But real questions don't work like that. "What was Company X's revenue growth compared to competitors in Q3 2025?" requires multiple retrievals, across different documents, with intermediate reasoning between each one. A single vector search can't handle that. It was never designed to.
Enter the Agent

Agentic RAG flips the relationship between retrieval and generation. Instead of retrieval being a preprocessing step, it becomes a tool that an agent decides when and how to use. The agent can retrieve once, evaluate what it got, decide the results are garbage, reformulate the query, try again, pull from a different source, validate the new results, and then generate an answer. Or it can skip retrieval entirely if the question doesn't need it.
The seminal survey on agentic RAG by Singh, Ehtesham, Kumar, and Talaei Khoei identifies four design patterns that give agents this flexibility: reflection, planning, tool use, and multi-agent collaboration. These aren't theoretical categories. They map directly onto how production systems are being built right now.
The key difference from basic RAG architecture patterns is agency. In a traditional iterative RAG system, the pipeline is hardcoded: retrieve, check, retrieve again, generate. The loop is fixed. In an agentic system, the agent decides whether to loop at all. It decides which tools to use, how many retrieval passes to make, and when to stop. That decision-making capacity is what separates "iterative" from "agentic."
The Four Patterns
Single-Agent RAG
The simplest agentic pattern puts one agent in charge of the entire retrieval-generation pipeline. The agent has access to tools: vector search, web search, SQL queries, calculators, whatever the task requires. It receives a question, reasons about what information it needs, selects the appropriate tool, evaluates the result, and decides what to do next.
This is what most people mean when they say "agentic RAG." Frameworks like LangGraph, which is now running in production at LinkedIn, Uber, and over 400 other companies, make this pattern accessible. The agent operates in a ReAct-style loop (reason, act, observe) and can course-correct mid-flight.
The strength is simplicity. One agent, one reasoning thread, full visibility into the decision chain. The weakness is bottlenecks. When the question requires expertise across multiple domains, a single agent can struggle to hold all the necessary context, exactly the goldfish brain problem applied to retrieval.
Multi-Agent RAG
The multi-agent pattern assigns specialized sub-agents to different retrieval tasks. A coordinator agent receives the query, decomposes it into sub-questions, and delegates each one to a specialist. One agent handles vector search over internal documents. Another runs SQL queries against structured data. A third searches the web for recent information. The coordinator aggregates the results and generates a unified answer.
This pattern shines when the question spans multiple data sources or requires different retrieval strategies for different parts of the answer. "Compare our Q3 revenue to industry benchmarks" needs internal financial data (structured query) and external market data (web search), and no single retrieval strategy handles both well.
The trade-off is coordination overhead. Every inter-agent message adds latency. Every delegation decision introduces a potential failure point. And debugging a multi-agent retrieval chain is significantly harder than debugging a single-agent loop. Sometimes the coordination tax exceeds the benefit, which is why single agents still beat swarms for many retrieval tasks.
Hierarchical RAG
Hierarchical agentic RAG adds explicit tiers to the multi-agent pattern. A top-level planner decomposes the task into high-level subtasks. Mid-level orchestrators manage groups of retrieval agents. Low-level executors handle individual search and extraction operations. Information flows up through the hierarchy, with each level performing validation and synthesis before passing results to the next.
The A-RAG framework from Du et al. demonstrates this approach with three hierarchical retrieval interfaces: keyword search, semantic search, and chunk read. The agent adaptively selects which granularity to use for each sub-query, retrieving across multiple levels before synthesizing. A-RAG consistently outperforms flat approaches while using comparable or fewer retrieved tokens, which matters when you're paying per token.
Hierarchical patterns report 5-13 percentage point improvements over flat or single-agent baselines in domain-specific and multimodal QA tasks. That's a meaningful gain. But the architecture is the most complex of the four patterns, and complexity has a cost that benchmarks don't always capture.
Adaptive RAG
Adaptive RAG is less a fixed architecture and more a routing strategy. The system classifies incoming queries by complexity and routes them to the appropriate retrieval pipeline. Simple factual questions get a single retrieval pass. Multi-hop reasoning questions trigger an iterative agentic loop. Questions that require real-time data get routed to web search agents.
The logic mirrors how human researchers work. You don't launch a full literature review to answer "What year was GPT-3 released?" But you also don't do a single Google search to answer "How has the transformer architecture influenced protein folding prediction methods?" The adaptive pattern applies the right amount of retrieval effort to each question.
Corrective RAG (CRAG), published by Yan et al., is a concrete implementation of this idea. CRAG uses a lightweight evaluator to score the relevance of retrieved documents and triggers different retrieval actions based on that score: keep the documents if they're good, supplement with web search if they're marginal, discard and re-retrieve if they're bad. CRAG outperformed standard RAG by 19% accuracy on PopQA and 36.6% on PubHealth. Self-RAG takes this further by training the model itself to decide when to retrieve, generating special reflection tokens that control retrieval behavior at inference time. Self-RAG at 7B parameters outperformed ChatGPT on fact verification and biography generation tasks, scoring 81% vs. 71% on fact checking.
What Production Actually Looks Like

The benchmark numbers are encouraging. The production reality is harder.
Agentic RAG on FinanceBench, a benchmark using real 10-K filings and earnings reports, pushed accuracy from 50% (traditional RAG) to roughly 70% through iterative, agent-driven retrieval. That's a 20-point jump, and it came from the agent's ability to decompose financial queries into atomic steps, validate intermediate results, and trigger additional retrievals when the first pass came back incomplete. For financial QA, where the cost of a wrong answer is measured in dollars, that kind of improvement justifies the added complexity.
But production deployments also report a 25-40% reduction in irrelevant retrievals alongside new failure modes that didn't exist in simpler systems. Retrieval loops, where the agent keeps searching without converging on an answer. Incorrect retrieval decisions, where the agent uses the wrong tool for the job. Over-retrieval, where broken confidence calibration causes the agent to pull far more context than necessary, blowing through token budgets.
LangChain's State of Agent Engineering survey found that 57% of organizations aren't fine-tuning models at all, relying instead on base models plus prompt engineering plus RAG. That makes retrieval quality the single most important variable in most production AI systems. Yet 70% of RAG systems still lack systematic evaluation frameworks, according to the same survey. Teams are building agentic retrieval pipelines without the instrumentation to know whether they're actually working.
The security surface also expands. Indirect prompt injection, already a concern with basic RAG, becomes more dangerous when an agent can autonomously decide to retrieve from external sources and then act on what it finds. A poisoned document in a vector database doesn't just corrupt a single answer. If the agent uses that answer as input for its next retrieval step, the corruption cascades.
The Evaluation Gap Nobody Talks About
Here's what frustrates me about the agentic RAG discourse: everyone benchmarks retrieval accuracy, and almost nobody benchmarks retrieval decisions. An agent that retrieves five times when two would have sufficed burned three unnecessary API calls, added seconds of latency, and ate through tokens that a customer is paying for. But there's no standard metric for retrieval efficiency in agentic systems.
The closest thing is "retrieved tokens per correct answer," which A-RAG tracks and optimizes. But that's one framework out of dozens. Most production teams measure whether the final answer is correct and call it a day. They don't measure whether the agent took a sensible path to get there. That's like evaluating a delivery driver solely on whether the package arrived, ignoring that they drove 200 miles to deliver something across town.
This matters because agentic RAG's cost profile is fundamentally different from traditional RAG. A naive pipeline has predictable costs: one embedding lookup, one generation call, done. An agentic pipeline's cost varies with the complexity of the question and the quality of the agent's decisions. A well-calibrated agent might make two retrieval calls. A poorly-calibrated one might make twelve, retrieve 50,000 tokens of marginally relevant context, and produce an answer that's only slightly better than what a single retrieval pass would have given.
Teams deploying agentic RAG need three metrics that most evaluation frameworks don't provide: retrieval decision quality (did the agent choose the right tool?), retrieval efficiency (did the agent use the minimum retrieval steps needed?), and confidence calibration (does the agent's certainty about its answer correlate with actual correctness?). Without these, you're flying blind with an expensive autopilot.
Where Agents Actually Help (and Where They Don't)
Agentic RAG's real value isn't in making every query more accurate. It's in handling the queries that traditional RAG can't handle at all.
Multi-hop questions that require information from multiple documents, synthesized across multiple reasoning steps, are where agents earn their keep. Questions that require different retrieval modalities (vector search plus SQL plus web search) benefit enormously from an agent that can select the right tool. Questions where the first retrieval attempt fails, and the system needs to try a different approach, only work with an agent in the loop.
For simple, single-hop factual questions, the overhead of an agent is pure waste. A well-tuned traditional RAG pipeline will answer "What's our return policy?" faster, cheaper, and just as accurately as an agentic system. The adaptive pattern exists precisely because not every question deserves the full agentic treatment.
The honest assessment: agentic RAG solves real problems that cost real money in production. The context window vs. RAG debate hasn't been settled, and agentic RAG adds a third option to the mix. But it also introduces failure modes, latency, and cost that traditional RAG avoids. The teams getting this right aren't the ones who slapped an agent on every retrieval pipeline. They're the ones who instrumented their systems, measured where traditional RAG was actually failing, and applied agentic patterns surgically to those specific failure points.
The 80% enterprise RAG failure rate won't drop to zero because we added agents to the mix. It'll drop because teams stop treating retrieval as a solved problem and start treating it as a system that needs monitoring, evaluation, and the same rigor we apply to any other critical infrastructure. Agents are a tool for that job. They're not a replacement for doing the job.
Sources
Research Papers:
- Agentic Retrieval-Augmented Generation: A Survey on Agentic RAG — Singh, Ehtesham, Kumar, Talaei Khoei (2025)
- A-RAG: Scaling Agentic RAG via Hierarchical Retrieval Interfaces — Du et al. (2026)
- Corrective Retrieval Augmented Generation — Yan, Gu, Zhu, Ling (2024)
- Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection — Asai, Wu, Wang, Sil, Hajishirzi (2023)
- RAG-E: Retriever-Generator Alignment Evaluation — (2026)
Industry / Case Studies:
- State of Agent Engineering — LangChain (2025)
- Agentic RAG for Financial QA: A SingleStore Optimization Approach — IET Conference Proceedings (2025)
Related Swarm Signal Coverage: