So Daniel sent us this one — and it's a topic that I think a lot of people building AI systems right now need to hear. He's pointing us at a landmark Google DeepMind and MIT study called "Towards a Science of Scaling Agent Systems," and the finding is genuinely counterintuitive: adding more agents to a system is about as likely to hurt performance as help it. The paper tested two hundred and sixty configurations across six benchmarks, and the numbers are striking — independent agents amplify errors seventeen times compared to a single agent, every multi-agent variant tested degraded sequential reasoning by thirty-nine to seventy percent, and the token cost runs one point six to six times higher for matched performance. Daniel's asking what this means for the whole vision of autonomous agent swarms, and whether we're actually heading toward a future of small, curated agent teams rather than large-scale agent societies. Lots to dig into here.
Herman Poppleberry, by the way, for anyone new. And yeah — this is one of those papers where the findings feel obvious in retrospect but run completely against how the industry has been building for the last two years.
The whole pitch has been additive, right? More agents, more specialization, better coverage. It's basically the engineering instinct applied to AI.
And that instinct comes from somewhere real. Microservices, human teams, parallel compute — decomposition works in a lot of contexts. The problem is that agents aren't microservices, and the research is now showing us exactly why in quantitative terms. By the way, today's script is courtesy of Claude Sonnet four point six — our AI collaborator down the road.
So let's start with the seventeen point two times error amplification number, because that's the one that jumped out at me. What's actually happening there mechanically?
So the study tested five canonical architectures. Single agent — one model with a unified memory stream. Independent — multiple agents running in parallel with no communication, results aggregated at the end. Centralized — hub and spoke, an orchestrator delegates to workers and synthesizes outputs. Decentralized — peer-to-peer mesh, agents communicate directly. And hybrid — hierarchical oversight combined with peer-to-peer coordination. The independent architecture is the one hitting seventeen point two times error amplification. And the mechanism is straightforward once you see it: each agent runs without visibility into what the others are doing, so when one agent makes an error, there's no correction signal. The aggregation step at the end doesn't fix errors — it averages them, or in the worst case, it amplifies the wrong answer because multiple agents independently converged on the same mistake.
So you're not getting independent verification, you're getting correlated failures.
That's the key insight. If the agents share the same underlying model, they share the same failure modes. They'll make the same mistakes on the same inputs. So independent doesn't mean independent in any meaningful statistical sense — it means you've multiplied your failure surface without adding any genuine diversity. The centralized architecture does better, four point four times amplification rather than seventeen, because the orchestrator acts as what the paper calls a "validation bottleneck" — it catches errors before they propagate downstream. But even four point four times is not a great number.
And then on sequential tasks it gets worse. The thirty-nine to seventy percent degradation figure — that's for every multi-agent variant, not just the independent one?
Every single one. That's what makes the sequential task finding so significant. The study used a benchmark called PlanCraft, which is planning in a Minecraft environment. The reason it's strictly sequential is that each action changes the inventory that later actions depend on. You can't parallelize that — the state of the world at step eight depends on what happened at steps one through seven. And when you fragment that reasoning across multiple agents, you're essentially forcing each agent to reconstruct context that the previous agent already had. That reconstruction is lossy. The paper quantifies the token cost of that reconstruction — thirty-seven percent of total tokens in multi-agent systems are what they call coordination tokens, paying to re-establish shared state rather than doing actual work.
Thirty-seven percent of your token budget is just... overhead. That's extraordinary when you say it plainly.
And it compounds. The paper's token efficiency numbers are stark. Single agent: sixty-seven tasks per thousand tokens. Centralized multi-agent: twenty-one tasks per thousand tokens. Hybrid: fourteen tasks per thousand tokens. So you're paying three to five times more per unit of work. There's a production case study from a document analysis workflow — three-agent pipeline running forty-seven thousand dollars a month, refactored single-agent version running twenty-two thousand seven hundred. The accuracy difference was two point one percentage points. The team spent three months building the pipeline before anyone measured it against a baseline.
That's a painful paragraph to read if you've just spent three months building a pipeline.
The engineering blog that reported it put it bluntly — "start with one agent, measure it honestly, and add coordination only when the data says you need it." Which is obvious advice in hindsight. But the organizational pressures run the other direction.
Say more about that, because I think this is underappreciated. There's a social dynamic here that the research doesn't fully capture.
The organizational pressure toward complexity is real and it's structural. A five-agent pipeline with a supervisor and a critic layer sounds sophisticated in a design review. "We have a single agent" sounds naive — like you haven't thought hard about the problem. Framework vendors have built their entire value proposition around orchestration as a differentiator. LangGraph, AutoGen, CrewAI — their business model is multi-agent coordination. So there's a whole ecosystem of incentives pushing practitioners toward architectural complexity before they've measured whether they need it.
It's the microservices era of AI, essentially. Everyone built microservices because it was the sophisticated thing to do, and then a decade later people started writing "maybe a monolith was fine actually" blog posts.
The parallel is pretty direct. And the paper from TianPan dot co makes exactly that point — specialization as an architectural strategy is running against the direction of model development, not with it. The single capable agent of today is broadly more capable than the three-agent pipeline of two years ago. As foundation models improve, the set of tasks where multi-agent coordination actually outperforms a capable single agent keeps shrinking.
Which brings us to the capability saturation finding, which I want to spend some time on because it's the most practically useful result in the paper.
The capability saturation effect — the paper calls it a threshold around forty-five percent single-agent accuracy. Once your single agent is hitting roughly forty-five percent on a task, adding more agents yields diminishing or negative returns. Below that threshold, multi-agent coordination can genuinely help — it's essentially compensating for model weakness. Above it, the coordination overhead starts to dominate. The theoretical implication is significant: multi-agent coordination is a workaround for capability gaps, not a permanent architectural advantage.
So the right question before you architect a multi-agent system is not "how do I decompose this task" — it's "what does my single-agent baseline actually score?"
That's the forty-five percent rule as an engineering heuristic. Measure first. Most teams aren't doing this — they're reaching for multi-agent complexity before establishing what a single capable agent can do. The paper found that their predictive model, which uses measurable task properties like tool count, decomposability, and sequential dependencies, correctly identifies the optimal coordination strategy for eighty-seven percent of unseen configurations. That's a genuinely useful result because it suggests we're moving toward principled agent design rather than guesswork.
Okay, but let's steelman the multi-agent case, because it's not all bad news. The paper does find an eighty percent improvement on parallelizable tasks.
The alignment principle is real and the numbers are large. On Finance-Agent, which is a financial reasoning benchmark with highly parallelizable subtasks — one agent analyzing revenue trends, another analyzing cost structures, another looking at market comparisons — centralized coordination improved performance by eighty point eight percent over single-agent. That's a genuine, substantial gain. The key is that the subtasks are truly independent: the revenue analysis doesn't depend on the cost analysis being done first. When you can decompose a problem into genuinely non-overlapping workstreams, parallel agents shine.
And the SWE-bench coding result is interesting too — that's a real gain on a real benchmark.
The coding case is instructive because of why it works. A multi-agent team — manager, researcher, engineer, reviewer — hit seventy-two point two percent on SWE-bench Verified, compared to sixty-five percent for a solo agent. Seven point two percentage points is meaningful. But the reason it works in code is that code has an objective verification signal. Tests either pass or fail. That gives the reviewer agent grounded feedback — it's not just arguing about the quality of reasoning, it's running the tests and reporting concrete results. The adversarial review pattern works when you have objective signals to anchor it.
Compare that to Multi-Agent Debate, which is the more general pattern, and the results are much weaker.
The ICLR 2025 evaluation of Multi-Agent Debate frameworks is blunt. They tested five frameworks — the original Du et al. MAD framework, Multi-Persona, Exchange-of-Thoughts, ChatEval, and AgentVerse — across nine benchmarks using GPT-4o-mini and Llama three point one eight billion. The verdict: current MAD frameworks fail to consistently outperform simple single-agent test-time computation strategies. Most can't beat Chain-of-Thought prompting. Most can't beat Self-Consistency, which just resamples from the same single agent multiple times. Increasing agent count or debate rounds does not reliably improve performance.
Self-Consistency beating a multi-agent debate system is a pretty damning result. You're literally just running the same model more times.
And it makes sense once you understand the failure mode. The evaluation found that MAD methods are "overly aggressive" — they turn correct answers into incorrect ones at a higher rate than they fix wrong answers. The Multi-Persona framework is the worst offender because the devil's advocate agent has a structural mandate to oppose, regardless of whether the original answer was right. You've built in a mechanism that degrades correct reasoning.
There's a finding from the MAST paper — the Berkeley forensic analysis — that connects here. They built a taxonomy of why multi-agent systems fail and the failure modes cluster around coordination, not capability.
The MAST paper is fascinating as a forensic document. They collected over sixteen hundred annotated failure traces across seven popular multi-agent frameworks and identified fourteen distinct failure modes in three categories. System design issues — problems baked in at architecture time. Inter-agent misalignment — coordination failures including things like conversation reset, task derailment, information withholding, and one they call "ignored other agent's input," which is exactly what it sounds like. And task verification failures — the system doesn't check its own outputs adequately. The inter-annotator agreement was zero point eight eight kappa, which is high — these failure modes are real and identifiable, not noise. And the finding is that most failures aren't "the model wasn't smart enough" — they're coordination and communication failures.
Which brings up the topology question, because not all multi-agent architectures are equally bad at coordination.
The MacNet paper from Tsinghua is the most optimistic of the studies, and it's worth taking seriously because it identifies where structure actually matters. MacNet organizes agents into directed acyclic graphs rather than fixed topologies. The key finding is that irregular topologies outperform regular ones. Fixed symmetric structures — rings, grids, star patterns — perform worse than adaptive irregular ones. And the intuition makes sense: irregular topologies allow information to flow along paths that match the actual dependency structure of the task. If your task has a specific dependency graph — step A must precede step B, but step C is independent of both — your communication topology should reflect that. Forcing information through a fixed grid means some of it travels through unnecessary hops.
So the right multi-agent architecture isn't a pattern you can pick off a shelf — it's something you derive from the task structure.
That's the practical implication. And it connects to the predictive model in the DeepMind paper — the task properties that predict optimal architecture are tool count, decomposability, and sequential dependency structure. If you map those properties for your specific task, you can derive the right topology rather than defaulting to whatever the framework's default configuration is. The MacNet paper also identifies a collaborative scaling law — performance follows a logistic growth curve as agents scale. Improvement up to a threshold, then plateau or decline. The sweet spot they identify is around four to five agents or three to four debate rounds before coordination overhead starts winning.
That logistic curve is interesting because it mirrors what we see in a lot of other scaling contexts. It's not that more is never better — it's that the curve has a ceiling that arrives faster than you'd expect.
The MacNet paper makes a specific point about this — collaborative emergence occurs earlier than traditional neural scaling emergence. Neural scaling intuitions from training compute suggest that gains compound over long ranges. Multi-agent collaboration hits its ceiling much faster. So if you're importing intuitions from "bigger models keep getting better," those intuitions don't transfer cleanly to "more agents keep getting better."
Let's talk about model diversity for a minute, because this is the finding that I think reframes the whole design question.
The MAD evaluation found that combining different foundation models — specifically GPT-4o-mini paired with Llama three point one seventy billion — shows more promise than same-model multi-agent debate. That's a significant result because it suggests the real variable isn't agent count, it's reasoning diversity. Two agents with genuinely different model architectures, training distributions, and failure modes can provide real error correction. Five agents all running the same model are just expensive correlated sampling.
So the design question isn't "how many agents" — it's "how different are the reasoning approaches in the system."
Which reframes the whole architecture conversation. If you're building a verification layer, the most valuable thing you can do is use a different foundation model for the critic than you used for the generator. Not because it's a larger model, but because its failure modes don't overlap. That's genuine independence in the statistical sense.
Where does this leave the broader vision of agent swarms? Because there's been a lot of investment — and a lot of hype — around the idea of emergent coordination at scale.
The research points toward something much more modest than swarms. Carefully curated small teams — four to five agents maximum — with explicit coordination protocols, clear task decomposition, and objective verification signals. The "bag of agents" approach, just throwing more agents at a problem, is what the TianPan piece calls the multi-agent equivalent of hoping a bigger model will fix poor prompt engineering. It's the same category of mistake — substituting scale for design.
And the Gartner forecast from last June is looking increasingly well-calibrated given all of this.
Over forty percent of agentic AI projects canceled by end of twenty twenty-seven. The reasoning was escalating costs, unclear business value, and inadequate risk controls. That forecast was made before the full DeepMind findings were widely circulated, and the research now gives you the mechanistic explanation for why those projects will fail. The reliability compounding problem alone is a serious enterprise risk. At ninety-seven percent per-step reliability — which sounds high — a ten-step agentic workflow has a seventy-four percent end-to-end success rate. At ninety-nine point five percent per step, you get ninety-five percent end-to-end. For enterprise deployments running millions of workflows, that gap is the difference between a reliable product and an unreliable one.
And I suspect most teams building these systems haven't actually measured their per-step reliability and done the compounding math.
Almost certainly not. The measurement culture in agentic AI is still immature. Teams are measuring benchmark performance on the overall task, not per-step reliability in their specific deployment environment. The MAST paper's forensic approach — annotating failure traces — is the kind of methodology that production teams need to adopt, but it requires actually collecting and analyzing failure data, which most teams aren't set up to do.
So practically, what does a sensible decision framework look like for an engineer building agentic systems right now?
The research converges on a pretty clear decision tree. Start with a single agent. Measure it honestly on your specific task — not on a published benchmark, on your actual task. If your single-agent baseline is already above forty-five percent accuracy, adding agents is likely to cost more than it gains. If the task is primarily sequential, single agent wins. If the tool set fits in one context window, single agent wins. If your latency budget is under five seconds, single agent wins. The cases where multi-agent genuinely earns its complexity are narrower than the industry assumed: demonstrably parallelizable subtasks with no ordering dependencies, objective verification signals that justify adversarial review, compliance requirements that mandate data isolation, or genuinely different foundation models in the critic and generator roles.
And the unified architecture results are worth flagging here because they're directly competitive with multi-agent systems on tasks that people assumed required orchestration.
Surfer two achieving ninety-seven point one percent on WebVoyager using a unified architecture. OpenAI's Deep Research system outperforming prior multi-agent research pipelines as a single model optimized for web browsing and data analysis. These aren't cases where single agent is "good enough" — they're cases where the unified architecture is actually better. The reasoning is that web navigation and research are fundamentally sequential — each page you visit changes what you know, which changes what you look for next. Fragmenting that across agents doesn't help, it just creates coordination overhead on top of a task that was already hard.
I want to come back to something you said earlier about the direction of model development, because I think it's the most underappreciated point in this whole discussion.
The capability trajectory is working against the multi-agent value proposition over time. As foundation models improve, the tasks where a single capable agent falls short — and where multi-agent coordination genuinely compensates — keep shrinking. The forty-five percent saturation threshold isn't a fixed number — as models get stronger, more tasks push above it. So an architecture that made sense in twenty twenty-four, when single-agent performance on complex tasks was genuinely limited, may not make sense now. The engineers who built those three-agent pipelines weren't wrong given what they knew — they were building with the tools available. The problem is that the tools improved faster than the architecture assumptions updated.
Which is actually a reasonable thing to have happen — the industry is learning in real time. The issue is the organizational inertia that keeps the old patterns in place after the evidence has moved.
The frameworks are already built, the teams are already trained on them, the design patterns are already documented. There's a switching cost to simplifying. And "we're going to remove the orchestration layer and replace it with a single agent" is a harder conversation to have than "we're adding a new capability." Simplification requires admitting that the complexity you built wasn't necessary, which is a harder organizational move than adding more complexity.
Alright, practical takeaways — what should someone listening to this actually do differently?
Three things. First: run your single-agent baseline before you architect anything. Not a toy baseline, a genuine measurement on your actual task. If it's above forty-five percent, you probably don't need multi-agent coordination. Second: if you do build multi-agent, use model diversity rather than agent count as your primary design variable. A two-model system with genuinely different architectures is more valuable than a five-agent system running the same model. Third: map your task's dependency structure before choosing a topology. Is it parallelizable? Sequential? What are the ordering dependencies? The architecture should reflect the task structure, not default to whatever the framework's default configuration is.
I'd add a fourth, which is: measure per-step reliability and do the compounding math before you deploy. The gap between ninety-seven percent and ninety-nine point five percent per step sounds small and the end-to-end difference is twenty-one percentage points on a ten-step workflow.
That's the one that will catch teams by surprise in production. The benchmark numbers look fine, the per-step reliability looks fine, and then the system fails one in four times at scale and nobody understands why. The math is not intuitive until you work through it explicitly.
The deeper implication here — and I think this is what Daniel is really gesturing at with his question about agent swarms — is that the vision of emergent coordination at scale may just not be how capable AI systems end up working. The research is pointing toward something more deliberate and more engineered.
The swarm vision assumed that coordination would emerge from scale the way capabilities emerge from scale in training. But coordination isn't an emergent property of agent count — it's an engineered property of task decomposition, topology, and verification. You can't throw agents at a problem and expect coordination to emerge. You have to design it. And once you accept that, the question stops being "how many agents" and starts being "what is the minimum coordination structure that solves this specific problem." That's a much more tractable engineering question, and the research is now giving us the tools to answer it.
Good framing to end on. Minimum viable coordination rather than maximum visible complexity.
That should be on a poster somewhere in every AI engineering team's office.
Alright. Real practical stuff in this one — if you're building agentic systems, this research is directly relevant to decisions you're making right now. Big thanks to our producer Hilbert Flumingtop for keeping the machine running. And thanks to Modal for the GPU credits that make this whole pipeline possible. This has been My Weird Prompts. If you haven't followed us on Spotify yet, we're there — search My Weird Prompts and hit follow so you don't miss an episode.
See you next time.