Why Your Multi-Agent System is Failing: Escaping the 17x Error Trap of the “Bag of Agents”

landed on arXiv just before Christmas 2025, very much an early present from the team at Google DeepMind, with the title “Towards a Science of Scaling Agent Systems.” I found this paper to be a genuinely useful read for engineers and data scientists. It’s peppered with concrete, measurement-driven advice, and packed with takeaways you can apply immediately. The authors run a large multi-factorial study, facilitated by the tremendous compute available at these frontier labs, to systematically varying key design parameters in order to really understand what drives performance in agentic systems.
Like many industry AI practitioners, I spend a lot of time building Multi-Agent Systems (MAS). This involves a taking complex, multi-step workflow and dividing it across a set of agents, each specialised for a specific task. While the dominant paradigm for AI chatbots is zero-shot, request-response interaction, Multi-Agent Systems (MAS) offer a more compelling promise: the ability to autonomously “divide and conquer” complex tasks. By parallelizing research, reasoning, and tool use, these systems significantly boost effectiveness over monolithic models.
To move beyond simple interactions, the DeepMind research highlights that MAS performance is determined by the interplay of four factors:
- Agent Quantity: The number of specialised units deployed.
- Coordination Structure: The topology (Centralised, Decentralised, etc.) governing how they interact.
- Model Capability: The baseline intelligence of the underlying LLMs.
- Task Properties: The inherent complexity and requirements of the work.
The Science of Scaling suggests that MAS performance success is found at the intersection of Quantity, Topology, Capability, and Task Complexity. If we get the balance wrong we end up scaling noise rather than the results. This post will help you find that secret sauce for your own tasks in a way that will reliably help you build a performant and robust MAS that will impress your stakeholders.
A compelling recent success story of where an optimal balance was found for a complex task comes from Cursor, the AI-powered software development company behind a popular IDE. They describe using large numbers of agents working in concert to automate complex tasks over extended runs, including generating substantial amounts of code to build a web browser (you can see the code here) and translating codebases (e.g., from Solid to React). Their write-up on scaling agentic AI to difficult tasks over weeks of computation is a fascinating read.
Cursor report that prompt engineering is critical to performance, and the specific agent coordination architecture is key. In particular, they report better results with a structured planner–worker decomposition than with a flat swarm (or bag) of agents. The role of coordination is particularly interesting, and it’s an aspect of MAS design we’ll return to in this article. Much like real-world teams benefit from a manager, the Cursor team found that a hierarchical setup, with a planner in control, was essential. This worked far better than a free-for-all in which agents picked tasks at will. The planner enabled controlled delegation and accountability, ensuring worker agents tackled the right sub-tasks and delivered concrete project progress. Interestingly they also find that it’s important to match the right model to the right, finding that GPT-5.2 is the best planner and worker agent compared to Claude Opus 4.5.
However, despite this early glimpse of success from the Cursor team, Multi-Agent System development in the real world continues to be at the boundary of scientific knowledge and therefore a challenging task. Multi-Agent Systems can be messy with unreliable outputs, token budgets lost to coordination chatter, and performance that drifts, sometimes worsening instead of improving. Careful thought is required into the design, mapping it to the particulars of a given use-case.
When developing a MAS, I kept coming back to the same questions: when should I split a step across multiple agents, and what criteria should drive that decision? What coordination architecture should I choose? With so many permutations of decomposition and agent roles, it’s easy to end up overwhelmed. Furthermore, how should I be thinking about the type of Agents available and their roles?
That gap between promise and reality is what makes developing a MAS such a compelling engineering and data science problem. Getting these systems to work reliably, and to deliver tangible business value, still involves a lot of trial, error, and hard-won intuition. In many ways, the field can feel like you’re operating off the beaten path, currently without enough shared theory or standard practice.
This paper by the DeepMind team helps a lot. It brings structure to the space and, importantly, proposes a quantitative way to predict when a given agent architecture is likely to shine and when it’s more likely to underperform.
In the rush to build complex AI, most developers fall into the ‘Bag of Agents’ trap by throwing more LLMs at a problem and hoping for emergent intelligence. But as the recent Science of Scaling research by DeepMind shows, a bag of agents isn’t an effective team, rather it can be a source of 17.2x error amplification. Without the fences and lanes of a formal topology to constrain the agentic interaction, we end up scaling noise rather than an intelligent capability that is likely to solve a business task.
I’ve no doubt that mastering the standardised build-out of Multi-Agent Systems is the next great technical moat. Companies that can quickly bridge the gap between ‘messy’ autonomous agents and rigorous, plane-based topologies will reap the dividends of extreme efficiency and high-fidelity output at scale, bringing massive competitive advantages within their market.
According to the DeepMind scaling research, multi-agent coordination yields the highest returns when a single-agent baseline is below 45%. If your base model is already hitting 80%, adding more agents might actually introduce more noise than value.
In this post I distill nuggets like the above into a playbook that provides you with the right mental map to build the best Multi-Agent Systems. You will find golden rules for constructing a Multi-Agent System for a given task, touching on what agents to build and how they should be coordinated. Further, I’ll define a set of ten core agent archetypes to help you map the landscape of capabilities and make it easier to choose the right setup for your use-case. I’ll then draw out the main design lessons from the DeepMind Science of Scaling paper, using them to show how to configure and coordinate these agents effectively for different kinds of work.
Defining a Taxonomy of Core Agent Archetypes
In this section we will map out the design space of Agents, focusing on the overarching types available for solving complex tasks. Its useful to distill the types of Agents down into 10 basic types, following closely the definitions in the 2023 autonomous agent survey of Wang et al.: Orchestrator, Planner, Executor, Evaluator, Synthesiser, Critic, Retriever, Memory Keeper, Mediator, and Monitor. In my experience, almost all useful Multi-Agent Systems can be designed by using a mixture of these Agents connected together into a specific topology.
So far we have a loose collection of different agent types, it is useful if we have an associated mental reference point for how they can be grouped and applied. We can organize these agents through two complementary lenses: a static architecture of Functional Control Planes and a dynamic runtime cycle of Plan–Do–Verify. The way to think of this is that the control planes provide the structure, while the runtime cycle drives the workflow:
- Plan: The Orchestrator (Control Plane) defines the high-level objective and constraints. It delegates to the Planner, which decomposes the objective into a task graph — mapping dependencies, priorities, and steps. As new information surfaces, the Orchestrator sequences and revises this plan dynamically.
- Do: The Executor translates abstract tasks into tangible outputs (artifacts, code changes, decisions). To ensure this work is grounded and efficient, the Retriever (Context Plane) supplies the Executor with just-in-time context, such as relevant files, documentation, or prior evidence.
- Verify: This is the quality gate. The Evaluator validates outputs against objective acceptance criteria, while the Critic probes for subjective weaknesses like edge cases or hidden assumptions. Feedback loops back to the Planner for iteration. Simultaneously, the Monitor watches for systemic health — tracking drift, stalls, or budget spikes — ready to trigger a reset if the cycle degrades.
As illustrated in our Multi-Agent Cognitive Architecture (Figure 2), these specialised agents are situated within Horizontal Control Planes, which group capabilities by functional responsibility. This structure transforms a chaotic “Bag of Agents” into a high-fidelity system by compartmentalising information flow.
To ground these technical layers, we can extend our human workplace analogy to define how these planes function:
The Control Layer — The Management
- The Orchestrator: Think of this as the Project Manager. It holds the high-level objective. It does not write code or search the web; its job is to delegate. It decides who does what next.
- The Monitor: This is the “Health & Safety” officer. It watches the Orchestrator. If the agent gets stuck in a loop, burns too much money (tokens), or drifts away from the original goal, the Monitor pulls the emergency brake or triggers an alert.
The Planning Layer — The Strategy
- The Planner: Before acting, the agent must think. The Planner takes the Orchestrator’s goal and breaks it down.
- Task Graph / Backlog (Artifact): This is the “To-Do List.” It is dynamic — as the agent learns new things, steps might be added or removed. The Planner constantly updates this graph so the Orchestrator knows the candidate next steps.
The Context Layer — The Memory
- The Retriever: The Librarian. It fetches specific information (docs, previous code) needed for the current task.
- The Memory Keeper: The Archivist. Not everything needs to be remembered. This role summarizes (compresses) what happened and decides what is important enough to store in the Context Store for the long term.
The Execution Layer — The Workers
- The Executor: The Specialist. This acts on the plan. It writes the code, calls the API, or generates the text.
- The Synthesiser: The Editor. The Executor’s output might be messy or too verbose. The Synthesiser cleans it up and formats it into a clear result for the Orchestrator to review.
The Assurance Layer — Quality Control
- The Evaluator: Checks for objective correctness. (e.g., “Did the code compile?” “Did the output adhere to the JSON schema?”)
- The Critic: Checks for subjective risks or edge cases. (e.g., “This code runs, but it has a security vulnerability,” or “This logic is flawed.”)
- Feedback Loops (Dotted Arrows): Notice the dotted lines going back up to the Planner in Figure 2. If the Assurance layer fails the work, the agent updates the plan to fix the specific errors found.
The Mediation Layer — Conflict Resolution
- The Mediator: Sometimes the Evaluator says “Pass” but the Critic says “Fail.” Or perhaps the Planner wants to do something the Monitor flags as risky. The Mediator acts as the tie-breaker to prevent the system from freezing in a deadlock.

To see these archetypes in action, we can trace a single request through the system. Imagine we submit the following objective: “Write a script to scrape pricing data from a competitor’s site and save it to our database.” The request doesn’t just go to a “worker”; it triggers a choreographed sequence across the functional planes.
Step 1: Initialisation — Control Lane
The Orchestrator receives the objective. It checks its “guardrails” (e.g., “Do we have permission to scrape? What is the budget?”).
- The Action: It hands the goal to the Planner.
- The Monitor starts a “stopwatch” and a “token counter” to ensure the agent doesn’t spend $50 trying to scrape a $5 site.
Step 2: Decomposition — Planning Lane
The Planner realises this is actually four sub-tasks: (1) Research the site structure, (2) Write the scraper, (3) Map the data schema, (4) Write the DB ingestion logic.
- The Action: It populates the Task Graph / Backlog with these steps and identifies dependencies (e.g., you can’t map the schema until you’ve researched the site).
Step 3: Grounding — Context Lane
Before the worker starts, the Retriever looks into the Context Store.
- The Action: It pulls our “Database Schema Docs” and “Scraping Best Practices” and hands them to the Executor.
- This prevents the agent from “hallucinating” a database structure that doesn’t exist.
Step 4: Production — Execution Lane
The Executor writes the Python code for Step 1 and 2.
- The Action: It places the code in the Workspace / Outputs.
- The Synthesiser might take that raw code and wrap it in a “Status Update” for the Orchestrator, saying “I have a draft script ready for verification.”
Step 5: The “Trial” — Assurance Lane
This is where the dotted feedback lines spring into action:
- The Evaluator runs the code. It fails because of a
403 Forbiddenerror (anti-scraping bot). It sends a Dotted Arrow back to the Planner: “Target site blocked us; we need a header-rotation strategy.” - The Critic looks at the code and sees that the database password is hardcoded. It sends a Dotted Arrow to the Planner: “Security risk: credentials must be environment variables.”
Step 6: Conflict & Resolution — Mediation Lane
Imagine the Planner tries to fix the Critic’s security concern, but the Evaluator says the new code now breaks the DB connection. They are stuck in a loop.
- The Action: The Mediator steps in, looks at the two conflicting “opinions,” and decides: “Prioritize the security fix; I will instruct the Planner to add a specific step for Debugging the DB Connection specifically.”
- The Orchestrator receives this resolution and updates the state.
Step 7: Final Delivery
The loop repeats until the Evaluator (Code works) and Critic (Code is safe) both give a “Pass.”
- The Action: The Synthesiser takes the final, verified code and returns it to the user.
- The Memory Keeper summarises the “403 Forbidden” encounter and stores it in the Context Store so that next time the agent is asked to scrape a site, it remembers to use header-rotation from the start.
Core Tool Archetypes: The 10 Building Blocks of Reliable Agentic Systems
In the same way we defined common Agent archetypes, we can undertake a similar exercise for their tools. Tool archetypes define how that work is grounded, executed, and verified and which failure modes are contained as the system scales.

As we cover in Figure 5 above, retrieval tools can prevent hallucination by forcing evidence; schema validators and test harnesses prevent silent failures by making correctness machine-checkable; budget meters and observability prevent runaway loops by exposing (and constraining) token burn and drift; and permission gates plus sandboxes prevent unsafe side effects by limiting what agents can do in the real world.
The Bag of Agents Anti-Pattern
Before exploring the Scaling Laws of Agency, it is instructive to define the anti-pattern currently stalling most agentic AI deployments and that we aim to improve upon: the “Bag of Agents.” In this naive setup, developers throw multiple LLMs at a problem without a formal topology, resulting in a system that typically exhibits three fatal characteristics:
- Flat Topology: Every agent has an open line to every other agent. There is no hierarchy, no gatekeeper, and no specialized planes to compartmentalize information flow.
- Noisy Chatter: Without an Orchestrator, agents descend into circular logic or “hallucination loops,” where they echo and validate each other’s mistakes rather than correcting them.
- Open-Loop Execution: Information flows unchecked through the group. There is no dedicated Assurance Plane (Evaluators or Critics) to verify data before it reaches the next stage of the workflow.
To use a workplace analogy: the jump from a “Bag” to a “System” is the same leap a startup makes when it hires its first manager. Early-stage teams quickly realize that headcount does not equal output without an organizational structure to contain it. As Brooks put it in The Mythical Man-Month, “Adding manpower to a late software project makes it later.”
But how much structure is enough?
The answer lies in the Scaling Laws of Agency from the DeepMind paper. This research uncovers the precise mathematical trade-offs between adding more LLM “brains” and the growing friction of their coordination.
The Scaling Laws of Agency: Coordination, Topology, and Trade-offs
The Science of Scaling Agent Systems’ paper’s core discovery is that cranking up the agent quantity is not a silver bullet for higher performance. Rather there exists a rigorous trade-off between coordination overhead and task complexity. Without a deliberate topology, adding agents is like adding engineers to a project without an orchestrating manager: you typically don’t get more valuable output; you will likely just get more meetings, undirected and potentially wastful work and noisy chatter.

Experimental Setup for MAS Evaluation
The DeepMind team evaluated their Multi-Agent Systems across four task suites and tested the agent designs across very different workloads to avoid tying the conclusions to a specific tasks or benchmark:
- BrowseComp-Plus (2025): Web browsing / information retrieval, framed as multi-website information location — a test of search, navigation, and evidence gathering.
- Finance-Agent (2025): Finance tasks designed to mimic entry-level analyst performance — tests structured reasoning, quantitative interpretation, and decision support.
- PlanCraft (2024): Agent planning in a Minecraft environment — a classic long-horizon planning setup with state, constraints, and sequencing.
- WorkBench (2024): Planning and tool selection for common business activities — tests whether agents can pick tools/actions and execute practical workflows.
Five coordination topologies are examined in Towards a Science of Scaling Agent Systems: a Single-Agent System (SAS)and four Multi-Agent System (MAS) designs — Independent, Decentralised, Centralised, and Hybrid. Topology matters because it determines whether adding agents buys useful parallel work or just buys more communication.
- SAS (Single Agent): One agent does everything sequentially. Minimal coordination overhead, but limited parallelism.
- MAS (Independent): Many agents work in parallel, then outputs are synthesised into a final answer. Strong for breadth (research, ideation), weaker for tightly coupled reasoning chains.
- MAS (Decentralised): Agents debate over multiple rounds and decide via majority vote. This can improve robustness, but communication grows quickly and errors can compound through repeated cross-talk.
- MAS (Centralised): A single Orchestrator coordinates specialist sub-agents. This “manager + team” design is typically more stable at scale because it constrains chatter and contains failure modes.
- MAS (Hybrid): Central orchestration plus targeted peer-to-peer exchange. More flexible, but also the most complex to manage and the easiest to overbuild.

This framing also clarifies why unstructured “bag of agents” designs can be very dangerous. Kim et al. report up to 17.2× error amplification in poorly coordinated networks, while centralised coordination contains this to ~4.4× by acting as a circuit breaker.
Importantly, the paper shows these dynamics are benchmark-dependent.
- On Finance-Agent (highly decomposable financial reasoning), MAS delivers the biggest gains — Centralised +80.8%, Decentralised +74.5%, Hybrid +73.1% over SAS — because agents can split the work into parallel analytic threads and then synthesise.
- On BrowseComp-Plus (dynamic web navigation and synthesis), improvements are modest and topology-sensitive: Decentralised performs best (+9.2%), while Centralised is essentially flat (+0.2%) and Independent can collapse (−35%) due to unchecked propagation and noisy cross-talk.
- Workbench sits in the middle, showing only marginal movement (~−11% to +6% overall; Decentralised +5.7%), suggesting a near-balance between orchestration benefit and coordination tax.
- And on PlanCraft (strictly sequential, state-dependent planning), every MAS variant degrades performance (~−39% to −70%), because coordination overhead consumes budget without providing real parallel advantage.
The practical antidote is to impose some structure in your MAS by mapping agents into functional planes and using central control to suppress error propagation and coordination sprawl.
That takes us to the paper’s core contribution: the Scaling Laws of Agency.
The Scaling Laws of Agency
Based on the findings from Kim et al., we can derive three Golden Rules for constructing an effective Agentic coordination architecture:
- The 17x Rule: Unstructured networks amplify errors exponentially. Our Centralized Control Plane (the Orchestrator) suppresses this by acting as a single point of verification.
- The Tool-Coordination Trade-off: More tools require more grounding. Our Context Plane (Retriever) ensures agents don’t “guess” how to use tools, reducing the noise that leads to overhead.
- The 45% Saturation Point: Agent coordination yields the highest returns when single-agent performance is low. As models get smarter, we must lean on the Monitor Agent to simplify the topology and avoid unnecessary complexity.

In my experience the Assurance layer is often the biggest differentiator to improving MAS performance. The “Assurance → Planner” loop transforms our MAS from an Open-Loop (fire and forget) system to a Closed-Loop (self-correcting) system that contains error propagation and allowing intelligence to scale to more complex tasks.
Mixed-Model Agent Teams: When Heterogeneous LLMs Help (and When They Hurt)
The DeepMind team explicitly test heterogeneous teams, in other words a different base LLM for the Orchestrator than for the sub-agents, and mixing capabilities in decentralised debate. The lessons here are very interesting from a practical standapoint. They report three main findings (shown on BrowseComp-Plus task/dataset):
- Centralised MAS: mixing can help or hurt depending on model family
- For Anthropic, a low-capability orchestrator + high-capability sub-agents beats an all–high-capability centralised team (0.42 vs 0.32, +31%).
- For OpenAI and Gemini, heterogeneous centralised setups degrade versus homogeneous high-capability.
Takeaway: a weak orchestrator can become a bottleneck in some families, even if the workers are strong (because it’s the routing/synthesis chokepoint).
2. Decentralised MAS: mixed-capability debate is surprisingly robust
- Mixed-capability decentralised debate is near-optimal or sometimes better than homogeneous high-capability baselines (they give examples: OpenAI 0.53 vs 0.50; Anthropic 0.47 vs 0.37; Gemini 0.42 vs 0.43).
Takeaway: voting/peer verification can “average out” weaker agents, so heterogeneity hurts less than you might expect.
3. In centralised systems, sub-agent capability matters more than orchestrator capability.
Across all families, configurations with high-capability sub-agents outperform those with high-capability orchestrators.
The practical overall key message from the DeepMind team is that if you’re spending money selectively, spend it on the workers (the agents producing the substance), not necessarily on the manager but validate on your model family because the “cheap orchestrator + strong workers” pattern did not generalise uniformly.
Understanding the Cost of a Multi-Agent System
A frequently asked question is how to define the cost of a MAS i.e., the token budget that ultimately translates into dollars. Topology determines whether adding agents buys parallel work or simply buys more communication. To make cost concrete, we can model it with a small set of knobs:
- k = max iterations per agent (how many plan/act/reflect steps each agent is allowed)
- n = number of agents (how many “workers” we spin up)
- r = orchestrator rounds (how many assign → collect → revise cycles we run)
- d = debate rounds (how many back-and-forth rounds before a vote/decision)
- p = peer-communication rounds (how often agents talk directly to each other)
- m = average peer requests per round (how many peer messages each agent sends per peer round)
A practical way to think about total cost is:
Total MAS cost ≈ Work cost + Coordination cost
- Work cost is driven mainly by n × k (how many agents you run, and how many steps each takes).
- Coordination cost is driven by rounds × fan-out, i.e. how many times we re-coordinate (r, d, p) multiplied by how many messages are exchanged (n, m) — plus the hidden tax of agents having to read all that extra context.
To convert this into dollars, first use the knobs (n, k, r, d, p, m) to estimate total input/output tokens generated and consumed, then multiply by your model’s per-token price:
$ Cost ≈ (InputTokens ÷ 1M × $/1M_input) + (OutputTokens ÷ 1M × $/1M_output)
Where:
- InputTokens include everything agents read (shared transcript, retrieved docs, tool outputs, other agents’ messages).
- OutputTokens include everything agents write (plans, intermediate reasoning, debate messages, final synthesis).
This is why decentralised and hybrid topologies can get expensive very fast: debate and peer messaging inflate both message volume and context length, so we pay twice as agents generate more text, and everyone has to read more text. In practice, once agents begin broadly communicating with each other, coordination can start to feel closer to an n² effect.
The key takeaway is that agent scaling is only helpful if the task gains more from parallelism than it loses to coordination overhead. We should use more agents when the work can be cleanly parallelised (research, search, independent solution attempts). Conversely we should be cautious when the task is sequential and tightly coupled (multi-step reasoning, long dependency chains), because extra rounds and cross-talk can break the logic chain and turn “more agents” into “more noise.”
MAS Architecture Scaling Law: Making MAS Design Data-Driven Instead of Exhaustive
A natural question is whether multi-agent systems have “architecture scaling laws,” analogous to the empirical scaling laws for LLM parameters. Kim et al. argue the answer is yes. To tackle the combinatorial search problem — topology × agent count × rounds × model family — they evaluated 180 configurations across four benchmarks, then trained a predictive model on coordination traces (e.g., efficiency vs. overhead, error amplification, redundancy). The model can forecast which topology is likely to perform best, reaching R² ≈ 0.513 and selecting the best coordination strategy for ~87% of held-out configurations. The practical shift is from “try everything” to running a small set of short probe configurations (much cheaper and faster), measure early coordination dynamics, and only then commit full budget to the architectures the model predicts will win.
Conclusions & Final Thoughts
In this post, we reviewed DeepMind’s Towards a Science of Scaling Agent Systems and distilled the most practical lessons for building higher-performing multi-agent systems. These design rules helps us avoid poking around in the dark and hoping for the best. The headline takeaway is that more agents is not a guaranteed path to better results. Agent scaling involves real trade-offs, governed by measurable scaling laws, and the “right” number of agents depends on task difficulty, the base model’s capability, and how the system is organised.
Here’s what the research suggests about agent quantity:
- Diminishing returns (saturation): Adding agents does not produce indefinite gains. In many experiments, performance rises initially, then plateaus — often around ~4 agents — after which additional agents contribute little.
- The “45% rule”: Extra agents help most when the base model performs poorly on the task (below ~45%). When the base model is already strong, adding agents can trigger capability saturation, where performance stagnates or becomes noisy rather than improving.
- Topology matters: Quantity alone is not the story; organisation dominates outcomes.
- Centralised designs tend to scale more reliably, with an Orchestrator helping contain errors and enforce structure.
- Decentralised “bag of agents” designs can become volatile as the group grows, sometimes amplifying errors instead of refining reasoning.
- Parallel vs. sequential work: More agents shine on parallelisable tasks (e.g., broad research), where they can materially increase coverage and throughput. But for sequential reasoning, adding agents can degrade performance as the “logic chain” weakens when passed through too many steps.
- The coordination tax: Every additional agent adds overhead. More messages, more latency, more opportunities for drift. If the task isn’t complex enough to justify that overhead, coordination costs outweigh the benefits of extra LLM “brains”.
With these Golden Rules of Agency in mind, I want to end by wishing you the best in your MAS build-out. Multi-Agent Systems sit right at the frontier of current applied AI, primed to bring the next level of business value in 2026 — and they come with the kinds of technical trade-offs that make this work genuinely interesting: balancing capability, coordination, and design to get reliable performance. In building your own MAS you will undoubtedly discover Golden Rules of your own that expand our knowledge over this uncharted territory.
A final thought on DeepMind’s “45%” threshold: multi-agent systems are, in many ways, a workaround for the limits of today’s LLMs. As base models become more capable, fewer tasks will sit in the low-accuracy regime where extra agents add real value. Over time, we may need less decomposition and coordination, and more problems may be solvable by a single model end-to-end, as we move toward artificial general intelligence (AGI).
Paraphrasing Tolkien: one model may yet rule them all.
📚 Further Learning
- Yubin Kim et al. (2025) — Towards a Science of Scaling Agent Systems — A large controlled study deriving quantitative scaling principles for agent systems across multiple coordination topologies, model families, and benchmarks (including saturation effects and coordination overhead).
- Cursor Team (2026) — Scaling long-running autonomous coding — A practical case study describing how Cursor coordinates large numbers of coding agents over extended runs (including their “build a web browser” experiment) and what that implies for planner–worker style topologies.
- Lei Wang et al. (2023) — A Survey on Large Language Model based Autonomous Agents — A comprehensive survey that maps the agent design space (planning, memory, tools, interaction patterns), useful for grounding your 10-archetype taxonomy in prior literature.
- Qingyun Wu et al. (2023) — AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation — A framework paper on programming multi-agent conversations and interaction patterns, with empirical results across tasks and agent configurations.
- Shunyu Yao et al. (2022) — ReAct: Synergizing Reasoning and Acting in Language Models — Introduces the interleaved “reason + act” loop that underpins many modern tool-using agent designs and helps reduce hallucination via grounded actions.
- Noah Shinn et al. (2023) — Reflexion: Language Agents with Verbal Reinforcement Learning — A clean template for closed-loop improvement: agents reflect on feedback and store “lessons learned” in memory to improve subsequent attempts without weight updates.
- LangChain (n.d.) — Multi-agent — Practical documentation covering common multi-agent patterns and how to structure interaction so you can avoid uncontrolled “bag of agents” chatter.
- LangChain (n.d.) — LangGraph overview — A focused overview of orchestration features that matter for real systems (durable execution, human-in-the-loop, streaming, and control-flow).
- langchain-ai (n.d.) — LangGraph (GitHub repository) — The reference implementation for a graph-based orchestration layer, useful if readers want to inspect concrete design choices and primitives for stateful agents.
- Frederick P. Brooks, Jr. (1975) — The Mythical Man-Month — The classic coordination lesson (Brooks’s Law) that translates surprisingly well to agent systems: adding more “workers” can increase overhead and slow progress without the right structure.



