The popular concept of multiple specialized agents collaborating in a "gossip protocol" is a misunderstanding of what currently works. A more practical and successful pattern for multi-agent systems is a hierarchical structure where a single supervisor agent breaks down a task and orchestrates multiple sub-agents to complete it.

Related Insights

Multi-agent systems work well for easily parallelizable, "read-only" tasks like research, where sub-agents gather context independently. They are much trickier for "write" tasks like coding, where conflicting decisions between agents create integration problems.

To build a useful multi-agent AI system, model the agents after your existing human team. Create specialized agents for distinct roles like 'approvals,' 'document drafting,' or 'administration' to replicate and automate a proven workflow, rather than designing a monolithic, abstract AI.

True Agentic AI isn't a single, all-powerful bot. It's an orchestrated system of multiple, specialized agents, each performing a single task (e.g., qualifying, booking, analyzing). This 'division of labor,' mirroring software engineering principles, creates a more robust, scalable, and manageable automation pipeline.

When building Spiral, a single large language model trying to both interview the user and write content failed due to "context rot." The solution was a multi-agent system where an "interviewer" agent hands off the full context to a separate "writer" agent, improving performance and reliability.

Early on, Google's Jules team built complex scaffolding with numerous sub-agents to compensate for model weaknesses. As models like Gemini improved, they found that simpler architectures performed better and were easier to maintain. The complex scaffolding was a temporary crutch, not a sustainable long-term solution.

Contrary to the trend toward multi-agent systems, Tasklet finds that one powerful agent with access to all context and tools is superior for a single user's goals. Splitting tasks among specialized agents is less effective than giving one generalist agent all information, as foundation models are already experts at everything.

The durable investment opportunities in agentic AI tooling fall into three categories that will persist across model generations. These are: 1) connecting agents to data for better context, 2) orchestrating and coordinating parallel agents, and 3) providing observability and monitoring to debug inevitable failures.

Separating AI agents into distinct roles (e.g., a technical expert and a customer-facing communicator) mirrors real-world team specializations. This allows for tailored configurations, like different 'temperature' settings for creativity versus accuracy, improving overall performance and preventing role confusion.

An experiment showed that given a fixed compute budget, training a population of 16 agents produced a top performer that beat a single agent trained with the entire budget. This suggests that the co-evolution and diversity of strategies in a multi-agent setup can be more effective than raw computational power alone.

Replit's leap in AI agent autonomy isn't from a single superior model, but from orchestrating multiple specialized agents using models from various providers. This multi-agent approach creates a different, faster scaling paradigm for task completion compared to single-model evaluations, suggesting a new direction for agent research.