Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

While intricate software "scaffolding" can boost an AI agent's performance, progress is overwhelmingly driven by the core model. A new model generation typically achieves the same capabilities with simple prompts that previously required complex engineering.

Related Insights

Overly structured, workflow-based systems that work with today's models will become bottlenecks tomorrow. Engineers must be prepared to shed abstractions and rebuild simpler, more general systems to capture the gains from exponentially improving models.

The true building block of an AI feature is the "agent"—a combination of the model, system prompts, tool descriptions, and feedback loops. Swapping an LLM is not a simple drop-in replacement; it breaks the agent's behavior and requires re-engineering the entire system around it.

The "bitter lesson" of AI applies to product development: complex scaffolding built around model limitations (like early vector stores or agent frameworks) will inevitably become obsolete as the models themselves get smarter and absorb those functions. Don't over-engineer solutions that a future model will solve natively.

Early on, Google's Jules team built complex scaffolding with numerous sub-agents to compensate for model weaknesses. As models like Gemini improved, they found that simpler architectures performed better and were easier to maintain. The complex scaffolding was a temporary crutch, not a sustainable long-term solution.

Early agent development used simple frameworks ("scaffolds") to structure model interactions. As LLMs grew more capable, the industry moved to "harnesses"—more opinionated, "batteries-included" systems that provide default tools (like planning and file systems) and handle complex tasks like context compaction automatically.

The pace of AI model improvement is faster than the ability to ship specific tools. By creating lower-level, generalizable tools, developers build a system that automatically becomes more powerful and adaptable as the underlying AI gets smarter, without requiring re-engineering.

AI development has evolved to where models can be directed using human-like language. Instead of complex prompt engineering or fine-tuning, developers can provide instructions, documentation, and context in plain English to guide the AI's behavior, democratizing access to sophisticated outcomes.

Obsessing over linear model benchmarks is becoming obsolete, akin to comparing dial-up speeds. The real value and locus of competition is moving to the "agentic layer." Future performance will be measured by the ability to orchestrate tools, memory, and sub-agents to create complex outcomes, not just generate high-quality token responses.

To avoid the rapid depreciation of hard-coded systems as LLMs improve, Blitzy's architecture is dynamic. Agents are generated just-in-time, with prompts written and tools selected by other agents based on the latest model capabilities and the specific task requirements.

Recent AI breakthroughs aren't just from better models, but from clever 'architecture' or 'scaffolding' around them. For example, Claude Code 'cheats' its context window limit by taking notes, clearing its memory, and then reading the notes to resume work. This architectural innovation drives performance.

New AI Models Make Complex Agent Scaffolding Obsolete Within Months | RiffOn