Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

Andrew Wilkinson reveals the hidden cost of using AI agents for automation. He spends the majority of his time debugging and improving them, with only a small fraction dedicated to actual productive output. This highlights the immaturity of current agent technology despite its power.

Related Insights

AI agents eliminate the physical work of typing and coding, but introduce a new form of burnout. The constraint on output is no longer time spent "doing," but the limited human capacity for high-stakes decision-making, context switching, and verification, which drains mental energy much faster.

Engineer productivity with AI agents hits a "valley of death" at medium autonomy. The tools excel at highly responsive, quick tasks (low autonomy) and fully delegated background jobs (high autonomy). The frustrating middle ground is where it's "not enough to delegate and not fun to wait," creating a key UX challenge.

AI is not a 'set and forget' solution. An agent's effectiveness directly correlates with the amount of time humans invest in training, iteration, and providing fresh context. Performance will ebb and flow with human oversight, with the best results coming from consistent, hands-on management.

The idea of an AI agent coding complex projects overnight often fails in practice. Real-world development is highly iterative, requiring constant feedback and design choices. This makes autonomous 'BuilderBots' less useful than interactive coding assistants for many common projects.

Some engineering teams use AI in a way that produces a high volume of code riddled with mistakes. This forces them to rewrite large portions, sometimes without AI assistance, ultimately slowing them down. The issue is not the tool, but the lack of best practices for its application.

AI performance on clean benchmarks overestimates real-world utility. In practice, tasks are "messy"—involving collaboration, large codebases, and adversarial situations—which current AIs handle poorly. This gap explains why productivity gains lag behind benchmark scores.

A Workday study reveals a critical blind spot in AI productivity metrics. While tools save time, roughly 37% of that saved time is offset by the need for rework—verifying information, correcting errors, and rewriting content. This dramatically reduces the net value and ROI of the technology.

Instead of leading to less work, agentic AI tools are causing users to work longer hours. The core reason is psychological: the tools are so effective at generating output that the opportunity cost of not working feels immense. This creates a hybrid of exhilaration and anxiety where time itself is the bottleneck.

While AI coding assistants appear to boost output, they introduce a "rework tax." A Stanford study found AI-generated code leads to significant downstream refactoring. A team might ship 40% more code, but if half of that increase is just fixing last week's AI-generated "slop," the real productivity gain is much lower than headlines suggest.

Developers using AI agents report unprecedented productivity but also a decline in job satisfaction. The creative act of writing code is replaced by the tedious task of reviewing vast amounts of AI-generated output, shifting their role to feel more like a middle manager of code.