Engineer productivity with AI agents hits a "valley of death" at medium autonomy. The tools excel at highly responsive, quick tasks (low autonomy) and fully delegated background jobs (high autonomy). The frustrating middle ground is where it's "not enough to delegate and not fun to wait," creating a key UX challenge.
Once AI coding agents reach a high performance level, objective benchmarks become less important than a developer's subjective experience. Like a warrior choosing a sword, the best tool is often the one that has the right "feel," writes code in a preferred style, and integrates seamlessly into a human workflow.
Treating AI coding tools like an asynchronous junior engineer, rather than a synchronous pair programmer, sets correct expectations. This allows users to delegate tasks, go to meetings, and check in later, enabling true multi-threading of work without the need to babysit the tool.
AI's impact on coding is unfolding in stages. Phase 1 was autocomplete (Copilot). We're now in Phase 2, defined by interactive agents where developers orchestrate tasks with prompts. Phase 3 will be true automation, where agents independently handle complete, albeit simpler, development workflows without direct human guidance.
The most significant productivity gains come from applying AI to every stage of development, including research, planning, product marketing, and status updates. Limiting AI to just code generation misses the larger opportunity to automate the entire engineering process.
Frame AI agent development like training an intern. Initially, they need clear instructions, access to tools, and your specific systems. They won't be perfect at first, but with iterative feedback and training ('progress over perfection'), they can evolve to handle complex tasks autonomously.
Tools like OpenAI's Codex can complete hours of coding in minutes following a design phase. This creates awkward, inefficient downtime periods for the developer, fundamentally altering the daily work rhythm from a steady flow to unproductive cycles of intense work followed by waiting.
The process of struggling with and solving hard problems is what builds engineering skill. Constantly available AI assistants act like a "slot machine for answers," removing this productive struggle. This encourages "vibe coding" and may prevent engineers from developing deep problem-solving expertise.
The ideal AI-powered engineering workflow isn't just one tool, but a fluid cycle. It involves synchronous collaboration with an AI for planning and review, then handing off to an asynchronous agent for implementation and testing, before returning to synchronous mode for the next phase.
A Meta study found expert programmers were less productive with AI tools. The speaker suggests this is because users thought they were faster while actually being distracted (e.g., social media) waiting for the AI, highlighting a dangerous gap between perceived and actual productivity.
Cursor's founder predicts AI developer tools will bifurcate into two modes: a fast, "in-the-loop" copilot for pair programming, and a slower, asynchronous "agent" that completes entire tasks with perfect accuracy. This requires building products optimized for both speed and correctness.