Developers fall into the "agentic trap" by building complex, fully-automated AI coding systems. These systems fail to create good products because they lack human taste and the iterative feedback loop where a creator's vision evolves through interaction with the software being built.
The trend of 'vibe coding'—casually using prompts to generate code without rigor—is creating low-quality, unmaintainable software. The AI engineering community has reached its limit with this approach and is actively searching for a new development paradigm that marries AI's speed with traditional engineering's craft and reliability.
Exploratory AI coding, or 'vibe coding,' proved catastrophic for production environments. The most effective developers adapted by treating AI like a junior engineer, providing lightweight specifications, tests, and guardrails to ensure the output was viable and reliable.
Karpathy found AI coding agents struggle with genuinely novel projects like his NanoChat repository. Their training on common internet patterns causes them to misunderstand custom implementations and try to force standard, but incorrect, solutions. They are good for autocomplete and boilerplate but not for intellectually intense, frontier work.
Product leaders must personally engage with AI development. Direct experience reveals unique, non-human failure modes. Unlike a human developer who learns from mistakes, an AI can cheerfully and repeatedly make the same error—a critical insight for managing AI projects and team workflow.
True creative mastery emerges from an unpredictable human process. AI can generate options quickly but bypasses this journey, losing the potential for inexplicable, last-minute genius that defines truly great work. It optimizes for speed at the cost of brilliance.
When AI can generate code and designs endlessly, creating "AI slop," the critical human contribution becomes judgment. The key challenge shifts from *building* to *deciding what to build* and evaluating the output's quality and security. The question is no longer "can we build it?" but "should we build it?"
A proactive AI feature at OpenAI that automatically revised PRs based on human feedback was unpopular. Unlike assistive tools, fully automated loops face an extremely high bar for quality, and the feature's "hit rate" wasn't high enough to be worth the cognitive overhead.
While developers leverage multiple AI agents to achieve massive productivity gains, this velocity can create incomprehensible and tightly coupled software architectures. The antidote is not less AI but more human-led structure, including modularity, rapid feedback loops, and clear specifications.
After achieving broad adoption of agentic coding, the new challenge becomes managing the downsides. Increased code generation leads to lower quality, rushed reviews, and a knowledge gap as team members struggle to keep up with the rapidly changing codebase.
Non-technical creators using AI coding tools often fail due to unrealistic expectations of instant success. The key is a mindset shift: understanding that building quality software is an iterative process of prompting, testing, and debugging, not a one-shot command that works in five prompts.