Current self-driving technology cannot solve the complex, unpredictable situations human drivers navigate daily. This is not a problem that more data or better algorithms can fix, but a fundamental limitation. According to the 'Journey of the Mind' theory, full autonomy will only be possible when vehicles can incorporate the actual mechanism of consciousness.
The integration of AI into human-led services will mirror Tesla's approach to self-driving. Humans will remain the primary interface (the "steering wheel"), while AI progressively automates backend tasks, enhancing capability rather than eliminating the human role entirely in the near term.
After proving its robo-taxis are 90% safer than human drivers, Waymo is now making them more "confidently assertive" to better navigate real-world traffic. This counter-intuitive shift from passive safety to calculated aggression is a necessary step to improve efficiency and reduce delays, highlighting the trade-offs required for autonomous vehicle integration.
The popular conception of AGI as a pre-trained system that knows everything is flawed. A more realistic and powerful goal is an AI with a human-like ability for continual learning. This system wouldn't be deployed as a finished product, but as a 'super-intelligent 15-year-old' that learns and adapts to specific roles.
Today's AI models are powerful but lack a true sense of causality, leading to illogical errors. Unconventional AI's Naveen Rao hypothesizes that building AI on substrates with inherent time and dynamics—mimicking the physical world—is the key to developing this missing causal understanding.
To determine if an AI has subjective experience, one could analyze its internal belief manifold for multi-tiered, self-referential homeostatic loops. Pain and pleasure, for example, can be seen as second-order derivatives of a system's internal states—a model of its own model. This provides a technical test for being-ness beyond simple behavior.
A useful mental model for AGI is child development. Just as a child can be left unsupervised for progressively longer periods, AI agents are seeing their autonomous runtimes increase. AGI arrives when it becomes economically profitable to let an AI work continuously without supervision, much like an independent adult.
A critical weakness of current AI models is their inefficient learning process. They require exponentially more experience—sometimes 100,000 times more data than a human encounters in a lifetime—to acquire their skills. This highlights a key difference from human cognition and a major hurdle for developing more advanced, human-like AI.
Self-driving cars, a 20-year journey so far, are relatively simple robots: metal boxes on 2D surfaces designed *not* to touch things. General-purpose robots operate in complex 3D environments with the primary goal of *touching* and manipulating objects. This highlights the immense, often underestimated, physical and algorithmic challenges facing robotics.
While AI models excel at gathering and synthesizing information ('knowing'), they are not yet reliable at executing actions in the real world ('doing'). True agentic systems require bridging this gap by adding crucial layers of validation and human intervention to ensure tasks are performed correctly and safely.
The primary obstacle to creating a fully autonomous AI software engineer isn't just model intelligence but "controlling entropy." This refers to the challenge of preventing the compounding accumulation of small, 1% errors that eventually derail a complex, multi-step task and get the agent irretrievably off track.