A developer combined Meta Ray-Ban glasses with a two-layer AI system. The first layer, Google's Gemini Live, handles real-time perception (vision and voice). It then delegates specific tasks to a second layer, OpenClaw, for execution and browser automation. This architecture effectively separates perception from action.
AI devices must be close to human senses to be effective. Glasses are the most natural form factor as they capture sight, sound, and are close to the mouth for speech. This sensory proximity gives them an advantage over other wearables like earbuds or pins.
The distinction between a "model" and an "agent" is dissolving. Google's new Interactions API provides a single interface for both, signaling a future where flagship releases are complete systems out-of-the-box, capable of both simple queries and complex, long-running tasks, blurring the lines for developers and users.
Demis Hassabis suggests that previous attempts at smart glasses like Google Glass were too early because they lacked a compelling use case. He believes a hands-free, always-on AI assistant like Project Astra provides the 'killer app' that will finally make smart glasses a mainstream consumer device.
As AI moves into collaborative 'multiplayer mode,' its user interface will evolve into a command center. This UI will explicitly separate tasks agents can execute autonomously from those requiring human intervention, which are flagged for review. This shifts the user's role from performing tasks to overseeing and approving AI's work.
Demis Hassabis claims previous smart glasses failed not just due to clunky hardware but because they lacked a compelling use case. He argues that a powerful, seamless AI assistant, integrated into daily life, is the "killer app" that will finally drive adoption for this form factor.
A new software paradigm, "agent-native architecture," treats AI as a core component, not an add-on. This progresses in levels: the agent can do any UI action, trigger any backend code, and finally, perform any developer task like writing and deploying new code, enabling user-driven app customization.
The next user interface paradigm is delegation, not direct manipulation. Humans will communicate with AI agents via voice, instructing them to perform complex tasks on computers. This will shift daily work from hours of clicking and typing to zero, fundamentally changing our relationship with technology.
While AI models excel at gathering and synthesizing information ('knowing'), they are not yet reliable at executing actions in the real world ('doing'). True agentic systems require bridging this gap by adding crucial layers of validation and human intervention to ensure tasks are performed correctly and safely.
Past smart glasses failed not because of the hardware, but the lack of a compelling use case. Hassabis argues a universal, context-aware digital assistant that works seamlessly across all devices is the true 'killer app' that will finally make wearables like smart glasses indispensable.
AI accelerates AR glasses adoption not by improving the display, but by changing how we compute. As AI agents operate software, our role shifts to monitoring, making a portable, multi-screen AR workstation more useful than a single-task phone.