Chatbots are fundamentally linear, which is ill-suited for complex tasks like planning a trip. The next generation of AI products will use AI as a co-creation tool within a more flexible canvas-like interface, allowing users to manipulate and organize AI-generated content non-linearly.
Current text-based prompting for AI is a primitive, temporary phase, similar to MS-DOS. The future lies in more intuitive, constrained, and creative interfaces that allow for richer, more visual exploration of a model's latent space, moving beyond just natural language.
The primary interface for managing AI agents won't be simple chat, but sophisticated IDE-like environments for all knowledge workers. This paradigm of "macro delegation, micro-steering" will create new software categories like the "accountant IDE" or "lawyer IDE" for orchestrating complex AI work.
The best UI for an AI tool is a direct function of the underlying model's power. A more capable model unlocks more autonomous 'form factors.' For example, the sudden rise of CLI agents was only possible once models like Claude 3 became capable enough to reliably handle multi-step tasks.
The best agentic UX isn't a generic chat overlay. Instead, identify where users struggle with complex inputs like formulas or code. Replace these friction points with a native, natural language interface that directly integrates the AI into the core product workflow, making it feel seamless and powerful.
Cues uses 'Visual Context Engineering' to let users communicate intent without complex text prompts. By using a 2D canvas for sketches, graphs, and spatial arrangements of objects, users can express relationships and structure visually, which the AI interprets for more precise outputs.
Open-ended prompts overwhelm new users who don't know what's possible. A better approach is to productize AI into specific features. Use familiar UI like sliders and dropdowns to gather user intent, which then constructs a complex prompt behind the scenes, making powerful AI accessible without requiring prompt engineering skills.
Agentic AI will evolve into a 'multi-agent ecosystem.' This means AI agents from different companies—like an airline and a hotel—will interact directly with each other to autonomously solve a customer's complex problem, freeing humans from multi-party coordination tasks.
With AI, designers are no longer just guessing user intent to build static interfaces. Their new primary role is to facilitate the interaction between a user and the AI model, helping users communicate their intent, understand the model's response, and build a trusted relationship with the system.
While chat works for human-AI interaction, the infinite canvas is a superior paradigm for multi-agent and human-AI collaboration. It allows for simultaneous, non-distracting parallel work, asynchronous handoffs, and persistent spatial context—all of which are difficult to achieve in a linear, turn-based chat interface.
The promise of AI shouldn't be a one-click solution that removes the user. Instead, AI should be a collaborative partner that augments human capacity. A successful AI product leaves room for user participation, making them feel like they are co-building the experience and have a stake in the outcome.