The next frontier for conversational AI is not just better text, but "Generative UI"—the ability to respond with interactive components. Instead of describing the weather, an AI can present a weather widget, merging the flexibility of chat with the richness of a graphical interface.
Figma CEO Dylan Field predicts we will look back at current text prompting for AI as a primitive, command-line interface, similar to MS-DOS. The next major opportunity is to create intuitive, use-case-specific interfaces—like a compass for AI's latent space—that allow for more precise control beyond text.
Current text-based prompting for AI is a primitive, temporary phase, similar to MS-DOS. The future lies in more intuitive, constrained, and creative interfaces that allow for richer, more visual exploration of a model's latent space, moving beyond just natural language.
For data-heavy queries like financial projections, AI responses should transcend static text. The ideal output is an interactive visualization, such as a chart or graph, that the user can directly manipulate. This empowers them to explore scenarios and gain a deeper understanding of the data.
The best UI for an AI tool is a direct function of the underlying model's power. A more capable model unlocks more autonomous 'form factors.' For example, the sudden rise of CLI agents was only possible once models like Claude 3 became capable enough to reliably handle multi-step tasks.
The best agentic UX isn't a generic chat overlay. Instead, identify where users struggle with complex inputs like formulas or code. Replace these friction points with a native, natural language interface that directly integrates the AI into the core product workflow, making it feel seamless and powerful.
Open-ended prompts overwhelm new users who don't know what's possible. A better approach is to productize AI into specific features. Use familiar UI like sliders and dropdowns to gather user intent, which then constructs a complex prompt behind the scenes, making powerful AI accessible without requiring prompt engineering skills.
AI will fundamentally change user interfaces. Instead of designers pre-building UIs, AI will generate the necessary "forms and lists" on the fly based on a user's natural language request. This means for the first time, the user, not the developer, will be the one creating the interface.
As AI models become proficient at generating high-quality UI from prompts, the value of manual design execution will diminish. A professional designer's key differentiator will become their ability to build the underlying, unique component libraries and design systems that AI will use to create those UIs.
With AI, designers are no longer just guessing user intent to build static interfaces. Their new primary role is to facilitate the interaction between a user and the AI model, helping users communicate their intent, understand the model's response, and build a trusted relationship with the system.
Chatbots are fundamentally linear, which is ill-suited for complex tasks like planning a trip. The next generation of AI products will use AI as a co-creation tool within a more flexible canvas-like interface, allowing users to manipulate and organize AI-generated content non-linearly.