Figma CEO Dylan Field predicts we will look back at current text prompting for AI as a primitive, command-line interface, similar to MS-DOS. The next major opportunity is to create intuitive, use-case-specific interfaces—like a compass for AI's latent space—that allow for more precise control beyond text.
Current text-based prompting for AI is a primitive, temporary phase, similar to MS-DOS. The future lies in more intuitive, constrained, and creative interfaces that allow for richer, more visual exploration of a model's latent space, moving beyond just natural language.
The primary interface for managing AI agents won't be simple chat, but sophisticated IDE-like environments for all knowledge workers. This paradigm of "macro delegation, micro-steering" will create new software categories like the "accountant IDE" or "lawyer IDE" for orchestrating complex AI work.
The early focus on crafting the perfect prompt is obsolete. Sophisticated AI interaction is now about 'context engineering': architecting the entire environment by providing models with the right tools, data, and retrieval mechanisms to guide their reasoning process effectively.
The best UI for an AI tool is a direct function of the underlying model's power. A more capable model unlocks more autonomous 'form factors.' For example, the sudden rise of CLI agents was only possible once models like Claude 3 became capable enough to reliably handle multi-step tasks.
Cues uses 'Visual Context Engineering' to let users communicate intent without complex text prompts. By using a 2D canvas for sketches, graphs, and spatial arrangements of objects, users can express relationships and structure visually, which the AI interprets for more precise outputs.
Open-ended prompts overwhelm new users who don't know what's possible. A better approach is to productize AI into specific features. Use familiar UI like sliders and dropdowns to gather user intent, which then constructs a complex prompt behind the scenes, making powerful AI accessible without requiring prompt engineering skills.
AI development has evolved to where models can be directed using human-like language. Instead of complex prompt engineering or fine-tuning, developers can provide instructions, documentation, and context in plain English to guide the AI's behavior, democratizing access to sophisticated outcomes.
With AI, designers are no longer just guessing user intent to build static interfaces. Their new primary role is to facilitate the interaction between a user and the AI model, helping users communicate their intent, understand the model's response, and build a trusted relationship with the system.
Chatbots are fundamentally linear, which is ill-suited for complex tasks like planning a trip. The next generation of AI products will use AI as a co-creation tool within a more flexible canvas-like interface, allowing users to manipulate and organize AI-generated content non-linearly.
AI tools that generate functional UIs from prompts are eliminating the 'language barrier' between marketing, design, and engineering teams. Marketers can now create visual prototypes of what they want instead of writing ambiguous text-based briefs, ensuring alignment and drastically reducing development cycles.