Waive integrates Vision-Language-Action models (VLAs) to create a conversational interface for the car. This allows users to talk to the AI chauffeur ("drive faster") and provides engineers with a powerful introspection tool to ask the system why it made a certain decision, demystifying its reasoning.
Instead of building its own capital-intensive robotaxi fleet, Waive's go-to-market strategy is to sell its autonomous driving stack to major auto manufacturers. This software-centric approach allows them to leverage the scale, distribution, and hardware infrastructure of established OEMs to reach millions of consumers.
The best UI for an AI tool is a direct function of the underlying model's power. A more capable model unlocks more autonomous 'form factors.' For example, the sudden rise of CLI agents was only possible once models like Claude 3 became capable enough to reliably handle multi-step tasks.
The AI's ability to handle novel situations isn't just an emergent property of scale. Waive actively trains "world models," which are internal generative simulators. This enables the AI to reason about what might happen next, leading to sophisticated behaviors like nudging into intersections or slowing in fog.
The evolution from simple voice assistants to 'omni intelligence' marks a critical shift where AI not only understands commands but can also take direct action through connected software and hardware. This capability, seen in new smart home and automotive applications, will embed intelligent automation into our physical environments.
The magic of ChatGPT's voice mode in a car is that it feels like another person in the conversation. Conversely, Meta's AI glasses failed when translating a menu because they acted like a screen reader, ignoring the human context of how people actually read menus. Context is everything for voice.
AI development has evolved to where models can be directed using human-like language. Instead of complex prompt engineering or fine-tuning, developers can provide instructions, documentation, and context in plain English to guide the AI's behavior, democratizing access to sophisticated outcomes.
Waive's core strategy is generalization. By training a single, large AI on diverse global data, vehicles, and sensor sets, they can adapt to new cars and countries in months, not years. This avoids the AV 1.0 pitfall of building bespoke, infrastructure-heavy solutions for each new market.
Waive treats the sensor debate as a distraction. Their goal is to build an AI flexible enough to work with any configuration—camera-only, camera-radar, or multi-sensor. This pragmatism allows them to adapt their software to different OEM partners and vehicle price points without being locked into a single hardware ideology.
A key learning from working with auto manufacturers is the desire for brand differentiation through driving personality. Waive can tailor its AI's behavior—from "helpfully assertive" to comfortably cautious—to match a brand's specific identity. This transforms the AI from a utility into a core part of the product experience.
Unlike older robots requiring precise maps and trajectory calculations, new robots use internet-scale common sense and learn motion by mimicking humans or simulations. This combination has “wiped the slate clean” for what is possible in the field.