The interface for physical machines is moving beyond buttons and touchscreens to multimodal interactions, primarily voice. This enables a "teaming" concept where a human operator collaborates with an AI agent, managing multiple machines and intervening only for critical decisions.
A common misconception is that simulation perfectly represents reality. In practice, it's a continuous loop: real-world data is required to tune simulator parameters, and this validation must be repeated until the gap between simulation and reality is small enough to trust the results.
While world models are powerful for understanding cause and effect, they are not a complete solution for deploying physical AI. Founders building real-world products must use a practical mix of technologies, as a pure world-model approach is too slow and expensive to be viable.
The physical AI industry is no longer in the fundamental research stage. It has entered a crucial "advanced engineering" phase between R&D and mass production. The focus is now on solving the subcomponent and reliability problems required to productionize existing technologies.
AI coding assistants are creating a massive productivity gap among engineers. This leads to a bimodal distribution where one group fully leverages the tools and becomes massively effective, while another falls far behind. Hiring must now select for this new skillset.
A core challenge in physical AI is the tension between large, powerful models (offboard, in a data center) and the need for low-latency models (onboard, on the machine). The key is using techniques like distillation to create smaller derivatives that run in milliseconds for safety-critical decisions.
In autonomous systems, LIDAR is invaluable during R&D to provide per-pixel depth data. This data trains models so that cheaper, camera-only production vehicles can accurately infer depth. This makes LIDAR a temporary means to an end, not the final sensor suite.
Unlike pure software, the value in physical AI and hard tech comes from long-term compounding of technology. Startups often fail because they don't survive long enough to see these returns. This makes early commercial discipline and constraints crucial for longevity.
To keep pace with rapid AI advancements, the company intentionally operates on a two-year horizon for its technology stack. This forces them to be dynamic and adapt to new research, rather than getting locked into outdated architectures, having completed four such evolutions so far.
Traditional vehicle safety (e.g., Euro NCAP) used a checklist of specific test cases with binary pass/fail answers. For AI systems, this is insufficient. The new paradigm is statistical validation, where the goal is to prove reliability to a certain number of "nines" across a vast range of scenarios.
While previously underwhelming, the latest generation of AI models are now surprisingly effective at highly specialized, low-level coding tasks such as writing GPU shaders. This shows that the "bitter lesson"—that general models scaling beats specialized approaches—applies even in embedded and systems programming.
Mirroring Google's Android strategy for mobile, Applied Intuition created a specialized OS to run AI across diverse hardware. This layer solves for safety-critical needs like real-time control, memory management, and reliable updates, which were previously impossible due to fragmentation across manufacturers.
