Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

Periodic Labs' co-founder states their work was not possible with the AI of late 2022. Advances in model reasoning, reliable tool use, and error correction over the subsequent years were foundational technologies necessary to connect AI systems to the physical world.

Related Insights

A genuine AI capabilities explosion won't happen just because models can write novel research papers. The bottleneck is the full automation of the R&D loop, which includes a long tail of "messy" real-world tasks like fixing failing GPUs in a data center or managing facility cooling. This physical and logistical grounding is often overlooked.

For vertical AI applications, foundation models are now sufficiently intelligent. The primary challenge is no longer model capability but building the surrounding software infrastructure—tools, UIs, and workflows—that lets models perform useful work reliably and trustworthily.

While language models understand the world through text, Demis Hassabis argues they lack an intuitive grasp of physics and spatial dynamics. He sees 'world models'—simulations that understand cause and effect in the physical world—as the critical technology needed to advance AI from digital tasks to effective robotics.

Startups and major labs are focusing on "world models," which simulate physical reality, cause, and effect. This is seen as the necessary step beyond text-based LLMs to create agents that can truly understand and interact with the physical world, a key step towards AGI.

Today's AI models are powerful but lack a true sense of causality, leading to illogical errors. Unconventional AI's Naveen Rao hypothesizes that building AI on substrates with inherent time and dynamics—mimicking the physical world—is the key to developing this missing causal understanding.

Unlike pre-programmed industrial robots, "Physical AI" systems sense their environment, make intelligent choices, and receive live feedback. This paradigm shift, similar to Waymo's self-driving cars versus simple cruise control, allows for autonomous and adaptive scientific experimentation rather than just repetitive tasks.

Ken Goldberg quantifies the challenge: the text data used to train LLMs would take a human 100,000 years to read. Equivalent data for robot manipulation (vision-to-control signals) doesn't exist online and must be generated from scratch, explaining the slower progress in physical AI.

In the rapidly advancing field of AI, building products around current model limitations is a losing strategy. The most successful AI startups anticipate the trajectory of model improvements, creating experiences that seem 80% complete today but become magical once future models unlock their full potential.

The perceived limits of today's AI are not inherent to the models themselves but to our failure to build the right "agentic scaffold" around them. There's a "model capability overhang" where much more potential can be unlocked with better prompting, context engineering, and tool integrations.

The founder of AI and robotics firm Medra argues that scientific progress is not limited by a lack of ideas or AI-generated hypotheses. Instead, the critical constraint is the physical capacity to test these ideas and generate high-quality data to train better AI models.