/
© 2026 RiffOn. All rights reserved.
  1. Latent Space: The AI Engineer Podcast
  2. After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs
After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs

Latent Space: The AI Engineer Podcast · Nov 25, 2025

World Labs founders Fei Fei & Justin explore spatial intelligence, the evolution of AI research, and their vision for world models with Marble.

Transformers Are Fundamentally Models of Sets, Not Sequences

Contrary to common perception shaped by their use in language, Transformers are not inherently sequential. Their core architecture operates on sets of tokens, with sequence information only injected via positional embeddings. This makes them powerful for non-sequential data like 3D objects or other unordered collections.

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs thumbnail

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs

Latent Space: The AI Engineer Podcast·3 months ago

"Pixel Maximalism" Argues Pixels Are a More Lossless World Representation Than Text

This idea posits that language is a lossy, discrete abstraction of reality. In contrast, pixels (visual input) are a more fundamental representation. We perceive language physically—as pixels on a page or sound waves—and tokenizing it discards rich information like font, layout, and visual context.

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs thumbnail

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs

Latent Space: The AI Engineer Podcast·3 months ago

GPU Scaling Limits May Force AI Architectures Beyond Transformers

The plateauing performance-per-watt of GPUs suggests that simply scaling current matrix multiplication-heavy architectures is unsustainable. This hardware limitation may necessitate research into new computational primitives and neural network designs built for large-scale distributed systems, not single devices.

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs thumbnail

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs

Latent Space: The AI Engineer Podcast·3 months ago

AI Needs "Spatial Intelligence" Because Language Is a Lossy Abstraction of Reality

World Labs argues that AI focused on language misses the fundamental "spatial intelligence" humans use to interact with the 3D world. This capability, which evolved over hundreds of millions of years, is crucial for true understanding and cannot be fully captured by 1D text, a lossy representation of physical reality.

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs thumbnail

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs

Latent Space: The AI Engineer Podcast·3 months ago

Massive VC Funding Creates a Resource Imbalance Threatening Academic AI Research

Fei-Fei Li expresses concern that the influx of commercial capital into AI isn't just creating pressure, but an "imbalanced resourcing" of academia. This starves universities of the compute and talent needed to pursue open, foundational science, potentially stifling the next wave of innovation that commercial labs build upon.

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs thumbnail

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs

Latent Space: The AI Engineer Podcast·3 months ago

Generative AI Lacks Causal Understanding, Limiting Its Use in High-Stakes Fields

While a world model can generate a physically plausible arch, it doesn't understand the underlying physics of force distribution. This gap between pattern matching and causal reasoning is a fundamental split between AI and human intelligence, making current models unsuitable for mission-critical applications like architecture.

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs thumbnail

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs

Latent Space: The AI Engineer Podcast·3 months ago

Academia's AI Role Has Shifted from Scaling Models to Exploring "Wacky" Foundational Ideas

With industry dominating large-scale compute, academia's function is no longer to train the biggest models. Instead, its value lies in pursuing unconventional, high-risk research in areas like new algorithms, architectures, and theoretical underpinnings that commercial labs, focused on scaling, might overlook.

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs thumbnail

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs

Latent Space: The AI Engineer Podcast·3 months ago

World Labs' Marble Uses Gaussian Splats as an Atomic Unit for Real-Time 3D Worlds

Unlike video models that generate frame-by-frame, Marble natively outputs Gaussian splats—tiny, semi-transparent particles. This data structure enables real-time rendering, interactive editing, and precise camera control on client devices like mobile phones, a fundamental architectural advantage for interactive 3D experiences.

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs thumbnail

After LLMs: Spatial Intelligence and World Models — Fei-Fei Li & Justin Johnson, World Labs

Latent Space: The AI Engineer Podcast·3 months ago