True co-design between AI models and chips is currently impossible due to an "asymmetric design cycle." AI models evolve much faster than chips can be designed. By using AI to drastically speed up chip design, it becomes possible to create a virtuous cycle of co-evolution.

Related Insights

To achieve 1000x efficiency, Unconventional AI is abandoning the digital abstraction (bits representing numbers) that has defined computing for 80 years. Instead, they are co-designing hardware and algorithms where the physics of the substrate itself defines the neural network, much like a biological brain.

New AI models are designed to perform well on available, dominant hardware like NVIDIA's GPUs. This creates a self-reinforcing cycle where the incumbent hardware dictates which model architectures succeed, making it difficult for superior but incompatible chip designs to gain traction.

Software companies struggle to build their own chips because their agile, sprint-based culture clashes with hardware development's demands. Chip design requires a "measure twice, cut once" mentality, as mistakes cost months and millions. This cultural mismatch is a primary reason for failure, even with immense resources.

Recursive Intelligence's AI develops unconventional, curved chip layouts that human designers considered too complex or risky. These "alien" designs optimize for power and speed by reducing wire lengths, demonstrating AI's ability to explore non-intuitive solution spaces beyond human creativity.

Musk states that designing the custom AI5 and AI6 chips is his 'biggest time allocation.' This focus on silicon, promising a 40x performance increase, reveals that Tesla's core strategy relies on vertically integrated hardware to solve autonomy and robotics, not just software.

Designing a chip is not a monolithic problem that a single AI model like an LLM can solve. It requires a hybrid approach. While LLMs excel at language and code-related stages, other components like physical layout are large-scale optimization problems best solved by specialized graph-based reinforcement learning agents.

Just as TSMC enabled "fabless" giants like NVIDIA, Recursive Intelligence envisions a "designless" paradigm. They aim to provide AI-driven chip design as a service, allowing companies to procure custom silicon without the massive overhead of hiring and managing large, specialized hardware engineering teams.

NVIDIA's commitment to programmable GPUs over fixed-function ASICs (like a "transformer chip") is a strategic bet on rapid AI innovation. Since models are evolving so quickly (e.g., hybrid SSM-transformers), a flexible architecture is necessary to capture future algorithmic breakthroughs.

OpenAI is designing its custom chip for flexibility, not just raw performance on current models. The team learned that major 100x efficiency gains come from evolving algorithms (e.g., dense to sparse transformers), so the hardware must be adaptable to these future architectural changes.

The current 2-3 year chip design cycle is a major bottleneck for AI progress, as hardware is always chasing outdated software needs. By using AI to slash this timeline, companies can enable a massive expansion of custom chips, optimizing performance for many at-scale software workloads.

Slow Chip Design Cycles Are the Primary Barrier to AI Hardware/Software Co-Design | RiffOn