Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

While NVIDIA's GPUs have been the primary AI constraint, the bottleneck is now moving to other essential subsystems. Memory, networking interconnects, and power management are emerging as the next critical choke points, signaling a new wave of investment opportunities in the hardware stack beyond core compute.

Related Insights

The AI supply chain is crunched not just by obvious components like TSMC wafers and HBM memory. A significant, often overlooked bottleneck is rack manufacturing—including high-speed cables, connectors, and even sheet metal—which are "sneaky hard" due to extreme power, heat, and signal integrity demands.

The growth of AI is constrained not by chip design but by inputs like energy and High Bandwidth Memory (HBM). This shifts power to component suppliers and energy providers, allowing them to gain leverage, demand equity, and influence the entire AI ecosystem, much like a central bank controls money.

The primary bottleneck for scaling AI over the next decade may be the difficulty of bringing gigawatt-scale power online to support data centers. Smart money is already focused on this challenge, which is more complex than silicon supply.

The AI industry's growth constraint is a swinging pendulum. While power and data center space are the current bottlenecks (2024-25), the energy supply chain is diverse. By 2027, the bottleneck will revert to semiconductor manufacturing, as leading-edge fab capacity (e.g., TSMC, HBM memory) is highly concentrated and takes years to expand.

The focus in AI has evolved from rapid software capability gains to the physical constraints of its adoption. The demand for compute power is expected to significantly outstrip supply, making infrastructure—not algorithms—the defining bottleneck for future growth.

The exponential growth in AI required moving beyond single GPUs. Mellanox's interconnect technology was critical for scaling to thousands of GPUs, effectively turning the entire data center into a single, high-performance computer and solving the post-Moore's Law scaling challenge.

While the world focused on GPU shortages, the real constraint on AI compute is now physical infrastructure. The bottleneck has moved to accessing power, building data centers, and finding specialized labor like electricians and acquiring basic materials like structural steel. Merely acquiring chips is no longer enough to scale.

The intense power demands of AI inference will push data centers to adopt the "heterogeneous compute" model from mobile phones. Instead of a single GPU architecture, data centers will use disaggregated, specialized chips for different tasks to maximize power efficiency, creating a post-GPU era.

Satya Nadella clarifies that the primary constraint on scaling AI compute is not the availability of GPUs, but the lack of power and physical data center infrastructure ("warm shelves") to install them. This highlights a critical, often overlooked dependency in the AI race: energy and real estate development speed.

Even if NVIDIA and TSMC solve wafer shortages, the AI industry faces a looming energy (watt) bottleneck. The inability to power new data centers could cap AI growth, shifting the primary constraint from semiconductor manufacturing to energy infrastructure and supply.

AI's Next Bottleneck Is Shifting From GPUs to Memory, Networking, and Power | RiffOn