Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

The inference market is too large to remain monolithic. It will fragment into specialized platforms for different use cases like real-time video, long-running agents, or language models. This specialization will extend to hardware, with high-throughput, low-latency-need tasks (like agents) favoring cheaper AMD/Intel chips over NVIDIA's top GPUs.

Related Insights

The AI inference process involves two distinct phases: "prefill" (reading the prompt, which is compute-bound) and "decode" (writing the response, which is memory-bound). NVIDIA GPUs excel at prefill, while companies like Grok optimize for decode. The Grok-NVIDIA deal signals a future of specialized, complementary hardware rather than one-size-fits-all chips.

While focus is on massive supercomputers for training next-gen models, the real supply chain constraint will be 'inference' chips—the GPUs needed to run models for billions of users. As adoption goes mainstream, demand for everyday AI use will far outstrip the supply of available hardware.

While NVIDIA's CUDA software provides a powerful lock-in for AI training, its advantage is much weaker in the rapidly growing inference market. New platforms are demonstrating that developers can and will adopt alternative software stacks for deployment, challenging the notion of an insurmountable software moat.

While NVIDIA dominates the AI chip market, tech giants like Meta and Google are developing custom silicon (ASICs). As the market matures and workloads segment, these highly optimized, cost-effective chips could erode NVIDIA's market share for tasks that don't require cutting-edge general-purpose GPUs.

Just as developers use various databases for different needs, AI applications will rely on a "constellation" of specialized models. Some tasks will require expensive, high-reasoning models, while others will prioritize low-latency or low-cost models. The market will become heterogeneous, not monolithic.

While Nvidia dominates the AI training chip market, this only represents about 1% of the total compute workload. The other 99% is inference. Nvidia's risk is that competitors and customers' in-house chips will create cheaper, more efficient inference solutions, bifurcating the market and eroding its monopoly.

The intense power demands of AI inference will push data centers to adopt the "heterogeneous compute" model from mobile phones. Instead of a single GPU architecture, data centers will use disaggregated, specialized chips for different tasks to maximize power efficiency, creating a post-GPU era.

The rise of agent orchestration using specialized, open-source models will drive demand for custom ASICs. Jerry Murdock argues that putting a model on a dedicated chip will be far cheaper and more tunable for specific workloads than using expensive, general-purpose GPUs like Nvidia's, spurring a hardware shift.

While the most powerful AI will reside in large "god models" (like supercomputers), the majority of the market volume will come from smaller, specialized models. These will cascade down in size and cost, eventually being embedded in every device, much like microchips proliferated from mainframes.

The narrative of endless demand for NVIDIA's high-end GPUs is flawed. It will be cracked by two forces: the shift of AI inference to on-device flash memory, reducing cloud reliance, and Google's ability to give away its increasingly powerful Gemini AI for free, undercutting the revenue models that fuel GPU demand.