We scan new podcasts and send you the top 5 insights daily.
In a significant strategic misstep, Google sold a large volume of its custom TPU accelerators to rival Anthropic. Immediately after, demand for Google's own Gemini model surged, leaving Google compute-constrained and trying to secure more capacity from a sold-out TSMC.
Google's strategy isn't just to sell AI chips; it's a platform play. By offering its powerful and potentially cheaper TPUs to companies, Google can create a powerful incentive for those customers to run their entire AI workloads on Google Cloud, creating a sticky, integrated ecosystem that challenges AWS and Azure.
Google is offering its TPUs externally for the first time as a strategic move to gain market share while it has a temporary hardware advantage over Nvidia. This classic tactic aims to build a crucial install base that can be upgraded later, even after its competitive performance edge inevitably narrows.
Anthropic is pioneering a new hardware strategy. Instead of just renting Tensor Processing Units (TPUs) from Google Cloud, it is buying the chips directly from co-designer Broadcom. This gives Anthropic more control over its infrastructure, a significant move away from the standard cloud-centric model for AI companies.
AI labs like Anthropic that were conservative in securing long-term compute now face a 'quality tax.' They must resort to lower-quality providers or pay significant markups and revenue-sharing deals for last-minute capacity, a cost their more aggressive competitors like OpenAI avoided by signing deals early.
Google successfully trained its top model, Gemini 3 Pro, on its own TPUs, proving a viable alternative to NVIDIA's chips. However, because Google doesn't sell these TPUs, NVIDIA retains its monopoly pricing power over every other company in the market.
Anthropic's choice to purchase Google's TPUs via Broadcom, rather than directly or by designing its own chips, indicates a new phase in the AI hardware market. It highlights the rise of specialized manufacturers as key suppliers, creating a more complex and diversified hardware ecosystem beyond just Nvidia and the major AI labs.
This theory suggests Google's refusal to sell TPUs is a strategic move to maintain a high market price for AI inference. By allowing NVIDIA's expensive GPUs to set the benchmark, Google can profit from its own lower-cost TPU-based inference services on GCP.
Google created its custom TPU chip not as a long-term strategy, but from an internal crisis. Engineer Jeff Dean calculated that scaling a new speech recognition feature to all Android phones would require doubling Google's entire data center footprint, forcing the company to design a more efficient, custom chip to avoid existential costs.
The narrative of NVIDIA's untouchable dominance is undermined by a critical fact: the world's leading models, including Google's Gemini 3 and Anthropic's Claude 4.5, are primarily trained on Google's TPUs and Amazon's Tranium chips. This proves that viable, high-performance alternatives already exist at the highest level of AI development.
While competitors like OpenAI must buy GPUs from NVIDIA, Google trains its frontier AI models (like Gemini) on its own custom Tensor Processing Units (TPUs). This vertical integration gives Google a significant, often overlooked, strategic advantage in cost, efficiency, and long-term innovation in the AI race.