Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

Google's cloud division (GCP), incentivized to sell compute, is allocating scarce TPU chips to external customer Anthropic. This directly constrains Google's own AI lab, Gemini, hindering its progress in the hyper-competitive AI race and revealing significant internal friction between business units with conflicting goals.

Related Insights

Google's strategy isn't just to sell AI chips; it's a platform play. By offering its powerful and potentially cheaper TPUs to companies, Google can create a powerful incentive for those customers to run their entire AI workloads on Google Cloud, creating a sticky, integrated ecosystem that challenges AWS and Azure.

Google is offering its TPUs externally for the first time as a strategic move to gain market share while it has a temporary hardware advantage over Nvidia. This classic tactic aims to build a crucial install base that can be upgraded later, even after its competitive performance edge inevitably narrows.

In a significant strategic misstep, Google sold a large volume of its custom TPU accelerators to rival Anthropic. Immediately after, demand for Google's own Gemini model surged, leaving Google compute-constrained and trying to secure more capacity from a sold-out TSMC.

Anthropic is pioneering a new hardware strategy. Instead of just renting Tensor Processing Units (TPUs) from Google Cloud, it is buying the chips directly from co-designer Broadcom. This gives Anthropic more control over its infrastructure, a significant move away from the standard cloud-centric model for AI companies.

Large tech companies are buying up compute from smaller cloud providers not for immediate need, but as a defensive strategy. By hoarding scarce GPU capacity, they prevent competitors from accessing critical resources, effectively cornering the market and stifling innovation from rivals.

For leading AI labs like Anthropic and OpenAI, the primary value from cloud partnerships isn't a sales channel but guaranteed access to scarce compute and GPUs. This turns negotiations into a complex, symbiotic bundle covering hardware access, cloud credits, and revenue sharing, where hardware is the most critical component.

As the current low-cost producer of AI tokens via its custom TPUs, Google's rational strategy is to operate at low or even negative margins. This "sucks the economic oxygen out of the AI ecosystem," making it difficult for capital-dependent competitors to justify their high costs and raise new funding rounds.

AI company Anthropic's potential multi-billion dollar compute deal with Google over AWS is a major strategic indicator. It suggests AWS's AI infrastructure is falling behind, and losing a cornerstone AI customer like Anthropic could mean its entire AI strategy is 'cooked,' signaling a shift in the cloud platform wars.

Sundar Pichai notes an ironic consequence of the AI boom: the scarcity of TPUs forces a more disciplined capital allocation process. Since all major projects, including Waymo, now compete for the same limited compute resources, the trade-offs are more explicit and front-of-mind than ever before.

While competitors like OpenAI must buy GPUs from NVIDIA, Google trains its frontier AI models (like Gemini) on its own custom Tensor Processing Units (TPUs). This vertical integration gives Google a significant, often overlooked, strategic advantage in cost, efficiency, and long-term innovation in the AI race.