The current GPU shortage is a temporary state. In any commodity-like market, a shortage creates a glut, and vice-versa. The immense profits generated by companies like NVIDIA are a "bat signal" for competition, ensuring massive future build-out and a subsequent drop in unit costs.
The massive capital expenditure by hyperscalers on AI will likely create an oversupply of capacity. This will crash prices, creating a golden opportunity for a new generation of companies to build innovative applications on cheap AI, much like Amazon utilized the cheap bandwidth left after the dot-com bust.
The AI infrastructure spending boom will continue robustly for at least two more years, creating a window where numerous chip startups can thrive in viable niches. While an eventual bubble pop and consolidation is guaranteed, the immediate future remains bright for even smaller players, challenging the winner-take-all narrative.
When power (watts) is the primary constraint for data centers, the total cost of compute becomes secondary. The crucial metric is performance-per-watt. This gives a massive pricing advantage to the most efficient chipmakers, as customers will pay anything for hardware that maximizes output from their limited power budget.
A single year of Nvidia's revenue is greater than the last 25 years of R&D and capex from the top five semiconductor equipment companies combined. This suggests a massive 'capex overhang,' meaning the primary bottleneck for AI compute isn't the ability to build fabs, but the financial arrangements to de-risk their construction.
While an AI bubble seems negative, the overproduction of compute power creates a favorable environment for companies that consume it. As prices for compute drop, their cost of goods sold decreases, leading to higher gross margins and better business fundamentals.
Hyperscalers face a strategic challenge: building massive data centers with current chips (e.g., H100) risks rapid depreciation as far more efficient chips (e.g., GB200) are imminent. This creates a 'pause' as they balance fulfilling current demand against future-proofing their costly infrastructure.
The comparison of the AI hardware buildout to the dot-com "dark fiber" bubble is flawed because there are no "dark GPUs"—all compute is being used. As hardware efficiency improves and token costs fall (Jevons paradox), it will unlock countless new AI applications, ensuring that demand continues to absorb all available supply.
The massive global investment required for AI will drive demand for GPUs so high that the annual market spend will exceed that of crude oil. This scale necessitates a dedicated futures market to allow participants, especially new cloud providers, to hedge price risk and lower their cost of capital.
The economic principle that 'shortages create gluts' is playing out in AI. The current scarcity of specialized talent and chips creates massive profit incentives for new supply to enter the market, which will eventually lead to an overcorrection and a future glut, as seen historically in the chip industry.
The narrative of endless demand for NVIDIA's high-end GPUs is flawed. It will be cracked by two forces: the shift of AI inference to on-device flash memory, reducing cloud reliance, and Google's ability to give away its increasingly powerful Gemini AI for free, undercutting the revenue models that fuel GPU demand.