We scan new podcasts and send you the top 5 insights daily.
Andreessen highlights a unique economic phenomenon: the pace of AI software improvement outstrips hardware depreciation. This means a three-year-old NVIDIA inference chip can generate more revenue today than when it was new, a complete reversal of typical tech hardware value cycles.
Unlike typical computer hardware that depreciates rapidly, H100 GPUs are trading above their launch price in secondary markets. This market anomaly, driven by the extreme and sustained compute shortage for AI, completely inverts traditional financial models for hardware assets.
AI software is improving so rapidly that older hardware, like a three-year-old NVIDIA inference chip, is now more profitable than it was when new. This phenomenon, where software advancements outpace hardware depreciation, is unprecedented and makes existing infrastructure increasingly valuable.
The current AI moment is unique because demand outstrips supply so dramatically that even previous-generation chips and models remain valuable. They are perfectly suited for running smaller models for simpler, high-volume applications like voice transcription, creating a broad-based boom across the entire hardware and model stack.
Despite the rapid pace of hardware innovation, the value of older NVIDIA GPUs like the H100 is holding strong. Cloud provider CoreWeave reports these chips are retaining 90-95% of their pricing power over a 5-6 year lifespan because compute demand far outstrips supply.
Contrary to typical hardware depreciation, GPUs like NVIDIA's H100 are becoming more valuable over time. This is because newer, more efficient AI models can generate significantly more output and value on the same hardware, tying the GPU's worth to its utility rather than its age.
While the industry standard is a six-year depreciation for data center hardware, analyst Dylan Patel warns this is risky for GPUs. Rapid annual performance gains from new models could render older chips economically useless long before they physically fail.
Contrary to the belief that AI chips quickly become obsolete, CoreWeave's CEO argues their value holds, citing average five-year client contracts as proof. Older chips like the A100 have even appreciated in price as new use cases emerge, making rapid depreciation a myth.
Hyperscalers are extending depreciation schedules for AI hardware. While this may look like "cooking the books" to inflate earnings, it's justified by the reality that even 7-8 year old TPUs and GPUs are still running at 100% utilization for less complex AI tasks, making them valuable for longer and validating the accounting change.
The useful life of an AI chip isn't a fixed period. It ends only when a new generation offers such a significant performance and efficiency boost that it becomes more economical to replace fully paid-off, older hardware. Slower generational improvements mean longer depreciation cycles.
Countering the narrative of rapid burnout, CoreWeave cites historical data showing a nearly 10-year service life for older NVIDIA GPUs (K80) in major clouds. Older chips remain valuable for less intensive tasks, creating a tiered system where new chips handle frontier models and older ones serve established workloads.