Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

The next wave of AI compute demand won't be from generating more outputs, but from agents performing exponentially more data collection for a single task. For example, a financial model could trigger an agent to analyze vast datasets, like satellite imagery, multiplying token usage for one result.

Related Insights

The shift from simple chatbots (one user request, one API call) to agentic AI systems will decouple inference requests from direct user actions. A single user request could trigger hundreds or thousands of automated model calls, leading to an exponential increase in compute demand and cost.

A paradox exists where the cost for a fixed level of AI capability (e.g., GPT-4 level) has dropped 100-1000x. However, overall enterprise spend is increasing because applications now use frontier models with massive contexts and multi-step agentic workflows, creating huge multipliers on token usage that drive up total costs.

While the growth of new consumer AI users is slowing into an S-curve, the compute consumption per user is still growing exponentially. This is driven by the shift from simple queries to complex, token-intensive tasks like reasoning and agents, sustaining massive demand for GPU infrastructure.

Ben Thompson argues the shift from simple chatbots to AI agents creates an exponential, non-speculative demand for compute. Agents automate complex, multi-step tasks, driving constant usage that justifies the massive capex investments by hyperscalers. This suggests the current spending is based on real demand, not bubble-fueled speculation.

The current AI data center arms race isn't about meeting today's demand for chatbots. It's fueled by companies like Meta betting on a future where personal AI agents run constantly, analyzing every interaction. This vision of persistent, parallel agents requires an exponential increase in compute, explaining why they will buy any available capacity.

While the cost to achieve a fixed capability level (e.g., GPT-4 at launch) has dropped over 100x, overall enterprise spending is increasing. This paradox is explained by powerful multipliers: demand for frontier models, longer reasoning chains, and multi-step agentic workflows that consume exponentially more tokens.

While user growth for apps like ChatGPT is slowing, per-user token consumption is skyrocketing as models shift from simple queries to complex reasoning and AI agents. This creates a hidden, exponential growth in compute demand, validating Oracle's massive infrastructure investment even as front-end adoption matures.

While the cost for GPT-4 level intelligence has dropped over 100x, total enterprise AI spend is rising. This is driven by multipliers: using larger frontier models for harder tasks, reasoning-heavy workflows that consume more tokens, and complex, multi-turn agentic systems.

Goldman's CIO predicts that while unit cost per token will decrease, the explosion in token usage from agentic systems will make total AI compute a major corporate expense. He suggests it should be compared to personnel costs, not traditional IT spending.

The success of personal AI assistants signals a massive shift in compute usage. While training models is resource-intensive, the next 10x in demand will come from widespread, continuous inference as millions of users run these agents. This effectively means consumers are buying fractions of datacenter GPUs like the GB200.