We scan new podcasts and send you the top 5 insights daily.
The traditional "data doubles every year" metric is outdated. The proliferation of AI agents running queries and generating activity will cause an exponential explosion in data volume, far exceeding human-generated data and approaching 10x annual growth.
The industry has already exhausted the public web data used to train foundational AI models, a point underscored by the phrase "we've already run out of data." The next leap in AI capability and business value will come from harnessing the vast, proprietary data currently locked behind corporate firewalls.
Leading LLMs can now replicate a two-hour human software engineering task with 50% accuracy. This capability is doubling every seven months, signaling an urgent need for organizations to adapt their data infrastructure, security, and governance to leverage this exponential growth.
The shift from simple chatbots (one user request, one API call) to agentic AI systems will decouple inference requests from direct user actions. A single user request could trigger hundreds or thousands of automated model calls, leading to an exponential increase in compute demand and cost.
Early AI adoption focused on saving time. The new wave, driven by agentic systems, derives its primary value from enabling completely new functions and significantly increasing throughput, representing a move from efficiency to opportunity-focused ROI.
As AI agents and developers operate increasingly within the terminal (CLI), demand for programmatic, API-driven data access will explode. This will replace clunky web UIs and credit card subscriptions with seamless, micro-transaction-based data consumption.
Ben Thompson argues the shift from simple chatbots to AI agents creates an exponential, non-speculative demand for compute. Agents automate complex, multi-step tasks, driving constant usage that justifies the massive capex investments by hyperscalers. This suggests the current spending is based on real demand, not bubble-fueled speculation.
The current AI data center arms race isn't about meeting today's demand for chatbots. It's fueled by companies like Meta betting on a future where personal AI agents run constantly, analyzing every interaction. This vision of persistent, parallel agents requires an exponential increase in compute, explaining why they will buy any available capacity.
The next wave of AI compute demand won't be from generating more outputs, but from agents performing exponentially more data collection for a single task. For example, a financial model could trigger an agent to analyze vast datasets, like satellite imagery, multiplying token usage for one result.
While user growth for apps like ChatGPT is slowing, per-user token consumption is skyrocketing as models shift from simple queries to complex reasoning and AI agents. This creates a hidden, exponential growth in compute demand, validating Oracle's massive infrastructure investment even as front-end adoption matures.
The success of personal AI assistants signals a massive shift in compute usage. While training models is resource-intensive, the next 10x in demand will come from widespread, continuous inference as millions of users run these agents. This effectively means consumers are buying fractions of datacenter GPUs like the GB200.