Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

An advanced user reveals their largest new expense from building AI agents isn't tokens, but database and storage costs. AI makes vast amounts of previously inert data useful, creating a surge in demand for storage solutions, which is where the real economic leverage lies.

Related Insights

The excitement around AI often overshadows its practical business implications. Implementing LLMs involves significant compute costs that scale with usage. Product leaders must analyze the ROI of different models to ensure financial viability before committing to a solution.

For years, access to compute was the primary bottleneck in AI development. Now, as public web data is largely exhausted, the limiting factor is access to high-quality, proprietary data from enterprises and human experts. This shifts the focus from building massive infrastructure to forming data partnerships and expertise.

A paradox exists where the cost for a fixed level of AI capability (e.g., GPT-4 level) has dropped 100-1000x. However, overall enterprise spend is increasing because applications now use frontier models with massive contexts and multi-step agentic workflows, creating huge multipliers on token usage that drive up total costs.

The next wave of AI compute demand won't be from generating more outputs, but from agents performing exponentially more data collection for a single task. For example, a financial model could trigger an agent to analyze vast datasets, like satellite imagery, multiplying token usage for one result.

The primary short-term risk for the AI sector isn't capital expenditure but the high cost of token generation. For AI applications to become ubiquitous, the unit economics must improve. If running a single query remains prohibitively expensive for businesses, widespread, sustainable adoption will be impossible, threatening the entire investment thesis.

Data is becoming more expensive not from scarcity, but because the work has evolved. Simple labeling is over. Costs are now driven by the need for pricey domain experts for specialized data preparation and creative teams to build complex, synthetic environments for training agents.

While user growth for apps like ChatGPT is slowing, per-user token consumption is skyrocketing as models shift from simple queries to complex reasoning and AI agents. This creates a hidden, exponential growth in compute demand, validating Oracle's massive infrastructure investment even as front-end adoption matures.

While the cost for GPT-4 level intelligence has dropped over 100x, total enterprise AI spend is rising. This is driven by multipliers: using larger frontier models for harder tasks, reasoning-heavy workflows that consume more tokens, and complex, multi-turn agentic systems.

Goldman's CIO predicts that while unit cost per token will decrease, the explosion in token usage from agentic systems will make total AI compute a major corporate expense. He suggests it should be compared to personnel costs, not traditional IT spending.

The shift to usage-based pricing for AI tools isn't just a revenue growth strategy. Enterprise vendors are adopting it to offset their own escalating cloud infrastructure costs, which scale directly with customer usage, thereby protecting their profit margins from their own suppliers.