AI companies operate under the assumption that LLM prices will trend towards zero. This strategic bet means they intentionally de-prioritize heavy investment in cost optimization today, focusing instead on capturing the market and building features, confident that future, cheaper models will solve their margin problems for them.

Related Insights

The AI race has been a prisoner's dilemma where companies spend massively, fearing competitors will pull ahead. As the cost of next-gen systems like Blackwell and Rubin becomes astronomical, the sheer economics will force a shift. Decision-making will be dominated by ROI calculations rather than the existential dread of slowing down.

While the US pursues cutting-edge AGI, China is competing aggressively on cost at the application layer. By making LLM tokens and energy dramatically cheaper (e.g., $1.10 vs. $10+ per million tokens), China is fostering mass adoption and rapid commercialization. This strategy aims to win the practical, economic side of the AI race, even with less powerful models.

AI is making core software functionality nearly free, creating an existential crisis for traditional SaaS companies. The old model of 90%+ gross margins is disappearing. The future will be dominated by a few large AI players with lower margins, alongside a strategic shift towards monetizing high-value services.

As the current low-cost producer of AI tokens via its custom TPUs, Google's rational strategy is to operate at low or even negative margins. This "sucks the economic oxygen out of the AI ecosystem," making it difficult for capital-dependent competitors to justify their high costs and raise new funding rounds.

Current AI spending appears bubble-like, but it's not propping up unprofitable operations. Inference is already profitable. The immense cash burn is a deliberate, forward-looking investment in developing future, more powerful models, not a sign of a failing business model. This re-frames the financial risk.

Unlike SaaS, where high gross margins are key, an AI company with very high margins likely isn't seeing significant use of its core AI features. Low margins signal that customers are actively using compute-intensive products, a positive early indicator.

Many AI startups prioritize growth, leading to unsustainable gross margins (below 15%) due to high compute costs. This is a ticking time bomb. Eventually, these companies must undertake a costly, time-consuming re-architecture to optimize for cost and build a viable business.

Traditional SaaS metrics like 80%+ gross margins are misleading for AI companies. High inference costs lower margins, but if the absolute gross profit per customer is multiples higher than a SaaS equivalent, it's a superior business. The focus should shift from margin percentages to absolute gross profit dollars and multiples.

The AI value chain flows from hardware (NVIDIA) to apps, with LLM providers currently capturing most of the margin. The long-term viability of app-layer businesses depends on a competitive model layer. This competition drives down API costs, preventing model providers from having excessive pricing power and allowing apps to build sustainable businesses.

An emerging AI growth strategy involves using expensive frontier models to acquire users and distribution at an explosive rate, accepting poor initial margins. Once critical mass is reached, the company introduces its own fine-tuned, cheaper model, drastically improving unit economics overnight and capitalizing on the established user base.