Users notice AI tools getting worse at simple tasks. This may not be a sign of technological regression, but rather a business decision by AI companies to run less powerful, cheaper models to reduce their astronomical operational costs, especially for free-tier users.

Related Insights

Reports that OpenAI hasn't completed a new full-scale pre-training run since May 2024 suggest a strategic shift. The race for raw model scale may be less critical than enhancing existing models with better reasoning and product features that customers demand. The business goal is profit, not necessarily achieving the next level of model intelligence.

The 'Andy Warhol Coke' era, where everyone could access the best AI for a low price, is over. As inference costs for more powerful models rise, companies are introducing expensive tiered access. This will create significant inequality in who can use frontier AI, with implications for transparency and regulation.

OpenAI found that significant upgrades to model intelligence, particularly for complex reasoning, did not improve user engagement. Users overwhelmingly prefer faster, simpler answers over more accurate but time-consuming responses, a disconnect that benefited competitors like Google.

The excitement around AI often overshadows its practical business implications. Implementing LLMs involves significant compute costs that scale with usage. Product leaders must analyze the ROI of different models to ensure financial viability before committing to a solution.

For the first time in years, the perceived leap in LLM capabilities has slowed. While models have improved, the cost increase (from $20 to $200/month for top-tier access) is not matched by a proportional increase in practical utility, suggesting a potential plateau or diminishing returns.

Companies like OpenAI and Anthropic are intentionally shrinking their flagship models (e.g., GPT-4.0 is smaller than GPT-4). The biggest constraint isn't creating more powerful models, but serving them at a speed users will tolerate. Slow models kill adoption, regardless of their intelligence.

AI companies operate under the assumption that LLM prices will trend towards zero. This strategic bet means they intentionally de-prioritize heavy investment in cost optimization today, focusing instead on capturing the market and building features, confident that future, cheaper models will solve their margin problems for them.

For consumer products like ChatGPT, models are already good enough for common queries. However, for complex enterprise tasks like coding, performance is far from solved. This gives model providers a durable path to sustained revenue growth through continued quality improvements aimed at professionals.

The common goal of increasing AI model efficiency could have a paradoxical outcome. If AI performance becomes radically cheaper ("too cheap to meter"), it could devalue the massive investments in compute and data center infrastructure, creating a financial crisis for the very companies that enabled the boom.

Contrary to the 'winner-takes-all' narrative, the rapid pace of innovation in AI is leading to a different outcome. As rival labs quickly match or exceed each other's model capabilities, the underlying Large Language Models (LLMs) risk becoming commodities, making it difficult for any single player to justify stratospheric valuations long-term.

Declining LLM Performance May Be an Economic Decision, Not a Technical Failure | RiffOn