While the most powerful AI will reside in large "god models" (like supercomputers), the majority of the market volume will come from smaller, specialized models. These will cascade down in size and cost, eventually being embedded in every device, much like microchips proliferated from mainframes.
The AI market is becoming "polytheistic," with numerous specialized models excelling at niche tasks, rather than "monotheistic," where a single super-model dominates. This fragmentation creates opportunities for differentiated startups to thrive by building effective models for specific use cases, as no single model has mastered everything.
The AI industry is hitting data limits for training massive, general-purpose models. The next wave of progress will likely come from creating highly specialized models for specific domains, similar to DeepMind's AlphaFold, which can achieve superhuman performance on narrow tasks.
The AI market is split between two strategies. Some companies build hyper-expensive, complex models (the "cappuccino machine") targeting the whole world. Others focus on cheaper, standardized, and accessible solutions (the "coffee pod"), creating a fundamental strategic divide for where value will accrue.
Models like Gemini 3 Flash show a key trend: making frontier intelligence faster, cheaper, and more efficient. The trajectory is for today's state-of-the-art models to become 10x cheaper within a year, enabling widespread, low-latency, and on-device deployment.
Instead of relying solely on massive, expensive, general-purpose LLMs, the trend is toward creating smaller, focused models trained on specific business data. These "niche" models are more cost-effective to run, less likely to hallucinate, and far more effective at performing specific, defined tasks for the enterprise.
The current focus on building massive, centralized AI training clusters represents the 'mainframe' era of AI. The next three years will see a shift toward a distributed model, similar to computing's move from mainframes to PCs. This involves pushing smaller, efficient inference models out to a wide array of devices.
The AI arms race will shift from building ever-larger general models to creating smaller, highly specialized models for domains like medicine and law. General AIs will evolve to act as "general contractors," routing user queries to the appropriate specialist model for deeper expertise.
Beyond the simple training-inference binary, Arm's CEO sees a third category: smaller, specialized models for reinforcement learning. These chips will handle both training and inference, acting like 'student teachers' taught by giant foundational models.
Block's CTO believes the key to building complex applications with AI isn't a single, powerful model. Instead, he predicts a future of "swarm intelligence"—where hundreds of smaller, cheaper, open-source agents work collaboratively, with their collective capability surpassing any individual large model.
The true commercial impact of AI will likely come from small, specialized "micro models" solving boring, high-volume business tasks. While highly valuable, these models are cheap to run and cannot economically justify the current massive capital expenditure on AGI-focused data centers.