Models like Gemini 3 Flash show a key trend: making frontier intelligence faster, cheaper, and more efficient. The trajectory is for today's state-of-the-art models to become 10x cheaper within a year, enabling widespread, low-latency, and on-device deployment.

Related Insights

The primary threat from competitors like Google may not be a superior model, but a more cost-efficient one. Google's Gemini 3 Flash offers "frontier-level intelligence" at a fraction of the cost. This shifts the competitive battleground from pure performance to price-performance, potentially undermining business models built on expensive, large-scale compute.

As frontier AI models reach a plateau of perceived intelligence, the key differentiator is shifting to user experience. Low-latency, reliable performance is becoming more critical than marginal gains on benchmarks, making speed the next major competitive vector for AI products like ChatGPT.

AI labs like Anthropic find that mid-tier models can be trained with reinforcement learning to outperform their largest, most expensive models in just a few months, accelerating the pace of capability improvements.

Apple's seemingly slow AI progress is likely a strategic bet that today's powerful cloud-based models will become efficient enough to run locally on devices within 12 months. This would allow them to offer powerful AI with superior privacy, potentially leapfrogging competitors.

Chinese AI models like Kimi achieve dramatic cost reductions through specific architectural choices, not just scale. Using a "mixture of experts" design, they only utilize a fraction of their total parameters for any given task, making them far more efficient to run than the "dense" models common in the West.

According to Ring's founder, the technology for ambitious AI features like "Dog Search Party" already exists. The real bottleneck is the cost of computation. Products that are technically possible today are often not launched because the processing expense makes them commercially unviable.

Arvind Krishna forecasts a 1000x drop in AI compute costs over five years. This won't just come from better chips (a 10x gain). It will be compounded by new processor architectures (another 10x) and major software optimizations like model compression and quantization (a final 10x).

A cost-effective AI architecture involves using a small, local model on the user's device to pre-process requests. This local AI can condense large inputs into an efficient, smaller prompt before sending it to the expensive, powerful cloud model, optimizing resource usage.

The true commercial impact of AI will likely come from small, specialized "micro models" solving boring, high-volume business tasks. While highly valuable, these models are cheap to run and cannot economically justify the current massive capital expenditure on AGI-focused data centers.

The narrative of endless demand for NVIDIA's high-end GPUs is flawed. It will be cracked by two forces: the shift of AI inference to on-device flash memory, reducing cloud reliance, and Google's ability to give away its increasingly powerful Gemini AI for free, undercutting the revenue models that fuel GPU demand.