Analyst Doug O'Loughlin questions why OpenAI hasn't announced a new, scaled-up base model pre-training run, unlike competitors such as Google with Gemini 3. He speculates this could indicate underlying issues, such as instability with NVIDIA's new GB200 chips, preventing them from successfully completing the next major training effort and potentially stalling their progress on the capability frontier.

Related Insights

Reports that OpenAI hasn't completed a new full-scale pre-training run since May 2024 suggest a strategic shift. The race for raw model scale may be less critical than enhancing existing models with better reasoning and product features that customers demand. The business goal is profit, not necessarily achieving the next level of model intelligence.

The era of advancing AI simply by scaling pre-training is ending due to data limits. The field is re-entering a research-heavy phase focused on novel, more efficient training paradigms beyond just adding more compute to existing recipes. The bottleneck is shifting from resources back to ideas.

The plateauing performance-per-watt of GPUs suggests that simply scaling current matrix multiplication-heavy architectures is unsustainable. This hardware limitation may necessitate research into new computational primitives and neural network designs built for large-scale distributed systems, not single devices.

Google successfully trained its top model, Gemini 3 Pro, on its own TPUs, proving a viable alternative to NVIDIA's chips. However, because Google doesn't sell these TPUs, NVIDIA retains its monopoly pricing power over every other company in the market.

The era of guaranteed progress by simply scaling up compute and data for pre-training is ending. With massive compute now available, the bottleneck is no longer resources but fundamental ideas. The AI field is re-entering a period where novel research, not just scaling existing recipes, will drive the next breakthroughs.

OpenAI favors "zero gradient" prompt optimization because serving thousands of unique, fine-tuned model snapshots is operationally very difficult. Prompt-based adjustments allow performance gains without the immense infrastructure burden, making it a more practical and scalable approach for both OpenAI and developers.

Google's latest AI model, Gemini 3, is perceived as so advanced that OpenAI's CEO privately warned staff to expect "rough vibes" and "temporary economic headwinds." This memo signals a significant competitive shift, acknowledging Google may have temporarily leapfrogged OpenAI in model development.

AI progress was expected to stall in 2024-2025 due to hardware limitations on pre-training scaling laws. However, breakthroughs in post-training techniques like reasoning and test-time compute provided a new vector for improvement, bridging the gap until next-generation chips like NVIDIA's Blackwell arrived.

OpenAI is now reacting to Google's advancements with Gemini 3, a complete reversal from three years ago. Google's strengths in infrastructure, proprietary chips, data, and financial stability are giving it a significant competitive edge, forcing OpenAI to delay initiatives and refocus on its core ChatGPT product.

While competitors like OpenAI must buy GPUs from NVIDIA, Google trains its frontier AI models (like Gemini) on its own custom Tensor Processing Units (TPUs). This vertical integration gives Google a significant, often overlooked, strategic advantage in cost, efficiency, and long-term innovation in the AI race.