Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

Coding assistant startup Cursor exemplifies a new AI playbook: start with a powerful open-weight base model (like China's Kimi), then apply significant reinforcement learning compute (3-4x the base model's) to achieve superior performance in a specific vertical. This strategy avoids the massive cost of pre-training a foundation model from scratch.

Related Insights

Startups like Cognition Labs find their edge not by competing on pre-training large models, but by mastering post-training. They build specialized reinforcement learning environments that teach models specific, real-world workflows (e.g., using Datadog for debugging), creating a defensible niche that larger players overlook.

China is gaining an efficiency edge in AI by using "distillation"—training smaller, cheaper models from larger ones. This "train the trainer" approach is much faster and challenges the capital-intensive US strategy, highlighting how inefficient and "bloated" current Western foundational models are.

AI labs like Anthropic find that mid-tier models can be trained with reinforcement learning to outperform their largest, most expensive models in just a few months, accelerating the pace of capability improvements.

Specialized models like Cursor's Composer 2 can achieve short-term dominance over general frontier models by hyper-focusing on a specific domain like coding. This 'hill climbing' strategy allows them to beat larger models on cost-performance, even if general models are predicted to win long-term.

The rise of Chinese AI models like DeepSeek and Kimmy in 2025 was driven by the startup and developer communities, not large enterprises. This bottom-up adoption pattern is reshaping the open-source landscape, creating a new competitive dynamic where nimble startups are leveraging these models long before they are vetted by corporate buyers.

Chinese AI models like Kimi achieve dramatic cost reductions through specific architectural choices, not just scale. Using a "mixture of experts" design, they only utilize a fraction of their total parameters for any given task, making them far more efficient to run than the "dense" models common in the West.

The belief that a single, god-level foundation model would dominate has proven false. Horowitz points to successful AI applications like Cursor, which uses 13 different models. This shows that value lies in the complex orchestration and design at the application layer, not just in having the largest single model.

Instead of relying on expensive, omni-purpose frontier models, companies can achieve better performance and lower costs. By creating a Reinforcement Learning (RL) environment specific to their application (e.g., a code editor), they can train smaller, specialized open-source models to excel at a fraction of the cost.

Leading Chinese AI models like Kimi appear to be primarily trained on the outputs of US models (a process called distillation) rather than being built from scratch. This suggests China's progress is constrained by its ability to scrape and fine-tune American APIs, indicating the U.S. still holds a significant architectural and innovation advantage in foundational AI.

To escape platform risk and high API costs, startups are building their own AI models. The strategy involves taking powerful, state-subsidized open-source models from China and fine-tuning them for specific use cases, creating a competitive alternative to relying on APIs from OpenAI or Anthropic.