Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

Beyond low electricity costs, Chinese AI models achieve a structural cost advantage through their "mixture of experts" architecture. This technical approach, spurred by US chip restrictions, requires less computing power to generate tokens compared to prevalent US systems.

Related Insights

While the US pursues cutting-edge AGI, China is competing aggressively on cost at the application layer. By making LLM tokens and energy dramatically cheaper (e.g., $1.10 vs. $10+ per million tokens), China is fostering mass adoption and rapid commercialization. This strategy aims to win the practical, economic side of the AI race, even with less powerful models.

China is gaining an efficiency edge in AI by using "distillation"—training smaller, cheaper models from larger ones. This "train the trainer" approach is much faster and challenges the capital-intensive US strategy, highlighting how inefficient and "bloated" current Western foundational models are.

Airbnb's reliance on Alibaba's QWEN 3 model as a more affordable alternative to US models signals a critical trend. As Chinese models approach performance parity, their significant cost advantage is making them a viable and attractive choice for Western companies, challenging the market dominance of US-based labs.

Model performance isn't just about architecture; it's also about compute budget. A less sophisticated AI model, if allowed to run for longer or iterate more times, can often match the output of a state-of-the-art model. This suggests access to cheap energy could be a greater advantage than access to the best chips.

Chinese AI models like Kimi achieve dramatic cost reductions through specific architectural choices, not just scale. Using a "mixture of experts" design, they only utilize a fraction of their total parameters for any given task, making them far more efficient to run than the "dense" models common in the West.

Faced with limited access to top-tier hardware, Chinese AI companies have been forced to innovate on model architecture to compete. They've developed superior techniques in memory management and multi-token prediction, making their models highly efficient and formidable competitors despite hardware constraints.

The shift from simple chatbots to task-oriented "agentic AI" dramatically increases the demand for AI tokens. This makes China's ability to produce tokens cheaply a more critical and growing strategic advantage, as the resource becomes increasingly scarce and valuable.

China is compensating for its deficit in cutting-edge semiconductors by pursuing an asymmetric strategy. It focuses on massive 'superclusters' of less advanced domestic chips and creating hyper-efficient, open-source AI models. This approach prioritizes widespread, low-cost adoption over chasing the absolute peak of performance like the US.

China is gaining a structural advantage in the global AI race by producing and exporting AI tokens—the computational fuel for LLMs—at a fraction of the cost of US alternatives. This is attracting global startups and creating geopolitical dependency on China's "new oil."

According to Jensen Huang, China's lack of cutting-edge chips is not a fatal flaw. Its abundant, cheap energy allows it to use a larger number of less-efficient chips in parallel to achieve the same computational output as labs using fewer, more advanced chips.