Chinese AI models like Kimi achieve dramatic cost reductions through specific architectural choices, not just scale. Using a "mixture of experts" design, they only utilize a fraction of their total parameters for any given task, making them far more efficient to run than the "dense" models common in the West.

Related Insights

While more data and compute yield linear improvements, true step-function advances in AI come from unpredictable algorithmic breakthroughs like Transformers. These creative ideas are the most difficult to innovate on and represent the highest-leverage, yet riskiest, area for investment and research focus.

While the US pursues cutting-edge AGI, China is competing aggressively on cost at the application layer. By making LLM tokens and energy dramatically cheaper (e.g., $1.10 vs. $10+ per million tokens), China is fostering mass adoption and rapid commercialization. This strategy aims to win the practical, economic side of the AI race, even with less powerful models.

China is pursuing a low-cost, open-source AI model, similar to Android's market strategy. This contrasts with the US's expensive, high-performance "iPhone" approach. This accessibility and cost-effectiveness could allow Chinese AI to dominate the global market, especially in developing nations.

The history of AI, such as the 2012 AlexNet breakthrough, demonstrates that scaling compute and data on simpler, older algorithms often yields greater advances than designing intricate new ones. This "bitter lesson" suggests prioritizing scalability over algorithmic complexity for future progress.

Model architecture decisions directly impact inference performance. AI company Zyphra pre-selects target hardware and then chooses model parameters—such as a hidden dimension with many powers of two—to align with how GPUs split up workloads, maximizing efficiency from day one.

A common misconception is that Chinese AI is fully open-source. The reality is they are often "open-weight," meaning training parameters (weights) are shared, but the underlying code and proprietary datasets are not. This provides a competitive advantage by enabling adoption while maintaining some control.

The exceptionally low cost of developing and operating AI models in China is forcing a reckoning in the US tech sector. American investors and companies are now questioning the high valuations and expensive operating costs of their domestic AI, creating fear that the US AI boom is a bubble inflated by high costs rather than superior technology.

China is compensating for its deficit in cutting-edge semiconductors by pursuing an asymmetric strategy. It focuses on massive 'superclusters' of less advanced domestic chips and creating hyper-efficient, open-source AI models. This approach prioritizes widespread, low-cost adoption over chasing the absolute peak of performance like the US.

Many AI startups prioritize growth, leading to unsustainable gross margins (below 15%) due to high compute costs. This is a ticking time bomb. Eventually, these companies must undertake a costly, time-consuming re-architecture to optimize for cost and build a viable business.

While the West may lead in AI models, China's key strategic advantage is its ability to 'embody' AI in hardware. Decades of de-industrialization in the U.S. have left a gap, while China's manufacturing dominance allows it to integrate AI into cars, drones, and robots at a scale the West cannot currently match.

Architectural Innovation Is Key to China's AI Cost Efficiency | RiffOn