The "AI wrapper" concern is mitigated by a multi-model strategy. A startup can integrate the best models from various providers for different tasks, creating a superior product. A platform like OpenAI is incentivized to only use its own models, creating a durable advantage for the startup.
To build a durable business on top of foundation models, go beyond a simple API call. Gamma creates a moat by deeply owning an entire workflow (visual communication) and orchestrating over 20 different specialized AI models, each chosen for a specific sub-task in the user journey.
Simply offering the latest model is no longer a competitive advantage. True value is created in the system built around the model—the system prompts, tools, and overall scaffolding. This 'harness' is what optimizes a model's performance for specific tasks and delivers a superior user experience.
The notion of building a business as a 'thin wrapper' around a foundational model like GPT is flawed. Truly defensible AI products, like Cursor, build numerous specific, fine-tuned models to deeply understand a user's domain. This creates a data and performance moat that a generic model cannot easily replicate, much like Salesforce was more than just a 'thin wrapper' on a database.
Incumbent companies are slowed by the need to retrofit AI into existing processes and tribal knowledge. AI-native startups, however, can build their entire operational model around agent-based, prompt-driven workflows from day one, creating a structural advantage that is difficult for larger companies to copy.
The enduring moat in the AI stack lies in what is hardest to replicate. Since building foundation models is significantly more difficult than building applications on top of them, the model layer is inherently more defensible and will naturally capture more value over time.
Startups are becoming wary of building on OpenAI's platform due to the significant risk of OpenAI launching competing applications (e.g., Sora for video), rendering their products obsolete. This "platform risk" is pushing developers toward neutral providers like Anthropic or open-source models to protect their businesses.
The choice between open and closed-source AI is not just technical but strategic. For startups, feeding proprietary data to a closed-source provider like OpenAI, which competes across many verticals, creates long-term risk. Open-source models offer "strategic autonomy" and prevent dependency on a potential future rival.
Initially, even OpenAI believed a single, ultimate 'model to rule them all' would emerge. This thinking has completely changed to favor a proliferation of specialized models, creating a healthier, less winner-take-all ecosystem where different models serve different needs.
Perplexity's CEO argues that building foundational models is not necessary for success. By focusing on the end-to-end consumer experience and leveraging increasingly commoditized models, startups can build a highly valuable business without needing billions in funding for model training.
Instead of offering a model selector, creating a proprietary, branded model allows a company to chain different specialized models for various sub-tasks (e.g., search, generation). This not only improves overall performance but also provides business independence from the pricing and launch cycles of a single frontier model lab.