AI models are becoming commodities; the real, defensible value lies in proprietary data and user context. The correct strategy is for companies to use LLMs to enhance their existing business and data, rather than selling their valuable context to model providers for pennies on the dollar.
LLMs have hit a wall by scraping nearly all available public data. The next phase of AI development and competitive differentiation will come from training models on high-quality, proprietary data generated by human experts. This creates a booming "data as a service" industry for companies like Micro One that recruit and manage these experts.
As startups build on commoditized AI platforms like GPT, product differentiation becomes less of a moat. Success now hinges on cracking growth faster than rivals. The new competitive advantages are proprietary data for training models and the deep domain expertise required to find unique growth levers.
The notion of building a business as a 'thin wrapper' around a foundational model like GPT is flawed. Truly defensible AI products, like Cursor, build numerous specific, fine-tuned models to deeply understand a user's domain. This creates a data and performance moat that a generic model cannot easily replicate, much like Salesforce was more than just a 'thin wrapper' on a database.
Instead of building AI models, a company can create immense value by being 'AI adjacent'. The strategy is to focus on enabling good AI by solving the foundational 'garbage in, garbage out' problem. Providing high-quality, complete, and well-understood data is a critical and defensible niche in the AI value chain.
Most successful SaaS companies weren't built on new core tech, but by packaging existing tech (like databases or CRMs) into solutions for specific industries. AI is no different. The opportunity lies in unbundling a general tool like ChatGPT and rebundling its capabilities into vertical-specific products.
Unlike sticky cloud infrastructure (AWS, GCP), LLMs are easily interchangeable via APIs, leading to customer "promiscuity." This commoditizes the model layer and forces providers like OpenAI to build defensible moats at the application layer (e.g., ChatGPT) where they can own the end user.
The true enterprise value of AI lies not in consuming third-party models, but in building internal capabilities to diffuse intelligence throughout the organization. This means creating proprietary "AI factories" rather than just using external tools and admiring others' success.
Instead of gating its valuable review data like traditional analyst firms, G2 strategically chose to syndicate it and make it available to LLMs. This ensures G2 remains a trusted, cited source within AI-generated answers, maintaining brand influence and relevance where buyers are now making decisions.
If a company and its competitor both ask a generic LLM for strategy, they'll get the same answer, erasing any edge. The only way to generate unique, defensible strategies is by building evolving models trained on a company's own private data.