We scan new podcasts and send you the top 5 insights daily.
As AI commoditizes software creation, the primary source of sustainable value shifts from the software itself to the unique, high-quality data that AI agents use for decision-making. Businesses must re-center their strategy around data as the core asset.
Public internet data has been largely exhausted for training AI models. The real competitive advantage and source for next-generation, specialized AI will be the vast, untapped reservoirs of proprietary data locked inside corporations, like R&D data from pharmaceutical or semiconductor companies.
As AI commoditizes user interfaces, enduring value will reside in the backend systems that are the authoritative source of data (e.g., payroll, financial records). These 'systems of record' are sticky due to regulation, business process integration, and high switching costs.
Instead of building AI models, a company can create immense value by being 'AI adjacent'. The strategy is to focus on enabling good AI by solving the foundational 'garbage in, garbage out' problem. Providing high-quality, complete, and well-understood data is a critical and defensible niche in the AI value chain.
As AI and better tools commoditize software creation, traditional technology moats are shrinking. The new defensible advantages are forms of liquidity: aggregated data, marketplace activity, or social interactions. These network effects are harder for competitors to replicate than code or features.
Since LLMs are commodities, sustainable competitive advantage in AI comes from leveraging proprietary data and unique business processes that competitors cannot replicate. Companies must focus on building AI that understands their specific "secret sauce."
For years, access to compute was the primary bottleneck in AI development. Now, as public web data is largely exhausted, the limiting factor is access to high-quality, proprietary data from enterprises and human experts. This shifts the focus from building massive infrastructure to forming data partnerships and expertise.
AI doesn't kill all software; it bifurcates the market. Companies with strong moats like distribution, proprietary data, and enterprise lock-in will thrive by integrating AI. However, companies whose only advantage was their software code will be wiped out as AI makes the code itself a commodity. The moat is no longer the software.
Software has long commanded premium valuations due to near-zero marginal distribution costs. AI breaks this model. The significant, variable cost of inference means expenses scale with usage, fundamentally altering software's economic profile and forcing valuations down toward those of traditional industries.
The 50-year supremacy of asset-light software may be an anomaly. If AI makes software creation nearly free, economic value will shift back to the historical mean: tangible assets like infrastructure, energy, and regulated, liability-bearing businesses that touch the physical world.
AI models are becoming commodities; the real, defensible value lies in proprietary data and user context. The correct strategy is for companies to use LLMs to enhance their existing business and data, rather than selling their valuable context to model providers for pennies on the dollar.