The industry has already exhausted the public web data used to train foundational AI models, a point underscored by the phrase "we've already run out of data." The next leap in AI capability and business value will come from harnessing the vast, proprietary data currently locked behind corporate firewalls.

Related Insights

LLMs have hit a wall by scraping nearly all available public data. The next phase of AI development and competitive differentiation will come from training models on high-quality, proprietary data generated by human experts. This creates a booming "data as a service" industry for companies like Micro One that recruit and manage these experts.

As startups build on commoditized AI platforms like GPT, product differentiation becomes less of a moat. Success now hinges on cracking growth faster than rivals. The new competitive advantages are proprietary data for training models and the deep domain expertise required to find unique growth levers.

The primary barrier to deploying AI agents at scale isn't the models but poor data infrastructure. The vast majority of organizations have immature data systems—uncatalogued, siloed, or outdated—making them unprepared for advanced AI and setting them up for failure.

The long-sought goal of "information at your fingertips," envisioned by Bill Gates, wasn't achieved through structured databases as expected. Instead, large neural networks unexpectedly became the key, capable of finding patterns in messy, unstructured enterprise data where rigid schemas failed.

For years, access to compute was the primary bottleneck in AI development. Now, as public web data is largely exhausted, the limiting factor is access to high-quality, proprietary data from enterprises and human experts. This shifts the focus from building massive infrastructure to forming data partnerships and expertise.

The future of valuable AI lies not in models trained on the abundant public internet, but in those built on scarce, proprietary data. For fields like robotics and biology, this data doesn't exist to be scraped; it must be actively created, making the data generation process itself the key competitive moat.

The true enterprise value of AI lies not in consuming third-party models, but in building internal capabilities to diffuse intelligence throughout the organization. This means creating proprietary "AI factories" rather than just using external tools and admiring others' success.

The primary reason multi-million dollar AI initiatives stall or fail is not the sophistication of the models, but the underlying data layer. Traditional data infrastructure creates delays in moving and duplicating information, preventing the real-time, comprehensive data access required for AI to deliver business value. The focus on algorithms misses this foundational roadblock.

If a company and its competitor both ask a generic LLM for strategy, they'll get the same answer, erasing any edge. The only way to generate unique, defensible strategies is by building evolving models trained on a company's own private data.

According to Salesforce's AI chief, the primary challenge for large companies deploying AI is harmonizing data across siloed departments, like sales and marketing. AI cannot operate effectively without connected, unified data, making data integration the crucial first step before any advanced AI implementation.

AI Model Progress Now Hinges on Unlocking Trapped Enterprise Data | RiffOn