AI's growth is hampered by a measurement problem, much like early digital advertising. The industry's acceleration won't come from better AI models alone, but from building a 'boring' infrastructure, like Comscore did for ads, to prove the tools actually work.

Related Insights

Companies feel immense pressure to integrate AI to stay competitive, leading to massive spending. However, this rush means they lack the infrastructure to measure ROI, creating a paradox of anxious investment without clear proof of value.

Instead of building AI models, a company can create immense value by being 'AI adjacent'. The strategy is to focus on enabling good AI by solving the foundational 'garbage in, garbage out' problem. Providing high-quality, complete, and well-understood data is a critical and defensible niche in the AI value chain.

Marketing leaders pressured to adopt AI are discovering the primary obstacle isn't the technology, but their own internal data infrastructure. Siloed, inconsistently structured data across teams prevents them from effectively leveraging AI for consumer insights and business growth.

To evaluate AI's role in building relationships, marketers must look beyond transactional KPIs. Leading indicators of success include sustained engagement, customers volunteering more information, and recommending the experience to others. These metrics quantify brand trust and empathy—proving the brand is earning belief, not just attention.

For years, access to compute was the primary bottleneck in AI development. Now, as public web data is largely exhausted, the limiting factor is access to high-quality, proprietary data from enterprises and human experts. This shifts the focus from building massive infrastructure to forming data partnerships and expertise.

Don't trust academic benchmarks. Labs often "hill climb" or game them for marketing purposes, which doesn't translate to real-world capability. Furthermore, many of these benchmarks contain incorrect answers and messy data, making them an unreliable measure of true AI advancement.

The primary bottleneck in improving AI is no longer data or compute, but the creation of 'evals'—tests that measure a model's capabilities. These evals act as product requirement documents (PRDs) for researchers, defining what success looks like and guiding the training process.

Instead of relying on hyped benchmarks, the truest measure of the AI industry's progress is the physical build-out of data centers. Tracking permits, power consumption, and satellite imagery reveals the concrete, multi-billion dollar bets being placed, offering a grounded view that challenges both extreme skeptics and believers.

The primary reason multi-million dollar AI initiatives stall or fail is not the sophistication of the models, but the underlying data layer. Traditional data infrastructure creates delays in moving and duplicating information, preventing the real-time, comprehensive data access required for AI to deliver business value. The focus on algorithms misses this foundational roadblock.

According to Salesforce's AI chief, the primary challenge for large companies deploying AI is harmonizing data across siloed departments, like sales and marketing. AI cannot operate effectively without connected, unified data, making data integration the crucial first step before any advanced AI implementation.