While a unified data platform is non-negotiable for AI, leaders should resist standardizing AI tools and frameworks too early. Given the rapid pace of innovation, it's better to allow for experimentation and "let the flowers bloom." This dual approach—a stable data foundation with flexible tooling—enables both governance and agility.

Related Insights

Companies struggle with AI not because of the models, but because their data is siloed. Adopting an 'integration-first' mindset is crucial for creating the unified data foundation AI requires.

Despite the hype, LinkedIn found that third-party AI tools for coding and design don't work out-of-the-box on their complex, legacy stack. Success requires deep customization, re-architecting internal platforms for AI reasoning, and working in "alpha mode" with vendors to adapt their tools.

Before implementing AI, organizations must first build a unified data platform. Many companies have multiple, inconsistent "data lakes" and lack basic definitions for concepts like "customer" or "transaction." Without this foundational data consolidation, any attempt to derive insights with AI is doomed to fail due to semantic mismatches.

AI agent platforms are typically priced by usage, not seats, making initial costs low. Instead of a top-down mandate for one tool, leaders should encourage teams to expense and experiment with several options. The best solution for the team will emerge organically through use.

Organizations fail when they push teams directly into using AI for business outcomes ("architect mode"). Instead, they must first provide dedicated time and resources for unstructured play ("sandbox mode"). This experimentation phase is essential for building the skills and comfort needed to apply AI effectively to strategic goals.

The history of AI tools shows that products launching with fewer restrictions to empower individual developers (e.g., Stable Diffusion) tend to capture mindshare and adoption faster than cautious, locked-down competitors (e.g., DALL-E). Early-stage velocity trumps enterprise-grade caution.

MLOps pipelines manage model deployment, but scaling AI requires a broader "AI Operating System." This system serves as a central governance and integration layer, ensuring every AI solution across the business inherits auditable data lineage, compliance, and standardized policies.

The primary reason multi-million dollar AI initiatives stall or fail is not the sophistication of the models, but the underlying data layer. Traditional data infrastructure creates delays in moving and duplicating information, preventing the real-time, comprehensive data access required for AI to deliver business value. The focus on algorithms misses this foundational roadblock.

To balance security with agility, enterprises should run two AI tracks. Let the CIO's office develop secure, custom models for sensitive data while simultaneously empowering business units like marketing to use approved, low-risk SaaS AI tools to maintain momentum and drive immediate value.

According to Salesforce's AI chief, the primary challenge for large companies deploying AI is harmonizing data across siloed departments, like sales and marketing. AI cannot operate effectively without connected, unified data, making data integration the crucial first step before any advanced AI implementation.

In Early-Stage AI, Standardize Data Platforms but Allow Tooling to Proliferate | RiffOn