The data catalog market struggled because major platforms like Snowflake and dbt absorbed discovery features, making standalone tools redundant for human users. The bigger, missed opportunity was creating catalogs for machine-to-machine interaction (e.g., microservices, agents) and solving for governance rather than just discoverability.
Despite promises of a single source of truth, modern data platforms like Snowflake are often deployed for specific departments (e.g., marketing, finance), creating larger, more entrenched silos. This decentralization paradox persists because different business functions like analytics and operations require purpose-built data repositories, preventing true enterprise-wide consolidation.
The primary barrier to deploying AI agents at scale isn't the models but poor data infrastructure. The vast majority of organizations have immature data systems—uncatalogued, siloed, or outdated—making them unprepared for advanced AI and setting them up for failure.
Before implementing AI, organizations must first build a unified data platform. Many companies have multiple, inconsistent "data lakes" and lack basic definitions for concepts like "customer" or "transaction." Without this foundational data consolidation, any attempt to derive insights with AI is doomed to fail due to semantic mismatches.
A key differentiator is that Katera's AI agents operate directly on a company's existing data infrastructure (Snowflake, Redshift). Enterprises prefer this model because it avoids the security risks and complexities of sending sensitive data to a third-party platform for processing.
Point-solution SaaS products are at a massive disadvantage in the age of AI because they lack the broad, integrated dataset needed to power effective features. Bundled platforms that 'own the mine' of data are best positioned to win, as AI can perform magic when it has access to a rich, semantic data layer.
Companies struggle to get value from AI because their data is fragmented across different systems (ERP, CRM, finance) with poor integrity. The primary challenge isn't the AI models themselves, but integrating these disparate data sets into a unified platform that agents can act upon.
The primary reason multi-million dollar AI initiatives stall or fail is not the sophistication of the models, but the underlying data layer. Traditional data infrastructure creates delays in moving and duplicating information, preventing the real-time, comprehensive data access required for AI to deliver business value. The focus on algorithms misses this foundational roadblock.
The boom in tools for data teams faded because the Total Addressable Market (TAM) was overestimated. Investors and founders pattern-matched the data space to larger markets like cloud and dev tools, but the actual number of teams with the budget and need for sophisticated data tooling proved to be much smaller.
Snowflake Intelligence is intentionally an "opinionated agentic platform." Unlike generic AI tools from cloud providers that aim to do everything, Snowflake focuses narrowly on helping users get value from their data. This avoids the paralysis of infinite choice and delivers more practical, immediate utility.
Many companies focus on AI models first, only to hit a wall. An "integration-first" approach is a strategic imperative. Connecting disparate systems *before* building agents ensures they have the necessary data to be effective, avoiding the "garbage in, garbage out" trap at a foundational level.