The most significant enterprise challenges for AI are the 'unstated constraints'—institutional knowledge, compliance nuances, and stakeholder dynamics not documented anywhere. The human operator who can identify and translate this implicit context for AI agents becomes indispensable.

Related Insights

The transformative power of AI agents is unlocked by professionals with deep domain knowledge who can craft highly specific, iterative prompts and integrate the agent into a valid workflow. The technology itself does not compensate for a lack of expertise or flawed underlying processes.

As AI evolves from single-task tools to autonomous agents, the human role transforms. Instead of simply using AI, professionals will need to manage and oversee multiple AI agents, ensuring their actions are safe, ethical, and aligned with business goals, acting as a critical control layer.

The effectiveness of enterprise AI agents is limited not by data access, but by the absence of context for *why* decisions were made. 'Context graphs' aim to solve this by capturing 'decision traces'—exceptions, precedents, and overrides that currently live in Slack threads and employee's heads, creating a true source of truth for automation.

AI models lack access to the rich, contextual signals from physical, real-world interactions. Humans will remain essential because their job is to participate in this world, gather unique context from experiences like customer conversations, and feed it into AI systems, which cannot glean it on their own.

Off-the-shelf AI models can only go so far. The true bottleneck for enterprise adoption is "digitizing judgment"—capturing the unique, context-specific expertise of employees within that company. A document's meaning can change entirely from one company to another, requiring internal labeling.

Despite AI's capabilities, it lacks the full context necessary for nuanced business decisions. The most valuable work happens when people with diverse perspectives convene to solve problems, leveraging a collective understanding that AI cannot access. Technology should augment this, not replace it.

To build coordinated AI agent systems, firms must first extract siloed operational knowledge. This involves not just digitizing documents but systematically observing employee actions like browser clicks and phone calls to capture unwritten processes, turning this tacit knowledge into usable context for AI.

AI tools like LLMs thrive on large, structured datasets. In manufacturing, critical information is often unstructured 'tribal knowledge' in workers' heads. Dirac’s strategy is to first build a software layer that captures and organizes this human expertise, creating the necessary context for AI to then analyze and add value.

GSB professors warn that professionals who merely use AI as a black box—passing queries and returning outputs—risk minimizing their own role. To remain valuable, leaders must understand the underlying models and assumptions to properly evaluate AI-generated solutions and maintain control of the decision-making process.

The most valuable AI systems are built by people with deep knowledge in a specific field (like pest control or law), not by engineers. This expertise is crucial for identifying the right problems and, more importantly, for creating effective evaluations to ensure the agent performs correctly.

AI Agents Reveal the Hidden Value of an Operator's 'Unstated Constraints' Knowledge | RiffOn