The Bottleneck Nobody Talks About

Enterprise AI has a dirty secret. The bottleneck is not model quality. GPT-4, Claude, Gemini, and their successors are remarkably capable. The bottleneck is not infrastructure. Kubernetes, cloud providers, and a mature ecosystem of deployment tools have made running AI workloads operationally tractable. The bottleneck is not even budget. Enterprises are spending freely.

The bottleneck is knowledge. Specifically, the inability to codify institutional knowledge — the rules, workflows, judgment calls, and domain expertise that make an organization function — into a form that AI systems can execute. Until that bottleneck breaks, most enterprise AI deployments will remain stuck at Assist: helpful copilots that suggest, but never truly act.


TL;DR

  • 88% of enterprises have adopted AI in some form, but the vast majority remain at the Assist level — AI as suggestion engine, human as decision-maker.
  • The productivity J-curve (Brynjolfsson, MIT) shows that firms layering AI on existing workflows see the smallest gains. Transformation requires organizational redesign.
  • Polanyi’s Paradox — “we can know more than we can tell” — explains why expert knowledge resists codification. 42% of organizational expertise is held only by the individual who possesses it.
  • Fortune 500 companies lose an estimated $31.5 billion annually from failure to share knowledge effectively.
  • Every previous automation wave — from Taylorism to expert systems to RPA — required a preceding codification phase. AI is no different.

The J-Curve Problem

Erik Brynjolfsson at the MIT Initiative on the Digital Economy has documented what he calls the Productivity J-Curve. The core finding: when firms layer AI on top of existing workflows — plugging a chatbot into an existing support process, adding a copilot to an existing coding environment — the productivity gains are real but small.

The firms that climb the right side of the J-curve are those that redesign workflows around AI capabilities. They do not add AI to the existing process. They rebuild the process with AI as a core component. This is the difference between giving a customer service agent an AI copilot (Assist) and building a system where AI resolves tier-1 tickets end-to-end while routing complex cases to specialized human agents (Execute).

But here is the catch: redesigning workflows around AI requires first codifying those workflows in a machine-executable form. And that codification step is where most organizations stall.


Polanyi’s Paradox in the Enterprise

In 1966, the philosopher Michael Polanyi articulated what is now known as Polanyi’s Paradox: “We can know more than we can tell.” A master craftsman knows how to shape wood, but cannot fully articulate the micro-decisions that guide each cut. An experienced support agent knows when to escalate and when to persist, but the decision factors — tone of voice, customer history, issue trajectory — resist enumeration.

The Tacit Knowledge Problem

Research on organizational knowledge consistently finds that a staggering proportion of expertise is tacit:

  • 42% of organizational expertise is only known to the individual who possesses it (Panopto Workplace Knowledge Report)
  • Workers spend an average of 2.5 hours per day searching for information they need (McKinsey Global Institute)
  • Fortune 500 companies lose an estimated $31.5 billion annually from failure to share knowledge (IDC)

AI systems are only as good as the knowledge they can access. If 42% of the knowledge that drives your business exists only in the heads of your most experienced people, your AI system is operating with less than 60% of the information it needs.

What Tacit Knowledge Looks Like in Practice

Consider a customer service organization handling billing disputes. The documented process might say:

  1. Verify customer identity
  2. Review the disputed charge
  3. Check refund eligibility per policy
  4. Process refund or explain denial
  5. Offer escalation if customer is unsatisfied

This is the explicit knowledge. An AI trained on this process will execute it competently for straightforward cases.

But an experienced agent knows things the process document does not say:

  • Enterprise customers with annual contracts over $100K get more flexible refund policies, even though the policy document does not differentiate
  • Disputes related to the Q4 billing system migration should be auto-approved because the migration introduced a known bug
  • A customer who mentions “cancellation” in a billing dispute is a retention risk and should be routed to the retention team
  • The refund policy document says 30 days, but in practice, managers approve 60-day refunds for customers with clean payment histories

None of this is written down. All of it is critical to doing the job well.


The Historical Pattern

The knowledge codification challenge is not new to AI. Every automation wave in industrial history has followed the same pattern: codification must precede automation.

Taylorism and Scientific Management (1880s-1920s): Before factories could be automated with assembly lines, Frederick Taylor and his followers spent decades studying and codifying manual work processes.

Expert Systems (1970s-1990s): The first wave of enterprise AI attempted to codify expert knowledge into rule-based systems. The knowledge engineering bottleneck is the primary reason expert systems failed to achieve broad adoption.

Robotic Process Automation (2010s-2020s): RPA worked brilliantly for processes that were already codified. It failed for processes that required judgment. The 30-40% failure rate maps directly to the percentage of process knowledge that was tacit.

In each case: new automation technology emerges, early adopters achieve gains by automating already-codified processes, and scaling stalls when automation encounters tacit knowledge.

AI in 2026 is at stage 3. The technology is powerful enough to automate nearly anything that can be clearly specified. The problem is that most of what enterprises do is not clearly specified.


What Externalized Knowledge Looks Like for AI

When tacit knowledge is successfully externalized for AI consumption, it takes the form of:

  • Behavioral instructions: Not “be helpful” but “when a customer mentions cancellation in a billing dispute, acknowledge the frustration, resolve the billing issue, then proactively offer a retention incentive from the approved list”
  • Decision workflows: Explicit multi-step procedures that encode the judgment calls experts make, including exception paths and edge cases
  • Guardrails: Safety boundaries calibrated by domain context
  • Quality metrics: Codified definitions of what “good” looks like for each interaction type
  • Escalation criteria: Explicit rules for when the AI should hand off to a human, with context requirements for the handoff

This is not a knowledge base. A knowledge base stores information. This is operational knowledge — it encodes how to act, not just what to know.


What This Means for Your Organization

If your AI deployment is stuck at Assist — producing value but not transforming operations — the bottleneck is almost certainly knowledge codification, not technology:

1. Audit your tacit knowledge exposure. For each AI-powered process, ask: what percentage of the knowledge required is documented? If below 70%, the AI will fail on the undocumented cases — which are often the highest-stakes.

2. Invest in externalization before automation. Interview top-performing agents. Document exception paths. Codify the judgment calls. This is not glamorous work, but it is the prerequisite for everything that follows.

3. Package knowledge, not just connect systems. Most AI integration work focuses on connecting APIs and databases. The system connection gives AI access to data. Codified operational knowledge gives AI the ability to act on that data correctly.

4. Treat codification as continuous. Operational knowledge changes. Policies update. New edge cases emerge. The codification investment must be ongoing, with feedback loops from production interactions.

5. Measure the codification gap. Track what percentage of AI interactions require human intervention and categorize the reasons. If the AI escalates because it lacks knowledge (not because the case is genuinely complex), that is a codification gap you can close.


Sources