View all research work

Back

March 9, 2026

Most Agentic AI Never Makes It Past the Demo

Mark Esposito

,

Chief Economist at micro1

There is a version of agentic AI that looks extraordinary. The demos are compelling, the walkthroughs are persuasive, and the possibility space feels genuinely limitless. Then someone tries to deploy the thing in a real organization, and reality intervenes.

This is not a niche failure mode. It is the dominant experience of enterprises engaging with agentic systems today. The gap between what an agent can do in a controlled pre-production setting and what it actually does when exposed to live infrastructure, real APIs, and unpredictable users is not a minor calibration problem. It is a structural one. And until the field takes it seriously, the promise of agentic AI will continue to outrun its delivery.

The Problem Is Not the Model

The instinct when an agent fails in production is to blame the model. If the underlying system were smarter, the thinking goes, it would handle the edge cases better. But that instinct is usually wrong.

In controlled environments, agents are evaluated against synthetic conditions. APIs respond reliably. Schemas stay consistent. Permissions are clear. None of that holds once you move into production. APIs time out. Schemas change without notice. Inputs arrive malformed, incomplete, or adversarially crafted. The agent that performed beautifully in staging now drifts, fails silently, or compounds errors across a multi-step pipeline.

What we are dealing with, in most cases, is a systems problem masquerading as a model problem. The environment is what the agent was never designed for.

The Demo-to-Deployment Gap Has a Robotics Equivalent

The same failure pattern runs through physical AI. We have all watched the videos: robots performing flawlessly in lab conditions, navigating obstacles, sorting packages. Then someone introduces a new floor surface or an unexpected object, and the system collapses.

The underlying logic is identical. Pre-deployment testing creates an illusion of robustness by eliminating the variables that matter most in the real world. Whether the system is digital or physical, the moment you expose it to genuine environmental complexity, you find out what it actually knows. The lesson is not that deployment is impossible. It is that we have been measuring the wrong things and preparing for the wrong conditions.

Stress Testing Is Not a One-Time Event

Current practice in most organizations treats stress testing as a pre-deployment checkpoint. You run the agent against a set of scenarios, it passes, and you ship it. That approach is inadequate for systems operating in dynamic environments, which is essentially all of them.

The environments agents operate in are not static. Enterprise data drifts. User behavior evolves. Tool interfaces change. An agent certified as production-ready in January is operating in a meaningfully different environment by April. If your evaluation framework does not account for that, you are not evaluating the deployed system. You are evaluating a snapshot of it.

What is needed is continuous monitoring, not periodic audits. The evaluation infrastructure has to run alongside the system in production, catching failure modes as they emerge rather than after the damage is done. This shifts evaluation from a gate into a practice, which is a significant lift, but it is the only approach that reflects how these systems actually behave over time.

Benchmarks Are Not Reality

Benchmarks measure performance in isolation. They are designed to be reproducible and comparable, which means they are, by construction, stripped of the contextual complexity that characterizes real deployments.

When a model scores well on a benchmark, it tells you something useful about a narrow slice of capability. It tells you very little about how that model will behave as the reasoning core of a multi-tool agent interacting with an enterprise system under production conditions. Less than one percent of enterprise data has been incorporated into the frontier models we are deploying at scale. Models that perform well on generic tasks may perform quite differently when asked to handle the specific workflows and institutional knowledge of a given organization. The benchmark score does not transfer, and assuming it does is one of the more costly mistakes I see organizations make.

Human Oversight Is Strategic, Not Universal

The debate about humans in agentic pipelines is often framed in binary terms: either you have a human in the loop or you do not. That framing is too crude.

The more productive question is where human judgment adds irreplaceable value, and where it becomes a bottleneck that limits scalability. For deterministic tasks with verifiable outputs, automated evaluation can and should carry the weight. For tasks involving ambiguous judgment, domain-specific nuance, or high-stakes consequences, human oversight is not a concession. It is the appropriate design choice.

The goal is not to minimize human involvement as a matter of principle. It is to deploy human expertise where it is genuinely irreplaceable, and build automated infrastructure capable of handling everything else.

What Has to Be True for This to Work

Responsible deployment of agentic AI is not primarily a model problem or even a data problem. It is an infrastructure and design problem. Organizations that get this right will have built continuous monitoring frameworks capable of detecting drift, tool use failures, and unexpected input patterns in real time. They will have evaluation pipelines that distinguish between reasoning validity and agreement with a predetermined answer. And they will have human expertise positioned at the failure modes that genuinely require it, not spread thin across the entire system.

The agentic internet being built right now is not going to be won by the organizations with the most impressive demos. It will be won by those who solve the unsexy, difficult, infrastructure-level problems that separate a compelling proof of concept from a system people can actually trust.

The distance between those two things is larger than most organizations currently appreciate. Closing it is the work ahead.

This piece draws on a conversation I recently hosted for the micro1 Virtual Series with Jason Mayes (WebAI Lead, Google) and Imran Nasim (AI Researcher, micro1). Their insights sharpened much of the thinking here.

Mark Esposito

Dr. Mark Esposito is a public policy scholar and social scientist affiliated with Harvard’s Berkman Klein Center for Internet and Society and the Center for International Development at Harvard Kennedy School. He leads policy clinics on the governance of technology worldwide. He is a Professor at Hult International Business School and Adjunct Professor at Georgetown University. He has co-founded several AI ventures, including Nexus FrontierTech, the AI Native Foundation, and The Chart ThinkTank, and serves as Chief Economist at micro1, a Silicon Valley–based AI lab. He is a member of the World Economic Forum’s Global AI Alliance, a Senior Advisor at Strategy& (PwC), a professorial fellow of the Mohammed Bin Rashid School of Government, and the co-author of 14 books.