← Back to Thinking
Agents in Organizations

The problem with 'autonomous agents'

Tim Jordan · March 16, 2026 · 5 min read

Every demo I see pitches the same vision: an AI agent that can do everything on its own, where you give it a goal and walk away and come back to a finished product with no human in the loop and no oversight needed and full autonomy, and the demos are impressive and the pitch is compelling but the fundamental premise is wrong.

Autonomy isn’t the goal

I spent 25 years in business before touching AI, working with hundreds of employees and contractors and partners across 35+ countries, and I never once had a team member who was fully autonomous nor would I want one because autonomy in an organization doesn’t mean “does everything independently” but “makes appropriate decisions within a defined scope and escalates when needed and remains accountable for outcomes,” which is a very different thing.

A fully autonomous employee who never checks in and never escalates and never adjusts based on feedback isn’t a high performer, they’re a liability, and the same is true for AI agents.

What happens when autonomy has no structure

I’ve watched the demo-to-deployment gap in the AI agent space, and the pattern is consistent where the demo shows an agent handling a clean, well-defined task with obvious boundaries while deployment encounters ambiguity, edge cases, conflicting priorities, and situations where the “right” answer depends on organizational context the agent doesn’t have.

Without governance structures, the autonomous agent does one of two things, either it makes a confident decision that turns out to be wrong because it lacked context, or it gets stuck in a loop trying to reason its way through a problem that requires organizational judgment it doesn’t possess.

Both failure modes come from the same root cause where the agent has capability without structure, so it can reason but has no framework for deciding when to act versus when to escalate, and it can use tools but has no model for which decisions are within its authority and which aren’t.

The three things autonomy actually needs

When I think about what makes an organizational member effective with or without AI three structural elements keep showing up, and identity means the agent knows what it is and what role it plays and where it fits in the organizational context not as a system prompt but as a persistent definition that shapes every decision it makes so an agent with identity doesn’t need to be told “you’re a customer operations specialist” every conversation because it knows the way an employee knows their job title and what it means.

Governance means there are clear boundaries on what the agent can and can’t do, not guardrails that prevent specific bad outputs but structural constraints that define the agent’s authority by clarifying what decisions it can make independently and what requires escalation and what’s completely outside its scope where these constraints don’t reduce the agent’s value but make it predictable.

Accountability means the agent’s decisions have consequences where good decisions build trust and expand autonomy while bad decisions trigger review and potentially reduce scope, creating a feedback loop where the agent’s behavior improves over time not because someone rewrites the prompt but because the organizational structure incentivizes good judgment.

Why “human in the loop” isn’t enough

The standard response to the autonomy problem is “just keep a human in the loop,” reviewing every output and approving every action while the agent does the work and the human does the oversight, but this works at small scale and falls apart at the scale where agents are actually valuable because if you have 5 agents handling 10 tasks a day human review is feasible but if you have 50 agents handling 500 tasks a day human review becomes the bottleneck and the entire efficiency argument for AI agents collapses.

The solution isn’t more humans reviewing more outputs but governance structures that make the right level of oversight automatic, so some tasks get reviewed every time and some tasks get spot-checked and some tasks run independently because the agent has earned trust for that category of work, and that’s not removing the human from the loop but making the loop intelligent by matching oversight intensity to actual risk instead of applying the same level of review to everything.

The race to the wrong finish line

The companies racing to build “fully autonomous” agents are solving a problem that organizations don’t actually have because no organization wants a team member that’s fully autonomous but a team member that’s appropriately autonomous, that handles routine work independently and escalates the hard stuff and gets better over time, and building that requires architecture that most agent frameworks don’t even attempt with identity that persists across sessions and governance that adapts based on context and accountability that creates learning loops and trust that’s earned rather than configured.

It’s not as exciting as a demo where an agent does everything by itself, it’s how actual organizations work, and it’s what’s going to matter when the demos stop and the real deployments start.

← Back to Thinking