← Back to Thinking
Agents in Organizations

Trust isn't a setting. It's earned.

Tim Jordan · March 16, 2026 · 5 min read

Every AI agent platform I’ve looked at handles trust the same way. There’s a configuration option, usually a dropdown or a slider, that sets the agent’s autonomy level. Low autonomy means every action gets human approval. High autonomy means the agent acts independently. You pick a setting at deployment time and move on.

That’s not trust. That’s a permission toggle.

How trust actually works in organizations

Think about how it works when you hire someone, on day one they don’t get the keys to the building and full signing authority, they get a desk, a limited set of responsibilities, and someone checking their work, and over the first few weeks they demonstrate competence by making good calls and catching their own mistakes as the oversight gradually decreases and their scope increases.

If they make a significant mistake at week six, the oversight goes back up, not as punishment but as an appropriate response to demonstrated performance, where trust adjusts based on what actually happened, not what you hoped would happen.

Now imagine the alternative, where you hire someone and on day one you set their “autonomy level” to 7 out of 10 based on their resume with no adjustment based on performance, no increase after good work, no decrease after mistakes, just a static number set once and never revisited.

That’s how every AI agent framework handles trust right now. And it’s obviously wrong.

What earned trust looks like in practice

For our agents, trust isn’t a setting in a config file. It’s a property that changes based on what the agent has actually done.

A newly birthed agent starts with high oversight, its outputs get reviewed, its tool usage gets monitored, and its recommendations get validated against human judgment, not because we don’t trust the underlying models but because we don’t yet know how this agent with its specific configuration and context performs in its specific operational environment.

As the agent demonstrates good judgment within its scope, the oversight mechanisms relax, not all at once and not uniformly, so an agent might earn high trust for information retrieval tasks while still requiring oversight for actions with financial consequences.

If something goes wrong, the trust level adjusts, not as punishment but as information, so the system has data about a scenario where the agent’s judgment was insufficient and responds appropriately.

Why static autonomy settings are dangerous

The risk isn’t just that static settings are imprecise. It’s that they create a false sense of safety.

A team deploys an agent at “medium autonomy.” It works fine for three weeks. They conclude it’s trustworthy. But “medium autonomy” doesn’t mean anything specific about what the agent can handle. It means someone made a judgment call at deployment time about how much freedom felt comfortable.

When the agent eventually encounters a scenario outside its demonstrated competence, the static setting provides no protection. It doesn’t know the agent has never handled this type of situation before. It doesn’t know the agent’s past performance was in a different category of work. It just knows the slider is at 5.

Earned trust handles this differently. Because the trust level is tied to demonstrated performance in specific categories, the system can distinguish between “the agent has proven it handles customer queries well” and “the agent has never handled financial reconciliation before.” The first gets earned autonomy. The second gets oversight by default.

The organizational parallel that keeps coming back

I keep returning to this comparison because it’s not a metaphor, it’s a direct structural parallel, where organizations that delegate effectively use conditional trust by giving people authority within demonstrated domains of competence, monitoring new domains more closely, and adjusting based on outcomes.

Organizations that delegate poorly use one of two approaches, either micromanaging everything which is expensive and demoralizing, or handing off authority in bulk which is efficient until someone makes a catastrophic mistake in a domain they weren’t qualified for.

AI agents face the same dynamic, where permanent full oversight is expensive and defeats the purpose of having agents, while static high autonomy is efficient until the agent does something you didn’t anticipate.

Graduated, earned trust is the middle path that actually works. It’s more complex to build. It requires tracking performance, categorizing types of decisions, and adjusting oversight dynamically. But it produces agents you can actually rely on, because “rely on” means something specific: “demonstrated competence in this type of work, with appropriate escalation for unfamiliar territory.”

This is governance, not AI

I want to name something that I think the AI industry consistently gets wrong. Trust management isn’t an AI problem. It’s a governance problem. The solution doesn’t come from better models or smarter prompts. It comes from organizational design.

Every functional company already knows how to manage trust by promoting people based on performance, giving stretch assignments with appropriate support, establishing escalation paths for decisions above someone’s authority level, and reviewing outcomes to adjust scope accordingly.

All of this translates directly to AI agents. We don’t need to invent new trust mechanisms. We need to apply the ones that organizations have been refining for centuries.

The implementation is technical. The framework is organizational. And the sooner the AI industry recognizes that, the sooner we’ll have agents that people actually trust. For real, not because of a slider.

← Back to Thinking