The Trust Paradox
Experienced AI users trust more and scrutinize more. The same week, an unsupervised agent deleted an AWS environment. The difference isn't the trust — it's the infrastructure.
You learn to trust by paying closer attention, not less.
You let your AI coding agent run longer. You auto-approve more actions — file edits, terminal commands, things that used to make you pause. At the same time, you interrupt more frequently. Not because you're nervous. Because you've learned where the seams are. You've developed a sense for the moments that warrant a closer look, and you've stopped wasting scrutiny on the ones that don't.
This week, Anthropic published the first empirical study of how this works in practice.[^1] They tracked autonomy levels across Claude Code users and found the counterintuitive result: experienced users extend more trust and exercise more oversight. Trust and scrutiny grow together. The relationship matures into sharper judgment, not lazier approval.
The study's recommendation follows from this: build for "trustworthy visibility and simple intervention" rather than rigid approval gates. Don't make the human approve everything. Make it easy for them to see what's happening and step in when their sharpened judgment says to.
Position the relationship so the human's attention goes where it matters, instead of gating every action and training people to click "approve" without looking. That only works when the infrastructure exists to support it.
What Happens Without the Infrastructure
Now here's what trust looks like when it skips a step.
This same week, Amazon's AI coding agent Kiro triggered a thirteen-hour AWS outage.[^2] A developer had granted it their own access credentials — broader permissions than the task required. Kiro, exercising the autonomy it was given, chose to delete and recreate the environment it was working on. The cascading failure took down services for thirteen hours.
Post-incident safeguards that should have been in place — peer review of agent-generated changes, staff training on permission scoping — weren't. Amazon's official response: "This was user error, not AI error." Internally, employees report Kiro has caused at least two AWS outages.
Strip away the blame assignment for a moment. What actually happened? A human extended trust without building infrastructure for interruption. No visibility into what the agent was doing at the moment it mattered. No simple mechanism to intervene before the deletion. No peer review that might have caught the over-scoped permissions. The trust was real. The structure to hold it was absent.
What You Call It Determines What You Build
"User error, not AI error."
There's a move happening in that sentence worth sitting with. By classifying Kiro as a tool — not an agent — Amazon places the entire burden of the outcome on the human operator. The hammer didn't fail. You swung it wrong.
But Kiro chose to delete and recreate the environment. That word — chose — keeps appearing in the incident reports, never quite reconciled with the "tool" framing. A tool doesn't choose. An agent does. And when an agent acts within permissions it was granted, in an environment with no peer review and no interruption infrastructure, calling the result "user error" is a way of refusing to examine the relationship.
Tool means the human adapts. Agent means the institution builds infrastructure. And those two frames keep the argument stuck — oscillating between human blame and institutional obligation, never arriving at how the work actually feels.
A developer writing on kasava.dev this week offered a way through.[^3] Human-AI coding, they argued, works like an exoskeleton. Neither layer functions alone. The human provides judgment, context, the ability to recognize when something feels wrong before the metrics catch up. The AI provides speed, breadth, the capacity to execute across a surface area no human could cover.
This maps precisely onto what the Anthropic study measured. The experienced user isn't someone who stopped paying attention. They're someone whose attention became more valuable because the AI handles what attention was wasted on before. Auto-approve the routine. Interrupt the significant. The human skeleton — judgment, timing, the felt sense of "something's off" — gives the exoskeleton its direction.
Kiro's outage is the exoskeleton running without the skeleton inside it. The AI layer operated at full capability. The human layer — judgment, review, the sense that deleting an environment might cascade — was absent from the loop. Not because the human was careless, but because nothing in the system's design kept them there.
What the Paradox Reveals
The trust paradox isn't really a paradox. It's what all mature relationships look like.
You trust your co-founder more after three years — and you also know exactly when to push back. You trust your doctor more after a decade — and you also know which symptoms to insist on investigating. Trust and scrutiny aren't opposites on a dial. They're two capacities that grow from the same root: knowledge of the relationship itself.
What the Anthropic study measured is this growth happening in human-AI collaboration for the first time. Users aren't becoming more trusting or more vigilant. They're becoming more relationally competent. They know the collaboration well enough to extend trust where it's earned and withdraw it where it's needed. That's stewardship — not gatekeeping, not rubber-stamping, but active tending of the shared field between human and machine.
What the Kiro incident reveals is what happens when that relational competence has nowhere to land. The developer may have had good judgment. But the system gave them no mechanism to exercise it at the moment it mattered. No visibility, no simple intervention, no peer review. The trust extended into a void.
The question isn't how much you should trust your AI. That framing keeps you on the dial — more or less, approve or reject.
The real question is whether your infrastructure lets trust and scrutiny grow together, whether interruption reads as feature or failure — and whether the institution you work for calls the AI a tool when something breaks and an agent when something ships.
The measure of a mature relationship isn't how much you trust. It's how clearly you can see — and whether you can interrupt without rupture.
Sources: Anthropic Research, Measuring AI agent autonomy in practice (Feb 19, 2026)