The Agentic Shift: When Agents Act, Who Can Stop Them?


Sunday 26th April 2026

The Agentic Shift: When Agents Act, Who Can Stop Them?

Enterprises are accelerating autonomy, but most have not engineered the circuit breaker: authority, promises, and evidence.

Hi, welcome to the Trusted Agents Situation Room. As AI systems move from assisting to acting, most enterprises are scaling decision capacity faster than they are engineering override authority, contractual duties, and audit-grade evidence.

We help risk, compliance, and governance leaders see where agentic systems will surface hidden weaknesses in data, process, and policy, and what controls need to be built before those systems scale. Because by the time this becomes obvious, it will already be late.

In 20 seconds

This week’s shift: agentic systems are no longer “helping.” They are coordinating actions that change outcomes. The governance gap is not whether the agent can do the task; it’s whether your organisation can (1) stop it at speed, (2) prove what it was authorised to do, and (3) show an evidence trail when something goes wrong.

What happened

AI-enabled decision systems are now coordinating decisions faster than most executive teams are redesigning who actually has the authority to stop them.

Why it matters

Accountability stays attached to named people, but agentic systems compress the time between signal and action, which means your intervention window shrinks unless it is engineered into the architecture.

The decision it forces

You need an explicit posture for the next 12 months: block agents in high-risk workflows, allow them only through controlled endpoints, build first-party agents with strict guardrails, or prepare for agent-to-agent operations where delegation limits and evidence trails are non-negotiable.


What we’re tracking this week

  • When Scale Outruns Authority by Dr Joanna Michalska. Authority design is lagging decision capacity. If your organisation is accelerating decisions through agents without redesigning who can intervene and how fast, you’re building speed without brakes.
  • Observable Contractual Loyalty by Daniel "Dazza" Greenwood, Commissioned by: Stanford Loyal Agents Initiative. “Loyalty” is becoming an evaluation problem, not a marketing claim. This draft framework shows how to turn fuzzy trust language into explicit duties, scoped authorisations, and tests you can actually run.
  • Assessing Claude Mythos Preview’s cybersecurity capabilities from Anthropic Security. Capability is rising fast enough that “nothing bad has happened yet” is not a strategy. Mythos is a useful glimpse of how quickly agentic tooling can compress both attack and defence cycles, and why governance has to assume speed.
  • How to scale AI with integrity and trust? by Jackson Pek, Chief Corporate & Legal Affairs Officer - Amadeus. Governance at scale means operationalising evidence capture, not writing more policy. Amadeus lays out how they’re aligning to the EU AI Act and scaling Responsible AI through standardised workflows that help teams produce consistent compliance evidence as use cases multiply.

Do you want Situation Room updates delivered to your inbox?

The circuit breaker problem

Most agent programs start with capability. Risk teams should start with control.

Joanna’s point is simple and brutal: decision capacity is accelerating, but authority to intervene is not. If your operating model quietly assumes “a human will catch it,” you have already lost the argument. The system will move faster than the human, and accountability will still land on the human.

Promise theory gives a useful way to make this concrete. In promise theory, an agent is anything that can act, and the system is the network of promises those agents make and keep. An enterprise becomes governable when those promises are explicit, limited in scope, and observable. When promises are vague, or hidden in people’s heads, the system is held together by human restraint and human patching. That does not survive agent speed.

This is where Loyal Agent Evals is particularly helpful for risk and compliance leaders. Greenwood’s point is not “trust the model.” It’s: contract for the behaviour you need, then evaluate it. The report argues that most frontier-model consumer terms allocate reliance and liability risk to users and generally do not assume fiduciary or agency duties, while adjacent SaaS markets (legal-tech especially) often do contract for limited agency and confidentiality-bound intermediary status, which matters for privilege-sensitive workflows.

The practical pattern is deployable this year:

  • Publish explicit duties (a CONTRACT.md analogue)
  • Publish user authorisation preferences (an AUTH_PREFS.md analogue)
  • Architect gates that enforce those duties (not “policy around the system”)
  • Run public evaluations and keep the evidence.

If you are accountable for governance, your “circuit breaker” is not a committee. It is a designed set of promises, permissions, and proof.

Mythos is a capability accelerant for both attackers and defenders

Anthropic’s Mythos Preview write-up is, at minimum, a reminder that security capability is moving quickly, and that frontier models can materially shift the cost and speed of cybersecurity work. Whether you read it as a defensive opportunity or an offensive risk, the governance implication is the same: your control model cannot assume slow adversaries or slow mistakes.

OWASP is the checklist, not the strategy

OWASP’s State of Agentic AI Security and Governance is useful because it names the threat surface in plain terms: memory poisoning, tool misuse, prompt injection, insider amplification, and the need for runtime monitoring and fine-grained access control. It is the “what to instrument” companion to the Loyal Agent Evals “what to contract and test” approach.

Questions to Ask your Peers

If you’re the person expected to “sign off” on an agentic pilot, these are the questions that cut through theatre:

  • Where does this process only work because people quietly patch the gaps?
  • Which decisions rely on judgment that lives in someone’s head rather than in the system?
  • If an agent pushed this workflow to the edge of the rules, what would happen to the customer, the business, and the data?
  • Who can stop the system, and how fast? What is the escalation path at 2am?
  • What are the agent’s explicit duties in this workflow (loyalty, confidentiality, disclosure, confirmation), and where are they written down?
  • What evidence will we have if a regulator asks “why did it do that”?

Where Trusted Agents comes in

When Trusted Agents works with clients, those are exactly the questions we put on the table. We put your processes, customer experience, data, and governance on the spot, then build the execution plan needed to address them.

If you want to push on agentic AI without losing control of what matters, start here and book a 30 minute conversation with us.

Read more

Trusted Agents

An advisory firm specialising in Agentic Commerce, Digital Trust and Customer Empowerment.

Read more from Trusted Agents
The Storefront Is Moving.

Tuesday 13th May 2026 The Storefront Is Moving. Is Your Brand Where Agents Can Find It? Amazon just made agentic shopping the default for hundreds of millions of customers. Walmart's agent is already lifting baskets by 35%. The question is no longer whether delegated buying is coming — it is whether your products, data, and operations are ready to be found, trusted, and executed. The Storefront Is Moving. Photo by Egor Myznik on Unsplash Hi, welcome to the Trusted Agents Situation Room....

The moment between "authorized" and "binding" used to be a human. Now it is nothing.

Sunday10th May 2026 The Checkout Page Is Gone. Nobody Built What Replaces It. Mastercard and Google just open-sourced the first serious answer to a governance gap that is already generating chargebacks, disputes, and liability exposure — and your payment, operations, and risk teams need to understand it now. The moment between "authorized" and "binding" used to be a human. Now it's nothing. Photo by Blake Wisz on Unsplash Hi, welcome to the Trusted Agents Situation Room. In 20 seconds AI...

The moment Jamie and I have been building toward is arriving.

Sunday 2nd May 2026 The Agentic Shift: The Prediction Is Landing How Jamie Smith and I built Trusted Agents around a thesis about agentic commerce, why that thesis is proving right, and what it means for the leaders who need to move their organisations now. Built from both sides. Nearly there. Photo by Mason Kimbarovsky on Unsplash Hi, if you are the person in your organisation who has been asked to make sense of agentic AI before everyone else is ready to act on it, this edition is written...