Why governance has to go inside the system – O’Reilly

by
0 comments
Why governance has to go inside the system - O'Reilly

For much of the last decade, AI governance remained comfortably outside the systems that were supposed to regulate it. Policies were written. Reviews were conducted. The models were approved. The audit took place after the fact. As long as AI behaved like a tool – generating predictions or recommendations on demand – the separation mostly worked. That notion is being broken.

As AI systems move from supporting components to autonomous actors, externally imposed governance is no longer the norm. The problem is not that organizations lack policies or oversight frameworks. This means that those controls are abstracted away from where decisions are actually made. Increasingly, the only place where governance can work effectively is inside the AI ​​application itself, at runtime, while decisions are being made. This is not a philosophical change. This is an architecture.

When AI silently fails

One of the more troubling aspects of autonomous AI systems is that their most consequential failures rarely look like failures at all. Nothing crashes. Latency remains within limits. Logs look clean. The system behaves coherently—just not correctly. An agent advances the workflow that should have been contained. A recommendation gradually moves away from the policy intention. A device is applied in a context that no one has explicitly approved, yet no explicit rules are violated.

These failures are difficult to detect because they emerge from behavior, not bugs. Traditional governance mechanisms do not help much here. Pre-deployment reviews assume that the decision path can be predicted in advance. Static policies assume that behavior is predictable. Post hoc audit assumes that intent can be reconstructed from the output. Once the system reasons dynamically, retrieves context opportunistically, and acts continuously, none of those assumptions hold. At that point, the rule is not missing – it is simply in the wrong place.

No one has scaling problems

Most organizations already feel this tension, even if they don’t describe it in architectural terms. Security teams tighten access controls. Compliance teams expand review checklists. Platform teams add more logging and dashboards. Product teams add additional hurdles quickly. Each layer helps a little. None of them address the underlying issue.

What is really happening is that governance responsibility is being divided between teams that do not have end-to-end ownership of system behavior. No single layer can explain why the system worked – only that it worked. As autonomy increases, the gap between intention and execution widens and accountability diffuses. This is a classic scaling problem. And like many scaling problems before it, the solution isn’t more rules. It is a different system architecture.

A familiar pattern from infrastructure history

We have seen this before also. In early networking systems, control logic was tightly coupled with packet management. As the network grew, this became unmanageable. Separating the control plane from the data plane allowed policy to evolve independently of traffic and made failures diagnosable rather than mysterious.

Cloud platforms also went through a similar change. Resource scheduling, identification, quotas, and policy moved out of application code and into shared control systems. That separation is what makes hyperscale clouds viable. Autonomous AI systems are approaching a comparable inflection point.

Right now, governance logic is scattered across signals, application code, middleware, and organizational processes. None of these layers were designed to consistently claim authority over the system’s logic and function. There is a lack of a control plane for AI – not as a metaphor but as a real architectural limitation.

What does “governance within the system” really mean?

When people hear “governance inside AI,” they often imagine strict rules turning into signals or more conservative model constraints. That’s not what this thing is about.

Establishing governance within the system means separating decision execution from decision authority. Execution includes inference, retrieval, memory update, and tool invocation. Authorization includes policy evaluation, risk assessment, permitting and intervention. In most AI applications today, those concerns are implicit – or worse, implicit.

A control-plane-based design makes that separation explicit. The execution proceeds but under constant surveillance. Decisions are seen as they are made, not second-guessed after the fact. Constraints are evaluated dynamically, not considered ahead of time. Governance stops being a checklist and starts behaving like infrastructure.

Figure 1. Separating execution from governance in autonomous AI systems

Logic, retrieval, memory, and tool invocations operate in the execution plane, while a runtime control plane continuously evaluates policy, risk, and authorization—observing and intervening without embedding it in application logic.

Where the rule breaks first

In practice, governance failures in autonomous AI systems accumulate on three surfaces.

logic. Systems create intermediate goals, weigh alternatives, and make branching decisions internally. Without visibility into those pathways, teams can’t differentiate acceptable variation from systemic drift.

restoration. Autonomous systems pull context opportunistically. That context may be out of date, inappropriate, or out of scope – and once it enters the reasoning process, it effectively becomes invisible unless it is explicitly tracked.

action. The use of the tool is where the intention becomes the effect. Systems are increasingly invoking APIs, modifying records, triggering workflows or escalating issues without human review. Static authority models do not map clearly onto dynamic decision contexts.

These surfaces are interconnected, but they fail independently. Treating governance as a monolithic concern leads to poor design and false confidence.

Control plane as runtime feedback system

A useful way to think about AI control planes is not as gatekeepers but as feedback systems. Signals flow continuously from execution to governance: declining confidence, exceeding policy limits, recovery drift, and action escalation patterns. Those signals are evaluated in real time, not weeks later during an audit. Responses returned: throttling, interference, surge, or barrier adjustment.

This is fundamentally different from monitoring output. Output monitoring tells you what happened. Control plane telemetry tells you why this was allowed to happen. This difference matters when the system works consistently, and results add up over time.

Figure 2. Runtime governance as a feedback loop

Behavioral telemetry flows from execution to the control level, where policy and risk are continuously evaluated. Enforcement and intervention return to execution before failures become irreversible.

Want radar delivered straight to your inbox? Join us on Substack. Sign up here.

A failure story that should sound familiar

Consider a customer-support agent who works on billing, policy, and CRM systems.

Over several months, policy documents are updated. Some are quickly re-indexed. Others lag behind. The agent continues to coherently retrieve context and reason, but its decisions increasingly reflect old rules. None of the actions directly violate the policy. The metrics remain stable. Customer satisfaction is gradually decreasing.

Ultimately, an audit flags non-compliant actions. At that time, the teams scrimmage. Logs show what the agent did but not why. They cannot reconstruct which documents influenced which decisions, when those documents were last updated, or why the agent believed his actions were lawful at that time.

This is not a logging failure. This governance lacks feedback loops. A control plane won’t prevent every mistake, but it will surface early – when intervention is still cheap.

Why can’t external governance catch up?

It’s tempting to believe that better tooling, stricter reviews, or more frequent audits will solve this problem. They won’t.

External governance operates on snapshots. Works on autonomous AI stream. The mismatch is structural. By the time an external process notices a problem, the system has already progressed—often repeatedly. This does not mean that governance teams are failing. This means they are being asked to regulate systems whose operating models have superseded their devices. The only viable option is governance which runs at the same pace as execution.

Authority, not mere observability

A subtle but important point: control planes aren’t just about visibility. They are about authority.

Observability without enforcement creates a false sense of security. Seeing a problem after it occurs does not prevent it from occurring again. Control planes must be able to stop, redirect, interrupt, or escalate behavior in real time.

This raises uncomfortable questions. How much autonomy should the system retain? When should humans intervene? How much latency is acceptable for policy evaluation? There are no universal answers. But those trade-offs can only be managed if governance is designed as a first-order runtime concern, not an afterthought.

architectural changes ahead

The move from guardrails to control loops reflects earlier changes in infrastructure. Every time, the lesson was the same: Static rules do not underlie dynamic behavior. Gives feedback.

AI is now entering that phase. The system of governance will not disappear. But its form will change. It will run inside the system, work continuously, and take over at runtime. Organizations that treat this as an architectural problem – not as a compliance exercise – will adapt faster and fail more gracefully. Those who don’t will spend the next few years chasing phenomena they can see, but never explain.

closing thoughts

Autonomous AI does not need less governance. This requires governance that understands autonomy.

This means moving policies beyond documents and audits into events. This means designing systems where authority when making decisions is clear, observable, and enforceable. In other words, governance should become part of the system—not something imposed upon it.

Further reading

Related Articles

Leave a Comment