Security leaders need a way to measure where they stand on AI agent governance. Not a maturity model with five levels of aspirational descriptions. A practical scoring framework that produces a number, identifies gaps, and points to specific actions.

This is a framework I've developed from conversations with security teams at regulated European enterprises. It has eight dimensions, each scored 0 to 4. The total score ranges from 0 to 32. It is not comprehensive enough to be a formal standard and not shallow enough to be marketing. It sits in between: a useful starting point for a conversation between the CISO and the board, or between the security team and the engineering teams deploying AI agents.

The eight dimensions

Dimension 1

Inventory completeness

Do you know how many AI agents are running in your organization?

Dimension 2

Identity and access control

Can you distinguish one agent from another? Can you scope their permissions individually?

Dimension 3

Runtime monitoring

Can you see what your agents are doing right now?

Dimension 4

Policy enforcement

Are agent permissions enforced at runtime, or just documented?

Dimension 5

Audit trail quality

Could you reconstruct exactly what an agent did last Tuesday?

Dimension 6

Human oversight

Can a human intervene in agent operations when needed?

Dimension 7

Incident response

What happens when an agent does something wrong?

Dimension 8

Third-party risk management

How well do you manage the risks from model providers and AI tool vendors?

Interpreting the score

0-8: Unprotected. AI agents are operating without governance. The organization is exposed to financial, regulatory, and reputational risk from AI agent behavior. Immediate action required: start with inventory and identity.

9-16: Emerging. Some governance exists but significant gaps remain. Common pattern: monitoring exists but enforcement doesn't. Or policy exists but isn't enforced at runtime. Priority: close the enforcement gap and build the audit trail.

17-24: Established. Core governance controls are in place. The organization can monitor, control, and audit its AI agents. Remaining gaps are typically in incident response and third-party risk management. Priority: test the controls under realistic conditions.

25-32: Advanced. Comprehensive AI agent governance with dynamic policy, automated response, and auditor-ready evidence. The organization is positioned for regulatory compliance and can demonstrate governance maturity to customers and partners.

Where most organizations score

Based on conversations with security teams at 20+ European enterprises, the typical score is between 4 and 10. The most common pattern:

That totals 6 to 8 out of 32. Unprotected. This is not because these teams are negligent. It is because AI agent governance is a new discipline and the tooling is still emerging. Most organizations are doing what they can with what they have. The framework helps identify where the gaps are largest and where investment will have the most impact.

Using the score

The score is useful in three contexts.

Board reporting. A single number that communicates the organization's AI governance posture. "We scored 7 out of 32 on the AI readiness assessment. Our target is 20 by Q4 2027. Here's the plan." This is the kind of communication boards can act on.

Prioritization. The dimension with the lowest score is usually the highest priority. An organization that scores 3 on monitoring but 0 on enforcement should focus on enforcement. Monitoring without enforcement means you can see the problem but can't stop it.

Vendor evaluation. When evaluating AI governance tooling, map the vendor's capabilities to the eight dimensions. Which dimensions does the tool address? Which does it leave unaddressed? This prevents buying a monitoring tool when the gap is in enforcement, or an audit tool when the gap is in identity.


This framework is not a standard. It is a starting point. The dimensions may evolve as the regulatory landscape matures and as AI agent architectures change. But the underlying question is permanent: do you know what your AI agents are doing, can you control it, and can you prove it? The score gives you a number. The dimensions give you a map.

Get your readiness score

We'll walk through the 8 dimensions with your team and identify the highest-impact gaps. 20 minutes. Honest assessment.

Book a demo