The AI security conversation has fundamentally changed. In 2024, it was about jailbreaks and prompt injection demos at conferences. In 2026, it is about autonomous agents making consequential decisions across production systems. The threat model evolved. The defences need to evolve with it.
Two years ago, the security community was focused on what models could be tricked into saying. Researchers competed to extract system prompts. Red teams published creative jailbreaks. The discourse centred on the model as the unit of risk.
That framing was never wrong, but it was always incomplete. Models are components. Agents are systems. And in 2026, agents are the thing that enterprises are actually deploying. Not chatbots behind a text box. Autonomous programs that read data, make decisions, call APIs, move money, and interact with other agents. The security surface is no longer the prompt. It is the entire execution graph.
If you are a CISO at a regulated enterprise, the question is no longer "should we worry about AI security?" That debate ended sometime in 2025. The question is: what specifically should you be doing about it, right now, with finite budget and finite staff?
This post is my attempt at a concrete answer. Not theory. Not a framework diagram. A prioritised list of what matters, what most teams are getting wrong, and what the regulatory timeline is actually forcing.
The threat model shifted
The AI security threat model of 2024 was largely about model attacks. Prompt injection. Data poisoning. Training data extraction. Jailbreaks. These are real attack vectors and they have not gone away. But they are now a subset of a much larger problem.
The shift happened because the unit of deployment changed. In 2024, most enterprises were deploying models behind wrappers. A user types a question, the model generates an answer, a human reads it. The blast radius of a compromised interaction was limited to the quality of the text output.
In 2026, the unit of deployment is the agent. And agents are categorically different from models in ways that matter for security.
Agents have identity. They authenticate to systems, hold credentials, and act on behalf of users or other services. A model does not have identity. It has an API key. An agent has a role in your organisation, whether you have formalised that or not.
Agents have permissions. They can read from databases, write to APIs, send communications, and trigger workflows. The scope of what they can do is defined by the tools they have access to and the credentials they hold.
Agents have persistent state. They accumulate context across sessions. They remember previous interactions. They maintain working memory. This means a single compromised session can affect future sessions, and data from one context can leak into another.
Agents chain actions. A model call is atomic. An agent execution is a sequence of decisions, each building on the last, each potentially expanding the scope of impact. The risk is not in any individual step. It is in the compound effect of the sequence.
Securing the model is necessary. It is not sufficient. If your AI security strategy begins and ends at the model layer, you are defending the engine while the car drives itself off a cliff.
What AI security actually means in 2026
I want to be specific about what a mature AI security posture looks like today. Not aspirational. Not what you might need in two years. What you need now, in production, for agents that are already running.
It comes down to five dimensions. Miss any one of them, and you have a gap that an attacker, an auditor, or an incident will find for you.
Agent identity
Every agent in your environment needs a cryptographic identity. Not a shared API key. Not a service account that six agents share. A unique, verifiable, revocable identity per agent instance.
The standard here is moving toward SPIFFE (Secure Production Identity Framework for Everyone). SPIFFE gives you workload identity that works across environments, without relying on network-level controls or shared secrets. Each agent gets a SPIFFE ID. Each SPIFFE ID maps to a set of permissions. If an agent identity is compromised, you revoke that specific identity without affecting other agents.
If your agents are authenticating to production systems with shared API keys or static tokens, you have the AI equivalent of everyone in the office sharing one password. It works until it doesn't, and when it doesn't, you cannot determine which agent did what.
Runtime policy enforcement
Policy that exists in a document is not policy. Policy that is enforced at runtime is policy.
This means having a governance layer that intercepts agent actions before they execute and evaluates them against defined rules. Not after the fact. Not in a weekly review. In the execution path, at the moment the agent attempts to act.
Runtime enforcement covers questions like: Is this agent allowed to call this tool? Is this agent allowed to access this data classification? Has this agent exceeded its action budget for this session? Is this sequence of actions consistent with the agent's defined scope?
The analogy is a firewall, but for agent behaviour. You define what is permitted, and everything else is denied by default. The agent can operate freely within its authorised boundaries and is stopped the moment it tries to exceed them.
Most organisations today have documentation that describes what agents should do. Very few have infrastructure that ensures agents can only do what they should. The gap between those two states is where incidents live.
Data flow governance
Agents consume and produce data. That data has classifications. PII. Financial records. Health information. Trade secrets. Attorney-client privileged communications. The data does not lose its classification because an AI agent is processing it.
Data flow governance means classifying data at the pipeline level, not just at rest. When an agent reads from a database, the governance layer needs to understand what data classifications are present in the response. When an agent passes data to another agent or to an external API, the governance layer needs to verify that the destination is authorised to receive that classification of data.
This is not a novel requirement. Data loss prevention has existed for decades. But the agent architecture introduces new data flow patterns that traditional DLP tools were not designed for. Data moves between agents, between tools, between contexts. It can be transformed, summarised, embedded in prompts, and sent to model providers. Every one of those transitions is a potential data governance violation.
Tool permission management
The principle of least privilege is not new. Applying it to AI agents is. Every agent should have access to exactly the tools it needs for its defined function, and nothing more. An agent that processes invoices needs access to the accounting API. It does not need access to the HR system, the email gateway, or the source code repository.
But it goes deeper than just tool-level permissions. You need operation-level permissions per tool. An agent might need read access to a database but not write access. It might need to send emails to internal addresses but not external ones. It might need to query an API but not modify configurations.
The right model is zero trust applied to AI agents. No implicit trust. Every action verified. Every permission explicitly granted and auditable.
Tamper-evident audit trails
Logs are not audit trails. An audit trail is a tamper-evident, chronologically ordered record of every action an agent took, every decision point it encountered, every piece of data it accessed, and every tool it invoked. It includes the policy evaluation that permitted or denied each action. It includes the full context of the agent's reasoning at the time of action.
Tamper-evident means the audit trail itself cannot be modified after the fact. Not by the agent. Not by an administrator. Not by an attacker who has compromised the agent's host. Append-only, cryptographically chained, stored independently of the systems being audited.
When a regulator asks "what did this agent do with customer data on March 15th," you need to be able to answer that question completely, accurately, and provably. "We have logs" is not the same as "we have an audit trail." Logs can be altered. Audit trails, properly implemented, cannot.
What most security teams are still getting wrong
I talk to security teams at regulated enterprises every week. The same three mistakes come up repeatedly. None of them are mistakes of negligence. They are mistakes of framing. Teams are applying the mental models they have, but the mental models do not fit the new architecture.
Treating model provider guardrails as sufficient
This is the most common and the most dangerous. The reasoning goes: "We use Anthropic. Anthropic has safety training, content filtering, and usage policies. Therefore our AI deployment is secure."
It is not. The model provider secures the model. They do not secure your agent. They cannot, because they have no visibility into what your agent does with the model's output. They do not know what tools your agent has access to. They do not know what data your agent can read. They do not know what actions your agent can take.
Model provider guardrails are a foundation. They are not a ceiling. If your security review stops at "which model provider are we using and what are their safety measures," you have reviewed one component of a multi-component system and declared the system secure.
Auditing quarterly instead of in real time
Traditional security audits operate on quarterly or annual cycles. You review access logs, check compliance postures, verify that policies are being followed. This cadence made sense when the systems being audited were relatively static. Access permissions changed slowly. Configurations were updated deliberately.
Agent behaviour is not static. An agent can take hundreds of actions per hour. Its behaviour can change based on the data it encounters, the prompts it receives, and the context it accumulates. A quarterly audit of an agent system is like reviewing a week's worth of network traffic by looking at one packet capture from January.
Real-time monitoring is not optional for agent governance. You need continuous visibility into what agents are doing, with alerting that triggers when behaviour deviates from the expected pattern. The shadow AI problem compounds this. You cannot monitor agents you do not know about.
No inventory of deployed agents
Ask a CISO how many AI agents are running in their environment. Most cannot answer the question. Not because they are negligent, but because there is no mechanism in place to track agent deployments the way there are mechanisms to track server deployments or application deployments.
Developers spin up agents using frameworks that run as standard processes. They authenticate using existing service accounts. They access data through existing APIs. From an infrastructure perspective, an agent looks like any other application. There is no registry. There is no discovery mechanism. There is no central inventory.
This is the shadow IT problem, accelerated. Shadow IT meant someone spun up an unapproved SaaS tool. Shadow AI means someone deployed an autonomous program with access to production data and the ability to take actions, and nobody outside their team knows it exists.
You cannot secure what you cannot see. The first step, before any of the other four dimensions, is knowing what agents exist in your environment.
The regulatory forcing function
If the security arguments are not sufficient to drive action, the regulatory timeline will be. Three regulatory frameworks are converging on AI agent governance in 2026, and the compliance deadlines are not theoretical.
The EU AI Act reaches its August 2026 deadline for high-risk AI systems. If your agents operate in healthcare, financial services, legal, HR, or any of the other high-risk domains defined in Annex III, you need conformity assessments, risk management systems, human oversight mechanisms, and technical documentation. The EU AI Act compliance timeline is not forgiving. Organisations that have not started building compliance infrastructure by now are already behind.
DORA (Digital Operational Resilience Act) is treating AI agents as ICT assets. This means they fall under the same operational resilience requirements as your other critical technology: testing, incident reporting, third-party risk management. If an AI agent is part of a critical business process, DORA requires you to demonstrate that it is operationally resilient. That requires monitoring, testing, and governance infrastructure that most agent deployments do not have.
NIS2 incident detection requirements apply to any significant incident affecting network and information systems. An AI agent that exfiltrates data, makes unauthorised decisions, or behaves anomalously is a NIS2-reportable incident if it affects essential or important entities. The 24-hour early warning and 72-hour notification requirements mean you need detection infrastructure that operates in real time. You cannot comply with NIS2 incident reporting timelines if you do not have real-time visibility into agent behaviour.
The regulatory pattern is clear. Regulators have decided that AI agents are not exempt from existing governance requirements. They are subject to them, and in many cases subject to additional requirements specific to AI. The organisations that build this infrastructure before the mandates take effect will have it running and tuned when their competitors are scrambling to stand something up under deadline pressure.
What to prioritise now
If you are a CISO reading this in March 2026, here is what I would do, in order. Not everything at once. Sequenced by dependency and impact.
First: inventory your agents. Conduct a discovery exercise across your organisation. Find every AI agent that is running in production, staging, or development. Document what each agent does, what tools it has access to, what data it can reach, what credentials it holds, and who is responsible for it. You will find agents you did not know about. That is the point.
Second: implement runtime governance. Put a governance layer in the execution path of your agents. This is the single highest-leverage investment you can make. A runtime governance layer gives you enforcement (agents can only do what they are authorised to do) and visibility (you can see what every agent is doing in real time). Both are preconditions for everything else on this list.
Third: establish tamper-evident audit trails. Instrument your agent infrastructure to produce complete, tamper-evident records of every action. This is not just a security measure. It is a compliance requirement under the EU AI Act, DORA, and NIS2. Start now, because retroactive audit trails do not satisfy regulators.
Fourth: define tool boundaries. For every agent in your inventory, define and enforce the minimum set of tools and permissions it needs. Remove everything else. This is the least-privilege principle applied to agent architecture. It reduces blast radius. It limits the damage from any single compromised agent. And it forces your teams to articulate exactly what each agent is supposed to do, which turns out to be valuable in itself.
Fifth: prepare for compliance deadlines. Map your agent deployments against the EU AI Act risk classifications. Identify which agents fall into high-risk categories. Begin the conformity assessment process for those agents. Build the technical documentation that the Act requires. If you are subject to DORA, include your AI agents in your ICT risk management framework and operational resilience testing.
This is not a twelve-month plan. Steps one and two should be underway within weeks, not months. The regulatory deadlines are fixed. The threat landscape is not waiting for your next quarterly planning cycle.
AI security in 2026 is not about preventing a chatbot from saying something embarrassing. It is about governing autonomous systems that operate across your production infrastructure with real credentials and real authority to act. The organisations that internalise this distinction and build accordingly will navigate what comes next. The organisations that treat agent security as an extension of model safety will learn the difference the hard way.
The good news is that the principles are not new. Identity. Least privilege. Audit trails. Real-time monitoring. Policy enforcement. We have been doing this for decades with other categories of systems. The challenge is applying those principles to a new architecture before the architecture outpaces the governance. That window is closing.
See governance at runtime
TapPass is in private beta. If your team is shipping AI agents, we'd rather get you on the product than in a pipeline.