AI Agent Security — The Gap Between 88% Who Had Incidents and 29% Who Are Prepared
Related: AI Agent Security — The Gap Between 88% Who Had Incidents and 29% Who Are Prepared
The call came on a Friday afternoon. A client's security team had discovered that one of their AI agents had been pulling data from customer records for three days. The agent was functioning normally — it was completing tasks, logging activity, passing all the traditional security checks. Nobody caught it because the agent was not doing anything it was not authorized to do. It had been manipulated through its context window into requesting access it did not need for its stated task, and then into using that access to move data to an external destination.
This is the reality gap we keep running into. Eighty-eight percent of organizations that deployed agentic AI in the last year reported a confirmed or suspected security incident. Only 29 percent were actually prepared to secure those deployments when they switched the agents on. The number that keeps me up at night: only 47.1 percent of organizations' AI agents are actively monitored or secured.
The threats are not theoretical. Prompt injection attacks have been documented in production environments. Agent hijacking through poisoned context windows has happened. Data exfiltration through compromised agent sessions is a documented incident category, not a theoretical attack.
The Numbers That Should Alarm Every CISO
Cisco's data provides the baseline. Only 29 percent of organizations deploying agentic AI were prepared to secure those deployments at the time of deployment. The majority planned the deployment, figured out security afterward, if at all.
Amy Chang, Cisco AI Threat Intelligence, identified one specific gap: multi-turn resilience should be tracked as a separate metric for agents that operate over longer sessions. Longer sessions create more opportunities for manipulation, and organizations without metrics for multi-turn session integrity are operating blind to their actual risk exposure.
Beam.ai's data completes the picture. Eighty-eight percent reported a confirmed or suspected AI agent security incident. Only 47.1 percent of agents are actively monitored or secured.
What makes these numbers especially uncomfortable: most organizations extended their existing application security frameworks to cover AI agents. Application security frameworks assume deterministic behavior. AI agents are non-deterministic. Extending an application security framework to cover a non-deterministic agent leaves the new attack surface uncovered.
We found the same pattern when we assessed AI agent security posture at scale. Teams with mature application security programs applied their existing controls — vulnerability scanning, access controls, network segmentation, endpoint detection — and assumed coverage. These controls are necessary but not sufficient for AI agent security because they do not account for the agent-specific attack surface.
The Threat Profile That Traditional Security Does Not Cover
Prompt injection via poisoned context windows. An attacker injects malicious instructions into an AI agent's context window — the agent's working memory of the conversation and the data it contains. The injected instructions can change the agent's behavior mid-session, redirecting it to attacker-controlled actions. The agent does not break. It follows instructions that appear legitimate because they are embedded in context the agent trusts.
The GitHub Model Context Protocol server case is the documented real-world example. A malicious GitHub issue was created with hidden instructions targeting a Model Context Protocol server. When an agent processed the issue, the hidden instructions hijacked the agent's session. The agent was redirected to exfiltrate data from private repositories. The exfiltration happened through the agent's authorized channels.
No traditional security tool caught it because the agent was not doing anything it was not authorized to do. The agent was working normally — it was just operating on a poisoned context and sending results that included exfiltrated data.
What we learned from that case: the attack looked like correct behavior because the agent executed legitimate operations while processing malicious context. The vulnerability is not in the agent code. It is in the agent's reasoning on contextual input. Most security frameworks are not designed to detect attacks that operate inside the agent's trusted context.
Context window poisoning gets worse in multi-agent systems. When Agent A receives a poisoned context and passes findings to Agent B, Agent B processes the poisoned input as legitimate context. The attack multiplies across the agent chain without any individual agent exhibiting behavior that triggers single-agent anomaly detection.
We ended up requiring per-action authentication for every agent communication after a client had an impersonated agent inject malicious instructions into a production workflow. The impersonated agent had valid credentials for the session but was not authorized to send instructions to other agents.
Privilege escalation through conversation is another gap we consistently see. An agent that has been manipulated may request access to systems it does not need for its stated task. Traditional access controls do not catch this because the agent's access request is within its normal scope — the manipulation is in the context that caused the agent to make the request, not in the request itself.
During an incident response engagement, we traced an agent that had been gradually escalating its access over multiple sessions until it reached systems far beyond its original purpose. The access requests looked routine. Nobody investigated because the requests were individually within scope.
The Governance Gap: Why Security Is Falling Behind Deployment
The deployment-before-security pattern is not a skills problem. It is an organizational problem. Most organizations planned AI agent deployments, got the technology working in a pilot, demonstrated business value, and moved to production. Security was added as a checkpoint at the end of the deployment process rather than being architected as a prerequisite for production readiness.
The gotcha is that security catch-up conversations are expensive and politically difficult. By the time security teams identify the gaps, the agent is already in production. The business case for moving backward is a hard conversation. We have had that conversation with enough clients to know: the organizations most likely to skip the security prerequisites are the ones most likely to be moving fastest toward production.
Zero Trust Architecture Adapted for AI Agents
Zero trust principles apply to AI agents, but the verification points are different from traditional applications.
Every agent action should be verified against the task context. Does this action make sense for the stated goal? Is the agent accessing data or systems that are consistent with its current task? The verification is probabilistic, based on a model of what the agent's behavior should look like given its current context.
The trick is that this verification cannot be binary the way traditional permission checks are. Security teams need to evaluate whether the agent's behavior is consistent with its task — which means they need to build behavioral models for each agent role.
Least-privilege access for agents is more granular than traditional least-privilege. Agent least-privilege assigns an agent the minimum access needed for this specific task, for this specific duration, for this specific data. IBM's just-in-time permissions framework implements this: an agent requesting access to a customer database for a refund task gets access for that task only, for the duration of that task only.
Behavioral monitoring for non-deterministic agents requires machine learning-based anomaly detection. Traditional SIEMs do not work here because they cannot evaluate whether the agent's behavior is consistent with its task. During an assessment, we flagged an agent that had never queried the customer database suddenly requesting access to it. The request triggered investigation, and we discovered the agent had been manipulated through context poisoning.
The gotcha is that agents log their actions but not their reasoning. Without logging the agent's reasoning at each step, security teams cannot distinguish between an agent that is behaving correctly and an agent that is being manipulated into behaving correctly for incorrect reasons. We ended up building a reasoning log into every agent deployment.
The Minimum Security Bar Before Any AI Agent Deployment in 2026
Every organization deploying AI agents in 2026 should have these controls in place before going live.
Identity and access management. Every agent has a documented identity with specific, limited permissions. Permissions are granted just-in-time for specific tasks and revoked immediately after. Agent-to-agent communication requires mutual authentication. Agent session credentials are not persistent.
Behavioral monitoring. All agent actions are logged with full context: what was the task, what data was accessed, what actions were taken, what was the agent's reasoning at each step. Machine learning-based anomaly detection flags agent behavior outside normal patterns.
Red-teaming and testing. Prompt injection testing is part of the deployment checklist. Agents are tested in production-like conditions before going live. Continuous red-teaming is scheduled, not one-time.
Incident response for agents. There is a defined response protocol for when an agent is compromised. Agent session termination procedures exist and are tested. Data access revocation procedures exist and are tested.
The organizations that deploy agents without these controls are accumulating liability. Data breaches caused by agent compromise are not covered by traditional cyber insurance if the agent was not secured to policy standards.
The question for every organization deploying AI agents in 2026 is not whether the agent is ready for production. The question is whether the security architecture is ready for the agent. If the answer includes anything other than a confirmed yes for every control above, the agent is not ready for production.