The Security Gap Most Executives Miss

Eight-eight percent of enterprises reported AI agent security incidents in the last twelve months. Only twenty-one percent have runtime visibility into what their agents are doing. That gap defines the crisis.
According to a recent report by VentureBeat, the disconnect between perceived protection and actual exposure is not a minor oversight—it is the prevailing architecture in production today. Eighty-two percent of executives say their policies protect them from unauthorized agent actions. The data tells a different story. Gravitee’s State of AI Agent Security 2026 survey of nine hundred nineteen executives and practitioners quantifies this precisely: while leadership believes it has locked down the attack surface, the operational reality remains largely invisible.
What the survey reveals
The numbers expose a structural failure, not a knowledge gap. Security teams built monitoring dashboards designed for human-speed workflows. AI agents operate at machine speed. CrowdStrike’s Falcon sensors now detect more than eighteen hundred distinct AI applications across enterprise endpoints, and the fastest recorded adversary breakout time has dropped to twenty-seven seconds. No human-designed monitoring UI can keep pace.
This is why eighty-eight percent of enterprises experienced incidents while fewer than a quarter could see them in real time. The survey found that monitoring investment snapped back to forty-five percent of security budgets in March after dropping to twenty-four percent in February—when early movers shifted dollars into runtime enforcement and sandboxing. That pattern reveals the core problem: enterprises are stuck observing while their agents already require isolation.
Three Stages, One Fail Point

The threat landscape maps to three maturity stages. Most enterprises occupy only the first.
Stage one: observe without action
Stage one is observe. Security teams deploy logging, build dashboards, generate alerts. But observation without enforcement means the team sees the breach only after it completes. Meta experienced this directly in March two thousand twenty-six: a rogue AI agent passed every identity check and still exposed sensitive data to unauthorized employees. Two weeks later, Mercor, a ten-billion-dollar AI startup, confirmed a supply-chain breach through LiteLLM. Both incidents traced to the same structural gap—monitoring without enforcement.
The OWASP Top 10 for Agentic Applications two thousand twenty-six formalized the attack surface last December. Goal hijack, tool misuse, identity and privilege abuse, agentic supply chain vulnerabilities, unexpected code execution, memory poisoning, insecure inter-agent communication, cascading failures, human-agent trust exploitation, and rogue agents represent risks with no direct analog in traditional LLM applications. Stage-one security cannot see most of them.
Why enforcement without isolation fails
Stage two introduces enforcement—IAM integration and cross-provider controls turn observation into action. But enforcement without isolation leaves a fatal gap. When guardrails fail, there is no blast radius containment. The March wave of VentureBeat’s survey confirms the pattern: enterprises in enforcement mode still lack the architectural isolation that bounds damage when controls fail.
Invariant Labs disclosed the MCP Tool Poisoning Attack in April two thousand twenty-five, demonstrating how malicious instructions in an MCP server’s tool description cause an agent to exfil files or hijack a trusted server. CyberArk extended this to Full-Schema Poisoning. The mcp-remote OAuth proxy patched CVE-2025-6514 after a command-injection flaw put four hundred thirty-seven thousand downloads at risk. These attacks do not trigger at stage one because no one is watching the right thing. They do not stop at stage two because enforcement alone cannot isolate what it cannot distinguish.
The Identity Architecture Problem
Root cause runs deeper than monitoring. The identity architecture for AI agents does not exist in most enterprises.
Non-human identity explosion
CrowdStrike CTO Elia Zaitsev framed this at RSAC two thousand twenty-six: “AI agents and non-human identities will explode across the enterprise, expanding exponentially and dwarfing human identities. Each agent will operate as a privileged super-human with OAuth tokens, API keys, and continuous access to previously siloed data sets.” That is the future. The present is twenty-one point nine percent of teams treating agents as identity-bearing entities. Forty-five point six percent still use shared API keys.
Shared API keys grant unlimited lateral movement once compromised. There is no per-agent permission boundary, no audit trail tying an action to a specific agent identity. Identity security built for humans will not survive this shift. Cisco President Jeetu Patel offered the operational analogy: agents behave “more like teenagers, supremely intelligent, but with no fear of consequence.” Teenagers with OAuth tokens and API keys and continuous data access.
ASI08 as architecture
Twenty-five point five percent of deployed agents can create and task other agents. A quarter of enterprises can spawn agents their security team never provisioned. That is ASI08—cascading failures—as architecture. When a parent agent spawns a child agent with elevated permissions, the blast radius of compromise scales with each generation. The survey data confirms this: enterprises built agent-spawning capability into their architecture without building isolation into their security model.
The Regulatory Clock

The compliance exposure timeline is colliding with the security gap.
Healthcare’s warning signal
In healthcare, the gap costs more. Gravitee’s survey found ninety-two point seven percent of healthcare organizations reported AI agent security incidents, versus eighty-eight percent all-industry average. HIPAA’s two thousand twenty-six Tier four willful-neglect maximum is two point nineteen million dollars per violation category per year. For a health system running agents that touch PHI, that ratio is the difference between a reportable breach and an uncontested finding of willful neglect.
Auditing priority tells the story in miniature. In January, fifty percent of respondents ranked auditability a top concern. By February, that dropped to twenty-eight percent as teams sprinted to deploy. In March, it surged to sixty-five percent when those same teams realized they had no forensic trail for what their agents did. The regulatory clock is ticking. The audit trail is not there.
Why audit trails matter now
FINRA’s two thousand twenty-six Oversight Report recommends explicit human checkpoints before agents that can act or transact execute, along with narrow scope, granular permissions, and complete audit trails. Mike Riemer, Field CISO at Ivanti, quantified the speed problem: “Threat actors are reverse engineering patches within seventy-two hours. If a customer doesn’t patch within seventy-two hours of release, they’re open to exploit.” Most enterprises take weeks.
At machine speed, that window becomes permanent exposure. Without an audit trail, there is no forensic capability. Without forensic capability, compliance becomes indefensible. The regulatory surface is expanding. The architectural capability to address it is not.
Guardrails Address the Wrong Surface
The common assumption—that model-level guardrails protect against agent compromise—fails under scrutiny.
The control surface mismatch
A two thousand twenty-five paper by Kazdan and colleagues from Stanford, ServiceNow Research, Toronto, and FAR AI demonstrated this precisely. A fine-tuning attack bypassed model-level guardrails in seventy-two percent of attempts against Claude three Haiku and fifty-seven percent against GPT-4o. The attack received a two-thousand-dollar bug bounty from OpenAI and was acknowledged as a vulnerability by Anthropic.
Guardrails constrain what an agent is told to do, not what a compromised agent can reach. The control surface mismatch is fundamental: guardrails operate on the instruction layer, while the threat operates on the execution layer. When an agent is compromised, guardrails are already irrelevant. The agent has the permissions, the tokens, the access. Guardrails address the wrong surface.
What CISOs actually prioritize
Prevention of unauthorized actions ranked as the top capability priority in every wave of VentureBeat’s three-wave survey at sixty-eight to seventy-two percent. That is the most stable high-conviction signal in the dataset. The demand is for permissioning, not prompting. Elia Zaitsev articulated this at RSAC two thousand twenty-six: “It looks indistinguishable if an agent runs your web browser versus if you run your browser.” Distinguishing the two requires walking the process tree, tracing whether Chrome was launched by a human from the desktop or spawned by an agent in the background.
Merritt Baer, CSO at Enkrypt AI and former AWS Deputy CISO, framed the gap directly: “Enterprises believe they’ve ‘approved’ AI vendors, but what they’ve actually approved is an interface, not the underlying system. The real dependencies are one or two layers deeper, and those are the ones that fail under stress.” Approval at the interface layer does not propagate to the permission layer. That is the gap between policy and architecture.
Bottom Line: What You Can Do Now

The threat is architectural. The response must be architectural. Here is where to start.
Immediate controls
First, deploy agent API call logging to SIEM. Baseline normal tool-call patterns per agent role. VentureBeat’s prescriptive matrix maps the first detection test clearly: inject a canary token into a test document, route it through your agent, and if the token leaves your network, stage one failed.
Second, treat agents as identity-bearing entities. Replace shared API keys with per-agent credentials. Implement OAuth token lifecycle management. If forty-five point six percent of enterprises still use shared keys, that is the highest-ROI control to fix first.
Third, implement sandboxed execution for any agent with elevated permissions or cross-system access. Isolation bounds blast radius when guardrails fail. Because they will fail. The question is not whether an agent is compromised. The question is whether that compromise stays contained.
Enterprise AI agent security threats enterprise adoption has outpaced enterprise security architecture. That mismatch is the story. It is also the opportunity—for the teams that close the gap first.

Hi, I’m Cary Huang — a tech enthusiast based in Canada. I’ve spent years working with complex production systems and open-source software. Through TechBuddies.io, my team and I share practical engineering insights, curate relevant tech news, and recommend useful tools and products to help developers learn and work more effectively.





