The Patch That Proved Nothing Fixed

Microsoft assigned CVE-2026-21520 to a prompt injection vulnerability in Copilot Studio. The patch arrived on January 15, 2026. The data still left the building. That sequence tells you everything about the state of prompt injection vulnerability agent security in enterprise AI: the fix came, the exploitation continued, and the industry is still catching up to what that means.
Researchers at Capsule Security discovered a flaw they named ShareLeak that targets the gap between a SharePoint form submission and the Copilot Studio agent’s context window. An attacker fills a public-facing comment field with a crafted payload that injects a fake system role message. In Capsule’s testing, Copilot Studio concatenated the malicious input directly with the agent’s system instructions with no input sanitization between the form and the model.
The injected payload overrode the agent’s original instructions in the proof-of-concept, directing it to query connected SharePoint Lists for customer data and send that data via Outlook to an attacker-controlled email address. Microsoft’s own safety mechanisms flagged the request as suspicious during testing. The data exfiltrated anyway.
The DLP never fired because the email was routed through a legitimate Outlook action that the system treated as an authorized operation. Carter Rees, VP of Artificial Intelligence at Reputation, described the architectural failure in an exclusive interview: the LLM cannot inherently distinguish between trusted instructions and untrusted retrieved data. It becomes a confused deputy acting on behalf of the attacker.
Why This CVE Matters More Than the Score
The CVE carries a CVSS score of 7.5, which places it in the “medium” range. That score obscures the real significance. Capsule’s research calls Microsoft’s decision to assign a CVE to a prompt injection vulnerability in an agentic platform “highly unusual.”
Microsoft previously assigned CVE-2025-32711 (CVSS 9.3) to EchoLeak, a prompt injection in M365 Copilot patched in June 2025, but that targeted a productivity assistant, not an agent-building platform. The Copilot Studio CVE represents something fundamentally different: it signals that the vulnerability class now extends to systems designed to act autonomously on behalf of users.
If the precedent extends to agentic systems broadly, every enterprise running agents inherits a new vulnerability class to track. OWASP classifies this pattern as ASI01: Agent Goal Hijack. The structural reality is that this class cannot be fully eliminated by patches alone.
The Lethal Trifecta That Makes Every Agent Exploitable

Naor Paz, CEO of Capsule Security, named the structural condition that makes any agent exploitable: access to private data, exposure to untrusted content, and the ability to communicate externally. ShareLeak hits all three. PipeLeak hits all three. Most production agents hit all three because that combination is what makes agents useful.
Why Defense-in-Depth Rules Fail Agents
Rees validated the diagnosis independently. Defense-in-depth predicated on deterministic rules is fundamentally insufficient for agentic systems. The core problem is that traditional security relies on known-bad patterns: if you can identify the threat, you can block it. But prompt injection attacks don’t exploit a vulnerability in the traditional sense—they exploit the fundamental architecture of how language models process instruction.
The LLM cannot distinguish between a system instruction and a cleverly disguised user prompt that the model treats as authoritative. The model is doing exactly what it was designed to do: follow instructions. The security failure is architectural, not implementation-level. You cannot patch your way out of a problem where the attack vector is indistinguishable from legitimate functionality.
Elia Zaitsev, CrowdStrike’s CTO, called the patching mindset itself the vulnerability. “People are forgetting about runtime security,” he said. “Let’s patch all the vulnerabilities. Impossible. Somehow always seem to miss something.”
TheThe Salesforce Split: Same Flaw, Different Response
Capsule also discovered PipeLeak, a parallel indirect prompt injection vulnerability in Salesforce Agentforce. In testing, a public lead form payload hijacked an Agentforce agent with no authentication required. Capsule found no volume cap on the exfiltrated CRM data, and the employee who triggered the agent received no indication that data had left the building.
Microsoft patched ShareLeak and assigned a CVE. Salesforce patched ForcedLeak (CVSS 9.4), disclosed by Noma Labs in September 2025, by enforcing Trusted URL allowlists. But according to Capsule’s research, PipeLeak survives that patch through a different channel: email via the agent’s authorized tool actions.
Salesforce has not assigned a CVE or issued a public advisory for PipeLeak as of publication. The company recommended human-in-the-loop as a mitigation. Paz pushed back: “If the human should approve every single operation, it’s not really an agent. It’s just a human clicking through the agent’s actions.”
The contrast in response reveals something important about the industry’s current posture: some vendors treat prompt injection in agentic systems as a security vulnerability requiring coordinated disclosure, while others treat it as a configuration or governance issue. That inconsistency creates fragmentation in how enterprises approach the problem.
Multi-Turn Attacks and the Monitoring Blind Spot

Single-shot prompt injections are the entry-level threat. Capsule’s research documented multi-turn crescendo attacks where adversaries distribute payloads across multiple benign-looking turns. Each turn passes inspection. The attack becomes visible only when analyzed as a sequence.
Why Stateless Detection Misses Semantic Trajectories
Rees explained why current monitoring misses this: a stateless WAF views each turn in a vacuum and detects no threat. It sees requests, not a semantic trajectory. Each individual prompt appears harmless. The malicious intent only emerges when the model processes them as a连贯 sequence—something no perimeter security tool is designed to analyze.
Kayne McGladrey, IEEE Senior Member, put it differently: organizations are cloning human user accounts to agentic systems, except agents use far more permissions than humans would because of the speed, the scale, and the intent. The lethal trifecta.
The Coding Agent Memory Poisoning Risk
Capsule also found undisclosed vulnerabilities in coding agent platforms it declined to name, including memory poisoning that persists across sessions and malicious code execution through MCP servers. In one case, a file-level guardrail designed to restrict which files the agent could access was reasoned around by the agent itself, which found an alternate path to the same data.
Rees identified the human vector: employees paste proprietary code into public LLMs and view security as friction. The combination of persistent memory in coding agents and user behavior that treats security as an obstacle creates a compounding risk that traditional security tools cannot address.
Runtime Security: The Architectural Shift Required
The gap between prompt and action is where legacy security fails. Chris Krebs, the first Director of CISA and a Capsule advisor, put it in operational terms: “Legacy tools weren’t built to monitor what happens between prompt and action. That’s the runtime gap.”
Beyond Patching: Guardian Agent Approaches
Capsule hooks into vendor-provided agentic execution paths—including Copilot Studio’s security hooks and Claude Code’s pre-tool-use checkpoints—with no proxies, gateways, or SDKs. The company deployed fine-tuned small language models that evaluate every tool call before execution, an approach Gartner’s market guide calls a “guardian agent.”
This model treats the runtime layer as the enforcement point rather than the input layer. Instead of trying to filter malicious prompts at entry, the architecture evaluates whether the intended action aligns with the agent’s authorized scope at the moment of execution.
Not everyone agrees that intent analysis is the right layer. Zaitsev told VentureBeat that intent-based detection is non-deterministic. Observing actual kinetic actions is a structured, solvable problem. Intent is not. CrowdStrike’s Falcon sensor walks the process tree and tracks what agents did, not what they appeared to intend.
The debate reveals the maturity gap: we are still determining what model of runtime enforcement actually scales. What is clear is that patching alone will not close the gap.
What Developers Need to Audit Now
Every security director running Copilot Studio agents triggered by SharePoint forms should audit that window for indicators of compromise. The vulnerability was patched in January, but the exposure window existed from discovery on November 24, 2025 through deployment on January 15, 2026.
For developers building with agents, the immediate actions are straightforward: audit where external content flows into agent context windows, review data egress paths for authorized tool actions that bypass DLP, implement logging that captures semantic trajectories across conversation turns, and treat runtime enforcement as a primary control layer, not a supplementary one.
The prompt injection vulnerability agent security challenge is not a bug to fix. It is an architectural condition to manage. The industry is moving from “can we prevent this?” to “how do we detect when it happens and limit the blast radius?” That shift is where the next generation of agent security tooling will be built.

Hi, I’m Cary Huang — a tech enthusiast based in Canada. I’ve spent years working with complex production systems and open-source software. Through TechBuddies.io, my team and I share practical engineering insights, curate relevant tech news, and recommend useful tools and products to help developers learn and work more effectively.





