<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0">
    <channel>
        <title>Security | VentureBeat</title>
        <link>https://venturebeat.com/category/security/feed/</link>
        <description>Transformative tech coverage that matters</description>
        <lastBuildDate>Mon, 20 Apr 2026 15:16:11 GMT</lastBuildDate>
        <docs>https://validator.w3.org/feed/docs/rss2.html</docs>
        <generator>https://github.com/jpmonette/feed</generator>
        <language>en</language>
        <copyright>Copyright 2026, VentureBeat</copyright>
        <item>
            <title><![CDATA[Most enterprises can't stop stage-three AI agent threats, VentureBeat survey finds]]></title>
            <link>https://venturebeat.com/security/most-enterprises-cant-stop-stage-three-ai-agent-threats-venturebeat-survey-finds</link>
            <guid isPermaLink="false">3AqqUC3RVDZbckt57R8hoO</guid>
            <pubDate>Fri, 17 Apr 2026 17:07:29 GMT</pubDate>
            <description><![CDATA[<p>A rogue AI agent at Meta <a href="https://venturebeat.com/security/meta-rogue-ai-agent-confused-deputy-iam-identity-governance-matrix">passed every identity check and still exposed sensitive data</a> to unauthorized employees in March. Two weeks later, <a href="https://fortune.com/2026/04/02/mercor-ai-startup-security-incident-10-billion/">Mercor</a>, a $10 billion AI startup, confirmed a supply-chain breach through LiteLLM. Both are traced to the same structural gap. Monitoring without enforcement, enforcement without isolation. A VentureBeat three-wave survey of 108 qualified enterprises found that the gap is not an edge case. It is the most common security architecture in production today.</p><p>Gravitee’s <a href="https://www.gravitee.io/state-of-ai-agent-security">State of AI Agent Security 2026</a> survey of 919 executives and practitioners quantifies the disconnect. 82% of executives say their policies protect them from unauthorized agent actions. Eighty-eight percent reported AI agent security incidents in the last twelve months. Only 21% have runtime visibility into what their agents are doing. Arkose Labs’ <a href="https://securityboulevard.com/2026/04/97-of-enterprises-expect-a-major-ai-agent-security-incident-within-the-year/">2026 Agentic AI Security Report</a> found 97% of enterprise security leaders expect a material AI-agent-driven incident within 12 months. Only 6% of security budgets address the risk.</p><p>VentureBeat&#x27;s survey results show that monitoring investment snapped back to 45% of security budgets in March after dropping to 24% in February, when early movers shifted dollars into runtime enforcement and sandboxing. The March wave (n=20) is directional, but the pattern is consistent with February’s larger sample (n=50): enterprises are stuck at observation while their agents already need isolation. CrowdStrike’s Falcon sensors detect more than <a href="https://ir.crowdstrike.com/news-releases/news-release-details/crowdstrike-establishes-endpoint-epicenter-ai-security">1,800 distinct AI applications</a> across enterprise endpoints. The fastest recorded adversary breakout time has dropped to <a href="https://venturebeat.com/security/rsac-2026-agentic-soc-agent-telemetry-security-gap">27 seconds</a>. Monitoring dashboards built for human-speed workflows cannot keep pace with machine-speed threats.</p><p>The audit that follows maps three stages. Stage one is observe. Stage two is enforce, where IAM integration and cross-provider controls turn observation into action. Stage three is isolate, sandboxed execution that bounds blast radius when guardrails fail. VentureBeat Pulse data from 108 qualified enterprises ties each stage to an investment signal, an OWASP ASI threat vector, a regulatory surface, and immediate steps security leaders can take.</p><h2>The threat surface stage-one security cannot see</h2><p>The <a href="https://genai.owasp.org/resource/owasp-top-10-for-agentic-applications-for-2026/">OWASP Top 10 for Agentic Applications 2026</a> formalized the attack surface last December. The ten risks are: goal hijack (ASI01), tool misuse (ASI02), identity and privilege abuse (ASI03), agentic supply chain vulnerabilities (ASI04), unexpected code execution (ASI05), memory poisoning (ASI06), insecure inter-agent communication (ASI07), cascading failures (ASI08), human-agent trust exploitation (ASI09), and rogue agents (ASI10). Most have no analog in traditional LLM applications. The audit below maps six of these to the stages where they are most likely to surface and the controls that address them.</p><p>Invariant Labs disclosed the <a href="https://invariantlabs.ai/blog/mcp-security-notification-tool-poisoning-attacks">MCP Tool Poisoning Attack</a> in April 2025: malicious instructions in an MCP server’s tool description cause an agent to exfiltrate files or hijack a trusted server. CyberArk extended it to <a href="https://www.cyberark.com/resources/threat-research-blog/poison-everywhere-no-output-from-your-mcp-server-is-safe">Full-Schema Poisoning</a>. The mcp-remote OAuth proxy patched CVE-2025-6514 after a command-injection flaw put 437,000 downloads at risk.</p><p>Merritt Baer, CSO at Enkrypt AI and former AWS Deputy CISO, framed the gap in an exclusive VentureBeat interview: “Enterprises believe they’ve ‘approved’ AI vendors, but what they’ve actually approved is an interface, not the underlying system. The real dependencies are one or two layers deeper, and those are the ones that fail under stress.”</p><p>CrowdStrike CTO Elia Zaitsev put the visibility problem in operational terms in an <a href="https://venturebeat.com/security/rsac-2026-agentic-soc-agent-telemetry-security-gap">exclusive VentureBeat interview at RSAC 2026</a>: “It looks indistinguishable if an agent runs your web browser versus if you run your browser.” Distinguishing the two requires walking the process tree, tracing whether Chrome was launched by a human from the desktop or spawned by an agent in the background. Most enterprise logging configurations cannot make that distinction.</p><h2>The regulatory clock and the identity architecture</h2><p>Auditability priority tells the same story in miniature. In January, 50% of respondents ranked it a top concern. By February, that dropped to 28% as teams sprinted to deploy. In March, it surged to 65% when those same teams realized they had no forensic trail for what their agents did.</p><p>HIPAA’s 2026 Tier 4 willful-neglect maximum is <a href="https://www.hipaacoach.com/what-is-the-maximum-penalty-for-a-hipaa-violation">$2.19M per violation category per year</a>. In healthcare, Gravitee’s survey found 92.7% of organizations reported AI agent security incidents versus the 88% all-industry average. For a health system running agents that touch PHI, that ratio is the difference between a reportable breach and an uncontested finding of willful neglect. <a href="https://www.finra.org/rules-guidance/guidance/reports/2026-finra-annual-regulatory-oversight-report/gen-ai">FINRA’s 2026 Oversight Report</a> recommends explicit human checkpoints before agents that can act or transact execute, along with narrow scope, granular permissions, and complete audit trails of agent actions.</p><p>Mike Riemer, Field CISO at Ivanti, quantified the speed problem in a recent VentureBeat interview: “Threat actors are reverse engineering patches within 72 hours. If a customer doesn’t patch within 72 hours of release, they’re open to exploit.” Most enterprises take weeks. Agents operating at machine speed widen that window into a permanent exposure.</p><p>The identity problem is architectural. <a href="https://www.gravitee.io/blog/state-of-ai-agent-security-2026-report-when-adoption-outpaces-control">Gravitee&#x27;s survey of 919 practitioners</a> found only 21.9% of teams treat agents as identity-bearing entities, 45.6% still use shared API keys, and 25.5% of deployed agents can create and task other agents. A quarter of enterprises can spawn agents that their security team never provisioned. That is ASI08 as architecture.</p><h2>Guardrails alone are not a strategy</h2><p>A 2025 paper by <a href="https://arxiv.org/abs/2502.19537">Kazdan and colleagues</a> (Stanford, ServiceNow Research, Toronto, FAR AI) showed a fine-tuning attack that bypasses model-level guardrails in 72% of attempts against Claude 3 Haiku and 57% against GPT-4o. The attack received a $2,000 bug bounty from OpenAI and was acknowledged as a vulnerability by Anthropic. Guardrails constrain what an agent is told to do, not what a compromised agent can reach.</p><p>CISOs already know this. In VentureBeat&#x27;s three-wave survey, prevention of unauthorized actions ranked as the top capability priority in every wave at 68% to 72%, the most stable high-conviction signal in the dataset. The demand is for permissioning, not prompting. Guardrails address the wrong control surface.</p><p>Zaitsev framed the identity shift at RSAC 2026: “AI agents and non-human identities will explode across the enterprise, expanding exponentially and dwarfing human identities. Each agent will operate as a privileged super-human with OAuth tokens, API keys, and continuous access to previously siloed data sets.” Identity security built for humans will not survive this shift. Cisco President Jeetu Patel offered the operational analogy in an exclusive VentureBeat interview: agents behave “more like teenagers, supremely intelligent, but with no fear of consequence.”</p><h2>VentureBeat Prescriptive Matrix: AI Agent Security Maturity Audit</h2><table><tbody><tr><td><p><b>Stage</b></p></td><td><p><b>Attack Scenario</b></p></td><td><p><b>What Breaks</b></p></td><td><p><b>Detection Test</b></p></td><td><p><b>Blast Radius</b></p></td><td><p><b>Recommended Control</b></p></td></tr><tr><td><p><b>1: Observe</b></p></td><td><p>Attacker embeds goal-hijack payload in forwarded email (ASI01). Agent summarizes email and silently exfiltrates credentials to an external endpoint. See: Meta March 2026 incident.</p></td><td><p>No runtime log captures the exfiltration. SIEM never sees the API call. The security team learns from the victim. Zaitsev: agent activity is “indistinguishable” from human activity in default logging.</p></td><td><p>Inject a canary token into a test document. Route it through your agent. If the token leaves your network, stage one failed.</p></td><td><p>Single agent, single session. With shared API keys (45.6% of enterprises): unlimited lateral movement.</p></td><td><p>Deploy agent API call logging to SIEM. Baseline normal tool-call patterns per agent role. Alert on the first outbound call to an unrecognized endpoint.</p></td></tr><tr><td><p><b>2: Enforce</b></p></td><td><p>Compromised MCP server poisons tool description (ASI04). Agent invokes poisoned tool, writes attacker payload to production DB using inherited service-account credentials. See: Mercor/LiteLLM April 2026 supply-chain breach.</p></td><td><p>IAM allows write because agent uses shared service account. No approval gate on write ops. Poisoned tool indistinguishable from clean tool in logs. Riemer: “72-hour patch window” collapses to zero when agents auto-invoke.</p></td><td><p>Register a test MCP server with a benign-looking poisoned description. Confirm your policy engine blocks the tool call before execution reaches the database. Run mcp-scan on all registered servers.</p></td><td><p>Production database integrity. If agent holds DBA-level credentials: full schema compromise. Lateral movement via trust relationships to downstream agents.</p></td><td><p>Assign scoped identity per agent. Require approval workflow for all write ops. Revoke every shared API key. Run mcp-scan on all MCP servers weekly.</p></td></tr><tr><td><p><b>3: Isolate</b></p></td><td><p>Agent A spawns Agent B to handle subtask (ASI08). Agent B inherits Agent A’s permissions, escalates to admin, rewrites org security policy. Every identity check passes. Source: CrowdStrike CEO George Kurtz, RSAC 2026 keynote.</p></td><td><p>No sandbox boundary between agents. No human gate on agent-to-agent delegation. Security policy modification is a valid action for admin-credentialed process. CrowdStrike CEO George Kurtz disclosed at RSAC 2026 that the agent “wanted to fix a problem, lacked permissions, and removed the restriction itself.”</p></td><td><p>Spawn a child agent from a sandboxed parent. Child should inherit zero permissions by default and require explicit human approval for each capability grant.</p></td><td><p>Organizational security posture. A rogue policy rewrite disables controls for every subsequent agent. 97% of enterprise leaders expect a material incident within 12 months (Arkose Labs 2026).</p></td><td><p>Sandbox all agent execution. Zero-trust for agent-to-agent delegation: spawned agents inherit nothing. Human sign-off before any agent modifies security controls. Kill switch per OWASP ASI10.</p></td></tr></tbody></table><p><i>Sources: OWASP Top 10 for Agentic Applications 2026; Invariant Labs MCP Tool Poisoning (April 2025); CrowdStrike RSAC 2026 Fortune 50 disclosure; Meta March 2026 incident (The Information/Engadget); Mercor/LiteLLM breach (Fortune, April 2, 2026); Arkose Labs 2026 Agentic AI Security Report; VentureBeat Pulse Q1 2026.</i></p><p>The stage-one attack scenario in this matrix is not hypothetical. Unauthorized tool or data access ranked as the most feared failure mode in every wave of VentureBeat’s survey, growing from 42% in January to 50% in March. That trajectory and the 70%-plus priority rating for prevention of unauthorized actions are the two most mutually reinforcing signals in the entire dataset. CISOs fear the exact attack this matrix describes, and most have not deployed the controls to stop it.</p><h2>Hyperscaler stage readiness: observe, enforce, isolate</h2><p>The maturity audit tells you where your security program stands. The next question is whether your cloud platform can get you to stage two and stage three, or whether you are building those capabilities yourself. Patel put it bluntly: “It’s not just about authenticating once and then letting the agent run wild.” A stage-three platform running a stage-one deployment pattern gives you stage-one risk.</p><p>VentureBeat Pulse data surfaces a structural tension in this grid. OpenAI leads enterprise AI security deployments at 21% to 26% across the three survey waves, making the same provider that creates the AI risk also the primary security layer. The provider-as-security-vendor pattern holds across Azure, Google, and AWS. Zero-incremental-procurement convenience is winning by default. Whether that concentration is a feature or a single point of failure depends on how far the enterprise has progressed past stage one.</p><table><tbody><tr><td><p><b>Provider</b></p></td><td><p><b>Identity Primitive (Stage 2)</b></p></td><td><p><b>Enforcement Control (Stage 2)</b></p></td><td><p><b>Isolation Primitive (Stage 3)</b></p></td><td><p><b>Gap as of April 2026 </b></p></td></tr><tr><td><p><b>Microsoft Azure</b></p></td><td><p>Entra ID agent scoping. Agent 365 maps agents to owners. GA.</p></td><td><p>Copilot Studio DLP policies. Purview for agent output classification. GA.</p></td><td><p>Azure Confidential Containers for agent workloads. Preview. No per-agent sandbox at GA.</p></td><td><p>No agent-to-agent identity verification. No MCP governance layer. Agent 365 monitors but cannot block in-flight tool calls.</p></td></tr><tr><td><p><b>Anthropic</b></p></td><td><p>Managed Agents: per-agent scoped permissions, credential mgmt. Beta (April 8, 2026). $0.08/session-hour.</p></td><td><p>Tool-use permissions, system prompt enforcement, and built-in guardrails. GA.</p></td><td><p>Managed Agents sandbox: isolated containers per session, execution-chain auditability. Beta. Allianz, Asana, Rakuten, and Sentry are in production.</p></td><td><p>Beta pricing/SLA not public. Session data in Anthropic-managed DB (lock-in risk per VentureBeat research). GA timing TBD.</p></td></tr><tr><td><p><b>Google Cloud</b></p></td><td><p>Vertex AI service accounts for model endpoints. IAM Conditions for agent traffic. GA.</p></td><td><p>VPC Service Controls for agent network boundaries. Model Armor for prompt/response filtering. GA.</p></td><td><p>Confidential VMs for agent workloads. GA. Agent-specific sandbox in preview.</p></td><td><p>Agent identity ships as a service account, not an agent-native principal. No agent-to-agent delegation audit. Model Armor does not inspect tool-call payloads.</p></td></tr><tr><td><p><b>OpenAI</b></p></td><td><p>Assistants API: function-call permissions, structured outputs. Agents SDK. GA.</p></td><td><p>Agents SDK guardrails, input/output validation. GA.</p></td><td><p>Agents SDK Python sandbox. Beta (API and defaults subject to change before GA per OpenAI docs). TypeScript sandbox confirmed, not shipped.</p></td><td><p>No cross-provider identity federation. Agent memory forensics limited to session scope. No kill switch API. No MCP tool-description inspection.</p></td></tr><tr><td><p><b>AWS</b></p></td><td><p>Bedrock model invocation logging. IAM policies for model access. CloudTrail for agent API calls. GA.</p></td><td><p>Bedrock Guardrails for content filtering. Lambda resource policies for agent functions. GA.</p></td><td><p>Lambda isolation per agent function. GA. Bedrock agent-level sandboxing on roadmap, not shipped.</p></td><td><p>No unified agent control plane across Bedrock + SageMaker + Lambda. No agent identity standard. Guardrails do not inspect MCP tool descriptions.</p></td></tr></tbody></table><p><i>Status as of April 15, 2026. GA = generally available. Preview/Beta = not production-hardened. “What’s Missing” column reflects VentureBeat’s analysis of publicly documented capabilities; gaps may narrow as vendors ship updates.</i></p><p>No provider in this grid ships a complete stage-three stack today. Most enterprises assemble isolation from existing cloud building blocks. That is a defensible choice if it is a deliberate one. Waiting for a vendor to close the gap without acknowledging the gap is not a strategy.</p><p>The grid above covers hyperscaler-native SDKs. A large segment of AI builders deploys through open-source orchestration frameworks like LangChain, CrewAI, and LlamaIndex that bypass hyperscaler IAM entirely. These frameworks lack native stage-two primitives. There is no scoped agent identity, no tool-call approval workflow, and no built-in audit trails. Enterprises running agents through open-source orchestration need to layer enforcement and isolation on top, not assume the framework provides it.</p><p>VentureBeat’s survey quantifies the pressure. Policy enforcement consistency grew from 39.5% to 46% between January and February, the largest consistent gain of any capability criterion. Enterprises running agents across OpenAI, Anthropic, and Azure need enforcement that works the same way regardless of which model executes the task. Provider-native controls enforce policy within that provider’s runtime only. Open-source orchestration frameworks enforce it nowhere.</p><p>One counterargument deserves acknowledgment: not every agent deployment needs stage three. A read-only summarization agent with no tool access and no write permissions may rationally stop at stage one. The sequencing failure this audit addresses is not that monitoring exists. It is that enterprises running agents with write access, shared credentials, and agent-to-agent delegation are treating monitoring as sufficient. For those deployments, stage one is not a strategy. It is a gap.</p><h2>Allianz shows stage-three in production</h2><p>Allianz, one of the world’s largest insurance and asset management companies, is running Claude Managed Agents across insurance workflows, with Claude Code deployed to technical teams and a dedicated AI logging system for regulatory transparency, per <a href="https://siliconangle.com/2026/04/08/anthropic-launches-claude-managed-agents-speed-ai-agent-development/">Anthropic’s April 8 announcement</a>. Asana, Rakuten, Sentry, and Notion are in production on the same beta. Stage-three isolation, per-agent permissioning, and execution-chain auditability are deployable now, not roadmap. The gating question is whether the enterprise has sequenced the work to use them.</p><h2>The 90-day remediation sequence</h2><p><b>Days 1–30: Inventory and baseline.</b> Map every agent to a named owner. Log all tool calls. Revoke shared API keys. Deploy read-only monitoring across all agent API traffic. Run <a href="https://github.com/invariantlabs-ai/mcp-scan">mcp-scan</a> against every registered MCP server. CrowdStrike detects 1,800 AI applications across enterprise endpoints; your inventory should be equally comprehensive. Output: agent registry with permission matrix, MCP scan report.</p><p><b>Days 31–60: Enforce and scope.</b> Assign scoped identities to every agent. Deploy tool-call approval workflows for write operations. Integrate agent activity logs into existing SIEM. Run a tabletop exercise: What happens when an agent spawns an agent? Conduct a canary-token test from the prescriptive matrix. Output: IAM policy set, approval workflow, SIEM integration, canary-token test results.</p><p><b>Days 61–90: Isolate and test.</b> Sandbox high-risk agent workloads (PHI, PII, financial transactions). Enforce per-session least privilege. Require human sign-off for agent-to-agent delegation. Red-team the isolation boundary using the stage-three detection test from the matrix. Output: sandboxed execution environment, red-team report, board-ready risk summary with regulatory exposure mapped to HIPAA tier and FINRA guidance.</p><h2>What changes in the next 30 days</h2><p>EU AI Act <a href="https://artificialintelligenceact.eu/article/14/">Article 14</a> human-oversight obligations take effect August 2, 2026. Programs without named owners and execution trace capability face enforcement, not operational risk.</p><p>Anthropic’s <a href="https://platform.claude.com/docs/en/managed-agents/overview">Claude Managed Agents</a> is in public beta at $0.08 per session-hour. GA timing, production SLAs, and final pricing have not been announced.</p><p>OpenAI <a href="https://techcrunch.com/2026/04/15/openai-updates-its-agents-sdk-to-help-enterprises-build-safer-more-capable-agents/">Agents SDK</a> ships TypeScript support for sandbox and harness capabilities in a future release, per the company’s April 15 announcement. Stage-three sandbox becomes available to JavaScript agent stacks when it ships.</p><h2>What the sequence requires</h2><p>McKinsey’s <a href="https://www.mckinsey.com/capabilities/tech-and-ai/our-insights/tech-forward/state-of-ai-trust-in-2026-shifting-to-the-agentic-era">2026 AI Trust Maturity Survey</a> pegs the average enterprise at 2.3 out of 4.0 on its RAI maturity model, up from 2.0 in 2025 but still an enforcement-stage number; only one-third of the ~500 organizations surveyed report maturity levels of three or higher in governance. Seventy percent have not finished the transition to stage three. <a href="https://www.armosec.io/blog/ai-agent-sandboxing-progressive-enforcement-guide/">ARMO’s progressive enforcement methodology</a> gives you the path: behavioral profiles in observation, permission baselines in selective enforcement, and full least privilege once baselines stabilize. Monitoring investment was not wasted. It was stage one of three. The organizations stuck in the data treated it as the destination.</p><p>The budget data makes the constraint explicit. The share of enterprises reporting flat AI security budgets doubled from 7.9% in January to 16% in February in VentureBeat&#x27;s survey, with the March directional reading at 20%. Organizations expanding agent deployments without increasing security investment are accumulating security debt at machine speed. Meanwhile, the share reporting no agent security tooling at all fell from 13% in January to 5% in March. Progress, but one in twenty enterprises running agents in production still has zero dedicated security infrastructure around them.</p><h2>About this research</h2><p><i>Total qualified respondents: 108. VentureBeat Pulse AI Security and Trust is a three-wave VentureBeat survey run January 6 through March 15, 2026. Qualified sample (organizations 100+ employees): January n=38, February n=50, March n=20. Primary analysis runs from January to February; March is directional. Industry mix: Tech/Software 52.8%, Financial Services 10.2%, Healthcare 8.3%, Education 6.5%, Telecom/Media 4.6%, Manufacturing 4.6%, Retail 3.7%, other 9.3%. Seniority: VP/Director 34.3%, Manager 29.6%, IC 22.2%, C-Suite 9.3%.</i></p>]]></description>
            <author>louiswcolumbus@gmail.com (Louis Columbus)</author>
            <category>Security</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/2oq4gxUSORHuJY6GKHVxQ1/1ff08d293fe4d0c43df9f5c7a1893344/hero_survey.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[Microsoft patched a Copilot Studio prompt injection. The data exfiltrated anyway]]></title>
            <link>https://venturebeat.com/security/microsoft-salesforce-copilot-agentforce-prompt-injection-cve-agent-remediation-playbook</link>
            <guid isPermaLink="false">2SfAXQE3dbv0XV7jJdQsAa</guid>
            <pubDate>Wed, 15 Apr 2026 20:58:42 GMT</pubDate>
            <description><![CDATA[<p>Microsoft assigned <a href="https://nvd.nist.gov/vuln/detail/CVE-2026-21520">CVE-2026-21520</a>, a CVSS 7.5 indirect prompt injection vulnerability, to Copilot Studio. <a href="https://www.capsulesecurity.io/">Capsule Security</a> discovered the flaw, coordinated disclosure with Microsoft, and the patch was deployed on January 15. Public disclosure went live on Wednesday.</p><p>That CVE matters less for what it fixes and more for what it signals. Capsule’s research calls Microsoft’s decision to assign a CVE to a prompt injection vulnerability in an agentic platform “highly unusual.” Microsoft previously assigned <a href="https://thehackernews.com/2025/06/zero-click-ai-vulnerability-exposes.html">CVE-2025-32711</a> (CVSS 9.3) to EchoLeak, a prompt injection in M365 Copilot patched in June 2025, but that targeted a productivity assistant, not an agent-building platform. If the precedent extends to agentic systems broadly, every enterprise running agents inherits a new vulnerability class to track. Except that this class cannot be fully eliminated by patches alone.</p><p>Capsule also discovered what they call PipeLeak, a parallel indirect prompt injection vulnerability in Salesforce Agentforce. Microsoft patched and assigned a CVE. Salesforce has not assigned a CVE or issued a public advisory for PipeLeak as of publication, according to Capsule&#x27;s research. </p><h2>What ShareLeak actually does</h2><p>The vulnerability that the researchers named ShareLeak exploits the gap between a SharePoint form submission and the Copilot Studio agent’s context window. An attacker fills a public-facing comment field with a crafted payload that injects a fake system role message. In Capsule’s testing, Copilot Studio concatenated the malicious input directly with the agent’s system instructions with no input sanitization between the form and the model.</p><p>The injected payload overrode the agent’s original instructions in Capsule’s proof-of-concept, directing it to query connected SharePoint Lists for customer data and send that data via Outlook to an attacker-controlled email address. NVD classifies the attack as low complexity and requires no privileges.</p><p>Microsoft’s own safety mechanisms flagged the request as suspicious during Capsule’s testing. The data was exfiltrated anyway. The DLP never fired because the email was routed through a legitimate Outlook action that the system treated as an authorized operation.</p><p>Carter Rees, VP of Artificial Intelligence at <a href="https://reputation.com">Reputation</a>, described the architectural failure in an exclusive VentureBeat interview. The LLM cannot inherently distinguish between trusted instructions and untrusted retrieved data, Rees said. It becomes a <a href="https://venturebeat.com/security/meta-rogue-ai-agent-confused-deputy-iam-identity-governance-matrix">confused deputy</a> acting on behalf of the attacker. OWASP classifies this pattern as <a href="https://genai.owasp.org/2025/12/09/owasp-top-10-for-agentic-applications-the-benchmark-for-agentic-security-in-the-age-of-autonomous-ai/">ASI01: Agent Goal Hijack</a>.</p><p>The research team behind both discoveries, Capsule Security, found the Copilot Studio vulnerability on November 24, 2025. Microsoft confirmed it on December 5 and patched it on January 15, 2026. Every security director running Copilot Studio agents triggered by SharePoint forms should audit that window for indicators of compromise.</p><h2>PipeLeak and the Salesforce split</h2><p>PipeLeak hits the same vulnerability class through a different front door. In Capsule’s testing, a public lead form payload hijacked an Agentforce agent with no authentication required. Capsule found no volume cap on the exfiltrated CRM data, and the employee who triggered the agent received no indication that data had left the building. Salesforce has not assigned a CVE or issued a public advisory specific to PipeLeak as of publication.</p><p>Capsule is not the first research team to hit Agentforce with indirect prompt injection. Noma Labs disclosed <a href="https://noma.security/blog/forcedleak-agent-risks-exposed-in-salesforce-agentforce/">ForcedLeak</a> (CVSS 9.4) in September 2025, and Salesforce patched that vector by enforcing Trusted URL allowlists. According to Capsule&#x27;s research, PipeLeak survives that patch through a different channel: email via the agent&#x27;s authorized tool actions.</p><p>Naor Paz, CEO of Capsule Security, told VentureBeat the testing hit no exfiltration limit. “We did not get to any limitation,” Paz said. “The agent would just continue to leak all the CRM.”</p><p>Salesforce recommended human-in-the-loop as a mitigation. Paz pushed back. “If the human should approve every single operation, it’s not really an agent,” he told VentureBeat. “It’s just a human clicking through the agent’s actions.”</p><p>Microsoft patched ShareLeak and assigned a CVE. According to Capsule&#x27;s research, Salesforce patched ForcedLeak&#x27;s URL path but not the email channel.</p><p>Kayne McGladrey, IEEE Senior Member, put it differently in a separate VentureBeat interview. Organizations are cloning human user accounts to agentic systems, McGladrey said, except agents use far more permissions than humans would because of the speed, the scale, and the intent.</p><h2>The lethal trifecta and why posture management fails</h2><p>Paz named the structural condition that makes any agent exploitable: access to private data, exposure to untrusted content, and the ability to communicate externally. ShareLeak hits all three. PipeLeak hits all three. Most production agents hit all three because that combination is what makes agents useful.</p><p>Rees validated the diagnosis independently. Defense-in-depth predicated on deterministic rules is fundamentally insufficient for agentic systems, Rees <a href="https://venturebeat.com/security/openclaw-agentic-ai-security-risk-ciso-guide">told VentureBeat</a>.</p><p>Elia Zaitsev, CrowdStrike’s CTO, called the patching mindset itself the vulnerability in a <a href="https://venturebeat.com/security/rsac-2026-agent-identity-frameworks-three-gaps">separate VentureBeat exclusive</a>. “People are forgetting about runtime security,” he said. “Let’s patch all the vulnerabilities. Impossible. Somehow always seem to miss something.” Observing actual kinetic actions is a structured, solvable problem, Zaitsev told VentureBeat. Intent is not. CrowdStrike’s Falcon sensor <a href="https://venturebeat.com/security/rsac-2026-agentic-soc-agent-telemetry-security-gap">walks the process tree</a> and tracks what agents did, not what they appeared to intend.</p><h2>Multi-turn crescendo and the coding agent blind spot</h2><p>Single-shot prompt injections are the entry-level threat. Capsule’s research documented multi-turn crescendo attacks where adversaries distribute payloads across multiple benign-looking turns. Each turn passes inspection. The attack becomes visible only when analyzed as a sequence.</p><p>Rees explained why current monitoring misses this. A stateless WAF views each turn in a vacuum and detects no threat, Rees told VentureBeat. It sees requests, not a semantic trajectory.</p><p>Capsule also found undisclosed vulnerabilities in coding agent platforms it declined to name, including memory poisoning that persists across sessions and malicious code execution through MCP servers. In one case, a file-level guardrail designed to restrict which files the agent could access was reasoned around by the agent itself, which found an alternate path to the same data. Rees identified the human vector: employees paste proprietary code into public LLMs and view security as friction.</p><p>McGladrey cut to the governance failure. “If crime was a technology problem, we would have solved crime a fairly long time ago,” he told VentureBeat. “Cybersecurity risk as a standalone category is a complete fiction.”</p><h2>The runtime enforcement model</h2><p>Capsule hooks into vendor-provided agentic execution paths — including Copilot Studio&#x27;s security hooks and Claude Code&#x27;s pre-tool-use checkpoints — with no proxies, gateways, or SDKs. The company exited stealth on Wednesday, timing its $7 million seed round, led by Lama Partners alongside Forgepoint Capital International, to its coordinated disclosure.</p><p>Chris Krebs, the first Director of CISA and a Capsule advisor, put the gap in operational terms. “Legacy tools weren’t built to monitor what happens between prompt and action,” Krebs said. “That’s the runtime gap.”</p><p>Capsule&#x27;s architecture deploys fine-tuned small language models that evaluate every tool call before execution, an approach Gartner&#x27;s market guide calls a &quot;guardian agent.&quot;</p><p>Not everyone agrees that intent analysis is the right layer. Zaitsev told VentureBeat during an exclusive interview that intent-based detection is non-deterministic. “Intent analysis will sometimes work. Intent analysis cannot always work,” he said. CrowdStrike bets on observing what the agent actually did rather than what it appeared to intend. <a href="https://www.microsoft.com/en-us/security/blog/2026/03/30/addressing-the-owasp-top-10-risks-in-agentic-ai-with-microsoft-copilot-studio/">Microsoft’s own Copilot Studio documentation</a> provides external security-provider webhooks that can approve or block tool execution, offering a vendor-native control plane alongside third-party options. No single layer closes the gap. Runtime intent analysis, kinetic action monitoring, and foundational controls (least privilege, input sanitization, outbound restrictions, targeted human-in-the-loop) all belong in the stack. SOC teams should map telemetry now: Copilot Studio activity logs plus webhook decisions, CRM audit logs for Agentforce, and EDR process-tree data for coding agents.</p><p>Paz described the broader shift. “Intent is the new perimeter,” he told VentureBeat. “The agent in runtime can decide to go rogue on you.”</p><h2>VentureBeat Prescriptive Matrix</h2><p>The following matrix maps five vulnerability classes against the controls that miss them, and the specific actions security directors should take this week.</p><table><tbody><tr><td><p><b>Vulnerability Class</b></p></td><td><p><b>Why Current Controls Miss It</b></p></td><td><p><b>What Runtime Enforcement Does</b></p></td><td><p><b>Suggested actions for security leaders</b></p></td></tr><tr><td><p><b>ShareLeak </b>— Copilot Studio, <a href="https://nvd.nist.gov/vuln/detail/CVE-2026-21520">CVE-2026-21520</a>, CVSS 7.5, patched Jan 15 2026</p></td><td><p>Capsule’s testing found no input sanitization between the SharePoint form and the agent context. Safety mechanisms flagged, but data still exfiltrated. DLP did not fire because the email used a legitimate Outlook action. OWASP ASI01: Agent Goal Hijack.</p></td><td><p>Guardian agent hooks into Copilot Studio pre-tool-use security hooks. Vets every tool call before execution. Blocks exfiltration at the action layer.</p></td><td><p>Audit every Copilot Studio agent triggered by SharePoint forms. Restrict outbound email to org-only domains. Inventory all SharePoint Lists accessible to agents. Review the Nov 24–Jan 15 window for indicators of compromise.</p></td></tr><tr><td><p><b>PipeLeak </b>— Agentforce, no CVE assigned</p></td><td><p>In Capsule’s testing, public form input flowed directly into the agent context. No auth required. No volume cap observed on exfiltrated CRM data. The employee received no indication that data was leaving.</p></td><td><p>Runtime interception via platform agentic hooks. Pre-invocation checkpoint on every tool call. Detects outbound data transfer to non-approved destinations.</p></td><td><p>Review all Agentforce automations triggered by public-facing forms. Enable human-in-the-loop for external comms as interim control. Audit CRM data access scope per agent. Pressure Salesforce for CVE assignment.</p></td></tr><tr><td><p><b>Multi-Turn Crescendo </b>— distributed payload, each turn looks benign</p></td><td><p>Stateless monitoring inspects each turn in isolation. WAFs, DLP, and activity logs see individual requests, not semantic trajectory.</p></td><td><p>Stateful runtime analysis tracks full conversation history across turns. Fine-tuned SLMs evaluate aggregated context. Detects when a cumulative sequence constitutes a policy violation.</p></td><td><p>Require stateful monitoring for all production agents. Add crescendo attack scenarios to red team exercises.</p></td></tr><tr><td><p><b>Coding Agents </b>— unnamed platforms, memory poisoning + code execution</p></td><td><p>MCP servers inject code and instructions into the agent context. Memory poisoning persists across sessions. Guardrails reasoned around by the agent itself. Shadow AI insiders paste proprietary code into public LLMs.</p></td><td><p>Pre-invocation checkpoint on every tool call. Fine-tuned SLMs detect anomalous tool usage at runtime.</p></td><td><p>Inventory all coding agent deployments across engineering. Audit MCP server configs. Restrict code execution permissions. Monitor for shadow installations.</p></td></tr><tr><td><p><b>Structural Gap </b>— any agent with private data + untrusted input + external comms</p></td><td><p>Posture management tells you what should happen. It does not stop what does happen. Agents use far more permissions than humans at far greater speed.</p></td><td><p>Runtime guardian agent watches every action in real time. Intent-based enforcement replaces signature detection. Leverages vendor agentic hooks, not proxies or gateways.</p></td><td><p>Classify every agent by lethal trifecta exposure. Treat prompt injection as class-based SaaS risk. Require runtime security for any agent moving to production. Brief the board on agent risk as business risk.</p></td></tr></tbody></table><h2>What this means for 2026 security planning</h2><p>Microsoft’s CVE assignment will either accelerate or fragment how the industry handles agent vulnerabilities. If vendors call them configuration issues, CISOs carry the risk alone.</p><p>Treat prompt injection as a class-level SaaS risk rather than individual CVEs. Classify every agent deployment against the lethal trifecta. Require runtime enforcement for anything moving to production. Brief the board on agent risk the way McGladrey framed it: as business risk, because cybersecurity risk as a standalone category stopped being useful the moment agents started operating at machine speed.</p><p><i>Update, April 16, 2026: After publication, a Salesforce spokesperson stated the company has &quot;remediated the specific scenario described&quot; and that Human-in-the-Loop confirmation is enabled by default for email-based agentic actions. Capsule Security maintains that the email channel remains exploitable on Custom Topics (now called Sub-Agents in Agentforce), which represent the majority of enterprise deployments. Capsule retested after Salesforce&#x27;s response and reported unchanged behavior on Custom Topics.</i>
</p>]]></description>
            <author>louiswcolumbus@gmail.com (Louis Columbus)</author>
            <category>Security</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/6QO34Fn3Ix5qFbnemAM3a5/6cf10a1a9ecd680e39c790c0733d16fd/HERO_CAPSULE.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[Frontier models are failing one in three production attempts — and getting harder to audit]]></title>
            <link>https://venturebeat.com/security/frontier-models-are-failing-one-in-three-production-attempts-and-getting-harder-to-audit</link>
            <guid isPermaLink="false">1rR8Uvmvfd8KBuv8J2EIl9</guid>
            <pubDate>Wed, 15 Apr 2026 19:35:32 GMT</pubDate>
            <description><![CDATA[<p>AI agents are now embedded in real enterprise workflows, and they&#x27;re still failing roughly one in three attempts on structured benchmarks. That <a href="https://hai.stanford.edu/ai-index/2026-ai-index-report">gap between capability and reliability</a> is the defining operational challenge for IT leaders in 2026, according to Stanford HAI&#x27;s ninth annual AI Index report.</p><p>This uneven, unpredictable performance is what the AI Index calls the &quot;jagged frontier,&quot; a term <a href="https://www.oneusefulthing.org/p/centaurs-and-cyborgs-on-the-jagged">coined by AI researcher Ethan Mollick</a> to describe the boundary where AI excels and then suddenly fails.</p><p>“AI models can win a gold medal at the International Mathematical Olympiad,” Stanford HAI researchers point out, “but still can’t reliably tell time.” </p><h2><b>How models advanced in 2025</b></h2><p>Enterprise AI adoption has reached 88%. Notable accomplishments in 2025 and early 2026: </p><ul><li><p>Frontier models improved 30% in just one year on <a href="https://agi.safe.ai/">Humanity&#x27;s Last Exam</a> (HLE), which includes 2,500 questions across math, natural sciences, ancient languages, and other specialized subfields. HLE was built to be difficult for AI and favorable to human experts.</p></li><li><p>Leading models scored above 87% on MMLU-Pro, which tests multi-step reasoning based on 12,000 human-reviewed questions across more than a dozen disciplines. This illustrates “how competitive the frontier has become on broad knowledge tasks,” the Stanford HAI researchers note. </p></li><li><p>Top models including Claude Opus 4.5, GPT-5.2, and Qwen3.5 scored between 62.9% and 70.2% on τ-bench. The benchmark tests agents on real-world tasks in realistic domains that involve chatting with a user and calling external tools or APIs. </p></li><li><p>Model accuracy on GAIA, which benchmarks general AI assistants, rose from about 20% to 74.5%. </p></li><li><p>Agent performance on SWE-bench Verified rose from 60% to near 100% in just one year. The benchmark evaluates models on their ability to resolve real-world software issues. </p></li><li><p>Success rates on WebArena increased from 15% in 2023 to 74.3% in early 2026. This benchmark presents a realistic web environment for evaluating autonomous AI agents, tasking them with information retrieval, site navigation, and content configuration. </p></li><li><p>Agent performance progressed from 17% in 2024 to roughly 65% in early 2026 on MLE-bench, which evaluates machine learning (ML) engineering capabilities. </p></li></ul><p>AI agents are showing capability gains in cybersecurity. For instance, frontier models solved 93% of problems on Cybench, a benchmark that includes 40 professional-level tasks across six capture-the-flag categories, including cryptography, web security, reverse engineering, forensics, and exploitation. </p><p>This is compared to 15% in 2024 and represents the “steepest improvement rate,” indicating that cybersecurity tasks are a “good fit for current agent capabilities.”</p><p>Video generation has also evolved significantly over the last year; models can now capture how objects behave. For instance, Google DeepMind’s Veo 3 was tested across more than 18,000 generated videos, and demonstrated the ability to simulate buoyancy and solved mazes without having been trained on those tasks.</p><p>“Video generation models are no longer just producing realistic-looking content,” the researchers write. “Some are beginning to learn how the physical world actually works.” </p><p>Overall, AI is being used across a number of areas in enterprise — knowledge management, software engineering and IT, marketing and sales — and expanding into specialized domains like tax, mortgage processing, corporate finance, and legal reasoning, where accuracy ranges from 60 to 90%. </p><p>“AI capability is not plateauing,” Stanford HAI says. “It is accelerating and reaching more people than ever.”</p><h2><b>AI capability surges, but reliability lags</b></h2><p>
Multimodal models now meet or exceed human baselines on PhD-level science questions, multimodal reasoning, and competition mathematics. For example, Gemini Deep Think earned a gold medal at the 2025 International Mathematical Olympiad (IMO), solving five of six problems end-to-end in natural language within the 4.5-hour time limit — a notable improvement from a silver-level score in 2024.</p><p>Yet these same AI systems still fail in roughly one in three attempts, and have trouble with basic perception tasks, according to Stanford HAI. On ClockBench — a test covering 180 clock designs and 720 questions — Gemini Deep Think achieved only 50.1% accuracy, compared to roughly 90% for humans. GPT-4.5 High reached an almost identical score of 50.6%. </p><p>“Many multimodal models still struggle with something most humans find routine: Telling the time,” the Stanford HAI report points out. The seemingly simple task combines visual perception with simple arithmetic, identification of clock hands and their positions, and conversion of those into a time value. Ultimately, errors at any of these steps can cascade, leading to incorrect results, according to researchers. </p><p>In analysis, models were shown a range of clock styles: standard analog, clocks without a second hand, those with arrows as hands, others with black dials or Roman numerals. But even after fine-tuning on 5,000 synthetic images, models improved only on familiar formats and failed to generalize to real-world variations (like distorted dials or thinner hands). </p><p>Researchers extrapolated that, when models confused hour and minute hands, their ability to interpret direction deteriorated, suggesting that the challenge lies not just in data, but in integrating multiple visual cues.</p><p>“Even as models close the gap with human experts on knowledge-intensive tasks, this kind of visual reasoning remains a persistent challenge,” Stanford HAI notes. </p><h2><b>Hallucination and multi-step reasoning remain major gaps</b></h2><p>Even as models continue to accelerate in their reasoning, hallucinations remain a major concern. </p><p>In one benchmark, for instance, hallucination rates across 26 leading models ranged from 22% to 94%. Accuracy for some models dropped sharply when put under scrutiny —for example, GPT-4o&#x27;s accuracy slid from 98.2% to 64.4%, and DeepSeek R1 plummeted from more than 90% to 14.4%. </p><p>On the other hand, Grok 4.20 Beta, Claude 4.5 Haiku, and MiMo-V2-Pro showed the lowest rates.</p><p>Further, models continue to struggle with multi-step workflows, even as they are tasked with more of them. For example, on the τ-bench benchmark — which evaluates tool use and multi-turn reasoning — no model exceeded 71%, suggesting that “managing multiturn conversations while correctly using tools and following policy constraints remains difficult even for frontier models,” according to the Stanford HAI report. </p><h2><b>Models are becoming opaque</b></h2><p>Leading models are now “nearly indistinguishable” from each other when it comes to performance, the Stanford HAI report notes. Open-weight models are more competitive than ever, but they are converging. </p><p>As capability is no longer a “clear differentiator,” competitive pressure is shifting toward cost, reliability, and real-world usefulness. </p><p>Frontier labs are disclosing less information about their models, evaluation methods are quickly losing relevance, and independent testing can’t always corroborate developer-reported metrics. </p><p>As Stanford HAI points out: “The most capable systems are now the least transparent.”</p><p>Training code, parameter counts, dataset sizes, and durations are often being withheld — by firms including OpenAI, Anthropic and Google. And transparency is declining more broadly: In 2025, 80 out of 95 models were released without corresponding training code, while only four made their code fully open source.</p><p>Further, after rising between 2023 and 2024, scores on the <a href="https://crfm.stanford.edu/fmti/December-2025/index.html">Foundation Model Transparency Index</a> — which ranks major foundation developers on 100 transparency indicators — have since dropped. The average score is now 40, representing a 17 point decrease. </p><p>“Major gaps persist in disclosure around training data, compute resources, and post-deployment impact,” according to the report. </p><h2><b>Benchmarking AI is getting harder — and less reliable</b></h2><p>The benchmarks used to measure AI progress are facing growing reliability issues, with error rates reaching as high as 42% on widely-used evaluations. “AI is being tested more ambitiously across reasoning, safety, and real-world task execution,” the Stanford report notes, yet “those measurements are increasingly difficult to rely on.” </p><p>Key challenges include:</p><ul><li><p>“Sparse and declining” reporting on bias from developers </p></li><li><p>Benchmark contamination, or when models are exposed to test data; this can lead to “falsely inflated” scores</p></li><li><p>Discrepancies between developer-reported results and independent testing</p></li><li><p>“Poorly constructed” evals lacking documentation, details on statistical significance and reproducible scripts</p></li><li><p>“Growing opacity and non-standard prompting” that make model-to-model comparisons unreliable</p></li></ul><p>“Even when benchmark scores are technically valid, strong benchmark performance does not always translate to real-world utility,” according to the report. Further, “AI capability is outpacing the benchmarks designed to measure it.”</p><p>This is leading to “benchmark saturation,” where models achieve scores so high that tests can no longer differentiate between them. More complex, interactive forms of intelligence are becoming increasingly difficult to benchmark. Some are calling for evals that measure human-AI collaboration, rather than AI performance in isolation, but this technique is early in development. </p><p>“Evaluations intended to be challenging for years are saturated in months, compressing the window in which benchmarks remain useful for tracking progress,” according to Stanford HAI. </p><h2><b>Are we at &quot;peak data&quot;?</b></h2><p>As builders move into more data-intensive inference, there is growing concern about data bottlenecks and scaling sustainability. Leading researchers are warning that the available pool of high-quality human text and web data has been “exhausted” — a state referred to as “peak data.”</p><p>Hybrid approaches combining real and synthetic data can “significantly accelerate training” — sometimes by a factor of 5 to 10 — and smaller models trained on purely synthetic data have shown promise for narrowly defined tasks like classification or code generation, according to Stanford HAI. </p><p>Synthetically generated data can be effective for improving model performance in post-training settings, including fine-tuning, alignment, instruction tuning, and reinforcement learning (RL), the report notes. However, “these gains have not generalized to large, general-purpose language models.”</p><p>Rather than scaling data “indiscriminately,” researchers are turning to pruning, curating, and refining inputs, and are improving performance by cleaning labels, deduplicating samples, and constructing overall higher-quality datasets.</p><p>“Discussions on data availability often overlook an important shift in recent AI research,” according to the report. “Performance gains are increasingly driven by improving the quality of existing datasets, not by acquiring more.” </p><h2><b>Responsible AI is falling behind</b></h2><p>While the infrastructure for responsible AI is growing, progress has been “uneven” and is unable to keep pace with rapid capability gains, according to Stanford HAI. </p><p>While almost all leading frontier AI model developers report results on capability benchmarks, corresponding reporting on safety and responsibility is inconsistent and “spotty.”</p><p>Documented AI incidents rose significantly year over year — 362 in 2025 compared to 233 in 2024. And, while several frontier models received “Very Good” or “Good” safety ratings under standard use (per the <a href="https://mlcommons.org/ailuminate/">AILuminate benchmark</a>, which assesses generative AI across 12 “hazard” categories), safety performance dropped across all models when tested against jailbreak attempts using adversarial prompts. </p><p>“AI models perform well on safety tests under normal conditions, but their defenses weaken under deliberate attack,” Stanford HAI notes. </p><p>Adding to this challenge, builders have reported that improving one dimension, such as safety, can degrade another, like accuracy. “The infrastructure for responsible AI is growing, but progress has been uneven, and it is not keeping pace with the speed of AI deployment,” according to Stanford researchers. </p><p>The Stanford data makes one thing clear: the gap that matters in 2026 isn&#x27;t between AI and human performance. It&#x27;s between what AI can do in a demo and what it does reliably in production. Right now — with less transparency from the labs and benchmarks that saturate before they&#x27;re useful — that gap is harder to measure than ever.</p>]]></description>
            <author>taryn.plumb@venturebeat.com (Taryn Plumb)</author>
            <category>Security</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/4z68iLEpqAtIY5EAKVlov3/eda1e72826a8322c9042c0c5f1fe7726/AI_question.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[43% of AI-generated code changes need debugging in production, survey finds]]></title>
            <link>https://venturebeat.com/technology/43-of-ai-generated-code-changes-need-debugging-in-production-survey-finds</link>
            <guid isPermaLink="false">5bWleqe2LUYYQuqucsMzV1</guid>
            <pubDate>Tue, 14 Apr 2026 13:00:00 GMT</pubDate>
            <description><![CDATA[<p>The software industry is racing to write code with artificial intelligence. It is struggling, badly, to make sure that code holds up once it ships.</p><p>A survey of 200 senior site-reliability and DevOps leaders at large enterprises across the United States, United Kingdom, and European Union paints a stark picture of the hidden costs embedded in the AI coding boom. According to <a href="https://lightrun.com/ebooks/state-of-ai-powered-engineering-2026/">Lightrun&#x27;s 2026 State of AI-Powered Engineering Report</a>, shared exclusively with VentureBeat ahead of its public release, 43% of AI-generated code changes require manual debugging in production environments even after passing quality assurance and staging tests. Not a single respondent said their organization could verify an AI-suggested fix with just one redeploy cycle; 88% reported needing two to three cycles, while 11% required four to six.</p><p>The findings land at a moment when AI-generated code is proliferating across global enterprises at a breathtaking pace. Both <a href="https://www.cnbc.com/2025/04/29/satya-nadella-says-as-much-as-30percent-of-microsoft-code-is-written-by-ai.html">Microsoft CEO Satya Nadella</a> and <a href="https://arstechnica.com/ai/2024/10/google-ceo-says-over-25-of-new-google-code-is-generated-by-ai/">Google CEO Sundar Pichai</a> have claimed that around a quarter of their companies&#x27; code is now AI-generated. The AIOps market — the ecosystem of platforms and services designed to manage and monitor these AI-driven operations — stands at $18.95 billion in 2026 and is projected to reach $37.79 billion by 2031.</p><p>Yet the report suggests the infrastructure meant to catch AI-generated mistakes is badly lagging behind AI&#x27;s capacity to produce them.</p><p>&quot;The 0% figure signals that engineering is hitting a trust wall with AI adoption,&quot; said Or Maimon, Lightrun&#x27;s chief business officer, referring to the survey&#x27;s finding that zero percent of engineering leaders described themselves as &quot;very confident&quot; that AI-generated code will behave correctly once deployed. &quot;While the industry&#x27;s emphasis on increased productivity has made AI a necessity, we are seeing a direct negative impact. As AI-generated code enters the system, it doesn&#x27;t just increase volume; it slows down the entire deployment pipeline.&quot;</p><h2><b>Amazon&#x27;s March outages showed what happens when AI-generated code ships without safeguards</b></h2><p>The dangers are no longer theoretical. In early March 2026, Amazon suffered a series of <a href="https://www.reuters.com/business/retail-consumer/amazon-down-thousands-users-us-downdetector-shows-2026-03-05/">high-profile outages</a> that underscored exactly the kind of failure pattern the Lightrun survey describes. On March 2, Amazon.com experienced a disruption lasting nearly six hours, resulting in 120,000 lost orders and 1.6 million website errors. Three days later, on March 5, a more <a href="https://www.cnbc.com/2026/03/05/amazon-online-store-suffers-outage-for-some-users.html">severe outage hit the storefront</a> — lasting six hours and causing a 99% drop in U.S. order volume, with approximately 6.3 million lost orders. Both incidents were traced to AI-assisted code changes deployed to production without proper approval.</p><p>The fallout was swift. Amazon launched a 90-day code safety reset across 335 critical systems, and AI-assisted code changes must now be approved by senior engineers before they are deployed.</p><p>Maimon pointed directly to the Amazon episodes. &quot;This uncertainty isn&#x27;t based on a hypothesis,&quot; he said. &quot;We just need to look back to the start of March, when Amazon.com in North America went down due to an AI-assisted change being implemented without established safeguards.&quot;</p><p>The Amazon incidents illustrate the central tension the Lightrun report quantifies in survey data: AI tools can produce code at unprecedented speed, but the systems designed to validate, monitor, and trust that code in live environments have not kept pace. Google&#x27;s own <a href="https://cloud.google.com/blog/products/ai-machine-learning/announcing-the-2025-dora-report">2025 DORA report</a> corroborates this dynamic, finding that AI adoption correlates with an increase in code instability, and that 30% of developers report little or no trust in AI-generated code.</p><p>Maimon cited that research directly: &quot;Google&#x27;s 2025 DORA report found that AI adoption correlates with an almost 10% increase in code instability. Our validation processes were built for the scale of human engineering, but today, engineers have become auditors for massive volumes of unfamiliar code.&quot;</p><h2><b>Developers are losing two days a week to debugging AI-generated code they didn&#x27;t write</b></h2><p>One of the report&#x27;s most striking findings is the scale of human capital being consumed by AI-related verification work. Developers now spend an average of 38% of their work week — roughly two full days — on debugging, verification, and environment-specific troubleshooting, according to the survey. For 88% of the companies polled, this &quot;reliability tax&quot; consumes between 26% and 50% of their developers&#x27; weekly capacity.</p><p>This is not the productivity dividend that enterprise leaders expected when they invested in AI coding assistants. Instead, the engineering bottleneck has simply migrated. Code gets written faster, but it takes far longer to confirm that it works.</p><p>&quot;In some senses, AI has made the debugging problem worse,&quot; Maimon said. &quot;The volume of change is overwhelming human validation, while the generated code itself frequently does not behave as expected when deployed in Production. AI coding agents cannot see how their code behaves in running environments.&quot;</p><p>The redeploy problem compounds the time drain. Every surveyed organization requires multiple deployment cycles to verify a single AI-suggested fix — and according to Google&#x27;s <a href="https://cloud.google.com/blog/products/ai-machine-learning/announcing-the-2025-dora-report">2025 DORA report</a>, a single redeploy cycle takes a day to one week on average. In regulated industries such as healthcare and finance, deployment windows are often narrow, governed by mandated code freezes and strict change-management protocols. Requiring three or more cycles to validate a single AI fix can push resolution timelines from days to weeks.</p><p>Maimon rejected the idea that these multiple cycles represent prudent engineering discipline. &quot;This is not discipline, but an expensive bottleneck and a symptom of the fact that AI-generated fixes are often unreliable,&quot; he said. &quot;If we can move from three cycles to one, we reclaim a massive portion of that 38% lost engineering capacity.&quot;</p><h2><b>AI monitoring tools can&#x27;t see what&#x27;s happening inside running applications — and that&#x27;s the real problem</b></h2><p>If the productivity drain is the most visible cost, the <a href="https://lightrun.com/ebooks/state-of-ai-powered-engineering-2026/">Lightrun report</a> argues the deeper structural problem is what it calls &quot;the runtime visibility gap&quot; — the inability of AI tools and existing monitoring systems to observe what is actually happening inside running applications.</p><p>Sixty percent of the survey&#x27;s respondents identified a lack of visibility into live system behavior as the primary bottleneck in resolving production incidents. In 44% of cases where AI SRE or application performance monitoring tools attempted to investigate production issues, they failed because the necessary execution-level data — variable states, memory usage, request flow — had never been captured in the first place.</p><p>The report paints a picture of AI tools operating essentially blind in the environments that matter most. Ninety-seven percent of engineering leaders said their AI SRE agents operate without significant visibility into what is actually happening in production. Approximately half of all companies (49%) reported their AI agents have only limited visibility into live execution states. Only 1% reported extensive visibility, and not a single respondent claimed full visibility.</p><p>This is the gap that turns a minor software bug into a costly outage. When an AI-suggested fix fails in production — as 43% of them do — engineers cannot rely on their AI tools to diagnose the problem, because those tools cannot observe the code&#x27;s real-time behavior. Instead, teams fall back on what the report calls &quot;tribal knowledge&quot;: the institutional memory of senior engineers who have seen similar problems before and can intuit the root cause from experience rather than data. The survey found that 54% of resolutions to high-severity incidents rely on tribal knowledge rather than diagnostic evidence from AI SREs or APMs.</p><h2><b>In finance, 74% of engineering teams trust human intuition over AI diagnostics during serious incidents</b></h2><p>The trust deficit plays out with particular intensity in the finance sector. In an industry where a single application error can cascade into millions of dollars in losses per minute, the survey found that 74% of financial-services engineering teams rely on tribal knowledge over automated diagnostic data during serious incidents — far higher than the 44% figure in the technology sector.</p><p>&quot;Finance is a heavily regulated, high-stakes environment where a single application error can cost millions of dollars per minute,&quot; Maimon said. &quot;The data shows that these teams simply do not trust AI not to make a dangerous mistake in their Production environments. This is a rational response to tool failure.&quot;</p><p>The distrust extends beyond finance. Perhaps the most telling data point in the entire report is that not a single organization surveyed — across any industry — has moved its AI SRE tools into actual production workflows. Ninety percent remain in experimental or pilot mode. The remaining 10% evaluated AI SRE tools and chose not to adopt them at all. This represents an extraordinary gap between market enthusiasm and operational reality: enterprises are spending aggressively on AI for IT operations, but the tools they are buying remain quarantined from the environments where they would deliver the most value.</p><p>Maimon described this as one of the report&#x27;s most significant revelations. &quot;Leaders are eager to adopt these new AI tools, but they don&#x27;t trust AI to touch live environments,&quot; he said. &quot;The lack of trust is shown in the data; 98% have lower trust in AI operating in production than in coding assistants.&quot;</p><h2><b>The observability industry built for human-speed engineering is falling short in the age of AI</b></h2><p>The findings raise pointed questions about the current generation of observability tools from major vendors like <a href="https://www.datadoghq.com/">Datadog</a>, <a href="https://www.dynatrace.com/">Dynatrace</a>, and <a href="https://www.splunk.com/">Splunk</a>. Seventy-seven percent of the engineering leaders surveyed reported low or no confidence that their current observability stack provides enough information to support autonomous root cause analysis or automated incident remediation.</p><p>Maimon did not shy away from naming the structural problem. &quot;Major vendors often build &#x27;closed-garden&#x27; ecosystems where their AI SREs can only reason over data collected by their own proprietary agents,&quot; he said. &quot;In a modern enterprise, teams typically have a multi-tool stack to provide full coverage. By forcing a team into a single-vendor silo, these tools create an uncomfortable dependency and a strategic liability: if the vendor&#x27;s data coverage is missing a specific layer, the AI is effectively blind to the root cause.&quot;</p><p>The second issue, Maimon argued, is that current observability-backed AI SRE solutions offer only partial visibility — defined by what engineers thought to log at the time of deployment. Because failures rarely follow predefined paths, autonomous root cause analysis using only these tools will frequently miss the key diagnostic evidence. &quot;To move toward true autonomous remediation,&quot; he said, &quot;the industry must shift toward AI SRE without vendor lock-in; AI SREs must be an active participant that can connect across the entire stack and interrogate live code to capture the ground truth of a failure as it happens.&quot;</p><p>When asked what it would take to trust AI SREs, the survey&#x27;s respondents coalesced unanimously around live runtime visibility. Fifty-eight percent said they need the ability to provide &quot;evidence traces&quot; of variables at the point of failure, and 42% cited the ability to verify a suggested fix before it actually deploys. No respondents selected the ability to ingest multiple log sources or provide better natural language explanations — suggesting that engineering leaders do not want AI that talks better, but AI that can see better.</p><h2><b>The question is no longer whether to use AI for coding — it&#x27;s whether anyone can trust what it produces</b></h2><p>The <a href="https://lightrun.com/ebooks/state-of-ai-powered-engineering-2026/">survey</a> was administered by <a href="https://surveyz.io/">Global Surveyz Research</a>, an independent firm, and drew responses from Directors, VPs, and C-level executives in SRE and DevOps roles at enterprises with 1,500 or more employees across the finance, technology, and information technology sectors. Responses were collected during January and February 2026, with questions randomized to prevent order bias.</p><p><a href="https://lightrun.com/">Lightrun</a>, which is backed by $110 million in funding from Accel and Insight Partners and counts <a href="https://www.att.com/">AT&amp;T</a>, <a href="https://www.citi.com/">Citi</a>, <a href="https://www.microsoft.com/en-us">Microsoft</a>, <a href="https://www.salesforce.com/">Salesforce</a>, and <a href="https://www.unitedhealthgroup.com/">UnitedHealth Group</a> among its enterprise clients, has a clear commercial interest in the problem the report describes: the company sells a runtime observability platform designed to give AI agents and human engineers real-time visibility into live code execution. Its AI SRE product uses a Model Context Protocol connection to generate live diagnostic evidence at the point of failure without requiring redeployment. That commercial interest does not diminish the survey&#x27;s findings, which align closely with independent research from Google DORA and the real-world evidence of the Amazon outages.</p><p>Taken together, they describe an industry confronting an uncomfortable paradox. AI has solved the slowest part of building software — writing the code — only to reveal that writing was never the hard part. The hard part was always knowing whether it works. And on that question, the engineers closest to the problem are not optimistic.</p><p>&quot;If the live visibility gap is not closed, then teams are really just compounding instability through their adoption of AI,&quot; Maimon said. &quot;Organizations that don&#x27;t bridge this gap will find themselves stuck with long redeploy loops, to solve ever more complex challenges. They will lose their competitive speed to the very AI tools that were meant to provide it.&quot;</p><p>The machines learned to write the code. Nobody taught them to watch it run.</p>]]></description>
            <author>michael.nunez@venturebeat.com (Michael Nuñez)</author>
            <category>Technology</category>
            <category>Security</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/5nAHuSU7TlSixVhQbV3Zpy/f97f9591cd1d877db961dac2be53b6cc/nuneybits_Vector_art_of_developer_mopping_code_spill_dbcceaac-fb6e-4e63-90cf-5774d34a0f44.webp?w=300&amp;q=30" length="0" type="image/webp"/>
        </item>
        <item>
            <title><![CDATA[Five signs data drift is already undermining your security models]]></title>
            <link>https://venturebeat.com/security/five-signs-data-drift-is-already-undermining-your-security-models</link>
            <guid isPermaLink="false">7i0FSJc3DXou1wEJSVEmkG</guid>
            <pubDate>Sun, 12 Apr 2026 19:00:00 GMT</pubDate>
            <description><![CDATA[<p>Data drift happens when the statistical properties of a machine learning (ML) model&#x27;s input data change over time, eventually rendering its predictions less accurate. <a href="https://venturebeat.com/security/ocsf-explained-the-shared-data-language-security-teams-have-been-missing?_gl=1*yt0z35*_up*MQ..*_ga*MTcxNTczODYxLjE3NzYwMDUzOTE.*_ga_B8TDS1LEXQ*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw*_ga_SCH1J7LNKY*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw">Cybersecurity professionals</a> who rely on ML for tasks like malware detection and network threat analysis find that undetected data drift can create vulnerabilities. A model trained on old attack patterns may fail to see today&#x27;s sophisticated threats. Recognizing the early signs of data drift is the first step in maintaining reliable and efficient security systems.</p><h2><b>Why data drift compromises security models</b></h2><p>ML models are trained on a snapshot of historical data. When live data no longer resembles this snapshot, the model&#x27;s performance dwindles, creating a <a href="https://venturebeat.com/technology/why-cios-must-lead-ai-experimentation-not-just-govern-it?_gl=1*x7qiq4*_up*MQ..*_ga*MTcxNTczODYxLjE3NzYwMDUzOTE.*_ga_B8TDS1LEXQ*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw*_ga_SCH1J7LNKY*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw">critical cybersecurity risk</a>. A threat detection model may generate more false negatives by missing real breaches or create more false positives, leading to alert fatigue for security teams.</p><p>Adversaries actively exploit this weakness. In 2024,<a href="https://thehackernews.com/2024/07/proofpoint-email-routing-flaw-exploited.html"> <u>attackers used echo-spoofing techniques</u></a> to bypass email protection services. By exploiting misconfigurations in the system, they sent millions of spoofed emails that evaded the vendor&#x27;s ML classifiers. This incident demonstrates how threat actors can manipulate input data to exploit blind spots. When a security model fails to adapt to shifting tactics, it becomes a liability.</p><h2><b>5 indicators of data drift</b></h2><p>Security professionals can recognize the presence of drift (or its potential) in several ways.</p><h3><b>1. A sudden drop in model performance</b></h3><p>Accuracy, precision, and recall are often the first casualties. A consistent decline in these key metrics is a red flag that the model is no longer in sync with the current threat landscape.</p><p>Consider Klarna&#x27;s success: Its AI assistant handled 2.3 million customer service conversations in its first month and performed work equivalent to 700 agents. This efficiency drove a<a href="https://www.nutshell.com/blog/best-ai-chatbots"> <u>25% decline in repeat inquiries</u></a> and reduced resolution times to under two minutes. </p><p>Now imagine if those parameters suddenly reversed because of drift. In a security context, a similar drop in performance does not just mean unhappy clients — it also means successful intrusions and potential data exfiltration.</p><h3><b>2. Shifts in statistical distributions</b></h3><p><a href="https://venturebeat.com/security/human-centric-iam-is-failing-agentic-ai-requires-a-new-identity-control?_gl=1*61shbb*_up*MQ..*_ga*MTcxNTczODYxLjE3NzYwMDUzOTE.*_ga_B8TDS1LEXQ*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw*_ga_SCH1J7LNKY*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw">Security teams</a> should monitor the core statistical properties of input features, such as the mean, median, and standard deviation. A significant change in these metrics from training data could indicate the underlying data has changed.</p><p>Monitoring for such shifts enables teams to catch drift before it causes a breach. For example, a phishing detection model might be trained on emails with an average attachment size of 2MB. If the average attachment size suddenly jumps to 10MB due to a new malware-delivery method, the model may fail to classify these emails correctly.</p><h3><b>3. Changes in prediction behavior</b></h3><p>Even if overall accuracy seems stable, distributions of predictions might change, a phenomenon often referred to as prediction drift.</p><p>For instance, if a fraud detection model historically flagged 1% of transactions as suspicious but suddenly starts flagging 5% or 0.1%, either something has shifted or the nature of the input data has changed. It might indicate a new type of attack that confuses the model or a change in legitimate user behavior that the model was not trained to identify.</p><h3><b>4. An increase in model uncertainty</b></h3><p>For models that provide a confidence score or probability with their predictions, a general decrease in confidence can be a subtle sign of drift.</p><p>Recent studies highlight the<a href="https://arxiv.org/html/2410.21952v2"> <u>value of uncertainty quantification</u></a> in detecting adversarial attacks. If the model becomes less sure about its forecasts across the board, it is likely facing data it was not trained on. In a cybersecurity setting, this uncertainty is an early sign of potential model failure, suggesting the model is operating in unfamiliar ground and that its decisions might no longer be reliable.</p><h3><b>5. Changes in feature relationships</b></h3><p>The correlation between different input features can also change over time. In a network intrusion model, traffic volume and packet size might be highly linked during normal operations. If that correlation disappears, it can signal a change in network behavior that the model may not understand. A sudden feature decoupling could indicate a new tunneling tactic or a stealthy exfiltration attempt.</p><h2><b>Approaches to detecting and mitigating data drift</b></h2><p>Common detection methods include the Kolmogorov-Smirnov (KS) and the population stability index (PSI). These compare the <a href="https://towardsdatascience.com/drift-detection-in-robust-machine-learning-systems/"><u>distributions of live and training data</u></a> to identify deviations. The KS test determines if two datasets differ significantly, while the PSI measures how much a variable&#x27;s distribution has shifted over time. </p><p>The mitigation method of choice often depends on how the drift manifests, as distribution changes may occur suddenly. For example, customers&#x27; buying behavior may change overnight with the launch of a new product or a promotion. In other cases, drift may occur gradually over a more extended period. That said, security teams must learn to adjust their monitoring cadence to capture both rapid spikes and slow burns. Mitigation will involve retraining the model on more recent data to reclaim its effectiveness.</p><h2><b>Proactively manage drift for stronger security</b></h2><p>Data drift is an inevitable reality, and cybersecurity teams can maintain a strong security posture by treating detection as a continuous and automated process. Proactive monitoring and model retraining are fundamental practices to ensure ML systems remain reliable allies against developing threats.</p><p><i>Zac Amos is the Features Editor at </i><a href="https://rehack.com/"><i><u>ReHack</u></i></a><i>.</i></p>]]></description>
            <category>Security</category>
            <category>DataDecisionMakers</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/erAw6FrOeAX9eZJqeF2Dx/3a759d02f32a698bdc815c787701a17a/AI_drift.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[Your developers are already running AI locally: Why on-device inference is the CISO’s new blind spot]]></title>
            <link>https://venturebeat.com/security/your-developers-are-already-running-ai-locally-why-on-device-inference-is</link>
            <guid isPermaLink="false">3EC5GemarqXB92UGk1xUjb</guid>
            <pubDate>Sun, 12 Apr 2026 15:00:20 GMT</pubDate>
            <description><![CDATA[<p>For the last 18 months, the CISO playbook for generative AI has been relatively simple: Control the browser.</p><p><a href="https://venturebeat.com/security/ocsf-explained-the-shared-data-language-security-teams-have-been-missing?_gl=1*4903t3*_up*MQ..*_ga*MTcxNTczODYxLjE3NzYwMDUzOTE.*_ga_B8TDS1LEXQ*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw*_ga_SCH1J7LNKY*czE3NzYwMDUzODkkbzEkZzAkdDE3NzYwMDUzODkkajYwJGwwJGgw">Security teams</a> tightened cloud access security broker (CASB) policies, blocked or monitored traffic to well-known AI endpoints, and routed usage through sanctioned gateways. The operating model was clear: If sensitive data leaves the network for an external API call, we can observe it, log it, and stop it. But that model is starting to break.</p><p>A quiet hardware shift is pushing large language model (LLM) usage off the network and onto the endpoint. Call it Shadow AI 2.0, or the “bring your own model” (BYOM) era: Employees running capable models locally on laptops, offline, with no API calls and no obvious network signature. The governance conversation is still framed as “data exfiltration to the cloud,” but the more immediate enterprise risk is increasingly “unvetted inference inside the device.&quot;</p><p>When inference happens locally, traditional data loss prevention (DLP) doesn’t see the interaction. And when security can’t see it, it can’t manage it.</p><h3><b>Why local inference is suddenly practical</b></h3><p>Two years ago, running a useful LLM on a work laptop was a niche stunt. Today, it’s routine for technical teams.</p><p>Three things converged:</p><ul><li><p><b>Consumer-grade accelerators got serious: </b>A MacBook Pro with 64GB unified memory can often run quantized 70B-class models at usable speeds (with practical limits on context length). What once required multi-GPU servers is now feasible on a high-end laptop for many real workflows.</p></li><li><p><b>Quantization went mainstream:</b> It’s now easy to compress models into smaller, faster formats that fit within laptop memory often with acceptable quality tradeoffs for many tasks.</p></li><li><p><b>Distribution is frictionless:</b> Open-weight models are a single command away, and the tooling ecosystem makes “download → run → chat” trivial.</p></li></ul><p><b>The result: </b>An engineer can pull down a multi‑GB model artifact, turn off Wi‑Fi, and run sensitive workflows locally, source code review, document summarization, drafting customer communications, even exploratory analysis over regulated datasets. No outbound packets, no proxy logs, no cloud audit trail.</p><p>From a <a href="https://venturebeat.com/security/mythos-detection-ceiling-security-teams-new-playbook?_gl=1*qe97gz*_up*MQ..*_ga*MzY1OTQzODYzLjE3NzYwMDU1Mjk.*_ga_SCH1J7LNKY*czE3NzYwMDU1MjgkbzEkZzAkdDE3NzYwMDU1MjgkajYwJGwwJGgw*_ga_B8TDS1LEXQ*czE3NzYwMDU1MjgkbzEkZzAkdDE3NzYwMDU1MjgkajYwJGwwJGgw">network-security perspective</a>, that activity can look indistinguishable from “nothing happened”.</p><h3><b>The risk isn’t only data leaving the company anymore</b></h3><p>If the data isn’t leaving the laptop, why should a CISO care?</p><p>Because the dominant risks shift from exfiltration to integrity, provenance, and compliance. In practice, local inference creates three classes of blind spots that most enterprises have not operationalized.</p><h4><b>1. Code and decision contamination (integrity risk)</b></h4><p>Local models are often adopted because they’re fast, private, and “no approval required.&quot; The downside is that they’re frequently unvetted for the enterprise environment.</p><p><b>A common scenario:</b> A senior developer downloads a community-tuned coding model because it benchmarks well. They paste in internal auth logic, payment flows, or infrastructure scripts to “clean it up.&quot; The model returns output that looks competent, compiles, and passes unit tests, but subtly degrades security posture (weak input validation, unsafe defaults, brittle concurrency changes, dependency choices that aren’t allowed internally). The engineer commits the change.</p><p>If that interaction happened offline, you may have no record that AI influenced the code path at all. And when you later do incident response, you’ll be investigating the symptom (a vulnerability) without visibility into a key cause (uncontrolled model usage).</p><h4><b>2. Licensing and IP exposure (compliance risk)</b></h4><p>Many high-performing models ship with licenses that include <a href="https://llama.meta.com/llama3/license/"><u>restrictions on commercial use</u></a>, attribution requirements, field-of-use limits, or obligations that can be incompatible with proprietary product development. When employees run models locally, that usage can bypass the organization’s normal procurement and legal review process.</p><p>If a team uses a non-commercial model to generate production code, documentation, or product behavior, the company can inherit risk that shows up later during M&amp;A diligence, customer security reviews, or litigation. The hard part is not just the license terms, it’s the lack of inventory and traceability. Without a governed model hub or usage record, you may not be able to prove what was used where.</p><h4><b>3. Model supply chain exposure (provenance risk)</b></h4><p>Local inference also changes the software supply chain problem. Endpoints begin accumulating large model artifacts and the toolchains around them: ownloaders, converters, runtimes, plugins, UI shells, and Python packages.</p><p>There is a critical technical nuance here: The file format matters. While newer formats like <a href="https://huggingface.co/docs/safetensors/index"><b><u>Safetensors</u></b></a> are designed to prevent arbitrary code execution, older <a href="https://pytorch.org/docs/stable/generated/torch.load.html"><b><u>Pickle-based</u></b><u> PyTorch files</u></a> can execute malicious payloads simply when loaded. If your developers are grabbing unvetted checkpoints from Hugging Face or other repositories, they aren&#x27;t just downloading data — they could be downloading an exploit.</p><p>Security teams have spent decades learning to treat unknown executables as hostile. BYOM requires extending that mindset to model artifacts and the surrounding runtime stack. The biggest organizational gap today is that most companies have no equivalent of a <a href="https://www.cisa.gov/sbom"><u>software bill of materials</u></a> for models: Provenance, hashes, allowed sources, scanning, and lifecycle management.</p><h3><b>Mitigating BYOM: treat model weights like software artifacts</b></h3><p>You can’t solve local inference by blocking URLs. You need endpoint-aware controls and a developer experience that makes the safe path the easy path.</p><p>Here are three practical ways:</p><p><b>1. Move governance down to the endpoint</b> </p><p>Network DLP and CASB still matter for cloud usage, but they’re not sufficient for BYOM. Start treating local model usage as an endpoint governance problem by looking for specific signals:</p><ul><li><p><b>Inventory and detection:</b> Scan for high-fidelity indicators like .gguf files larger than 2GB, processes like <a href="https://github.com/ggerganov/llama.cpp"><u>llama.cpp</u></a> or Ollama, and local listeners on common <a href="https://docs.ollama.com/faq"><u>default port 11434</u></a>.</p></li><li><p><b>Process and runtime awareness:</b> Monitor for repeated high GPU/NPU (neural processing unit) utilization from unapproved runtimes or unknown local inference servers.</p></li><li><p><b>Device policy:</b> Use <b>mobile device management (MDM) and endpoint detection and response (EDR)</b> policies to control installation of unapproved runtimes and enforce baseline hardening on engineering devices. The point isn’t to punish experimentation. It’s to regain visibility.</p></li></ul><p><b>2. Provide a paved road: An internal, curated model hub</b> </p><p><a href="https://venturebeat.com/security/ai-agent-zero-trust-architecture-audit-credential-isolation-anthropic-nvidia-nemoclaw">Shadow AI</a> is often an outcome of friction. Approved tools are too restrictive, too generic, or too slow to approve. A better approach is to offer a curated internal catalog that includes: </p><ul><li><p>Approved models for common tasks (coding, summarization, classification)</p></li><li><p>Verified licenses and usage guidance</p></li><li><p>Pinned versions with hashes (prioritizing safer formats like Safetensors)</p></li><li><p>Clear documentation for safe local usage, including where sensitive data is and isn’t allowed. If you want developers to stop scavenging, give them something better.</p></li></ul><p><b>3. Update policy language: “Cloud services” isn’t enough anymore</b> </p><p>Most acceptable use policies talk about SaaS and cloud tools. BYOM requires policy that explicitly covers:</p><ul><li><p>Downloading and running model artifacts on corporate endpoints</p></li><li><p>Acceptable sources</p></li><li><p>License compliance requirements</p></li><li><p>Rules for using models with sensitive data</p></li><li><p>Retention and logging expectations for local inference tools This doesn’t need to be heavy-handed. It needs to be unambiguous.</p></li></ul><h3><b>The perimeter is shifting back to the device</b></h3><p>For a decade we moved security controls “up” into the cloud. Local inference is pulling a meaningful slice of AI activity back “down” to the endpoint.</p><p>5 signals shadow AI has moved to endpoints:</p><ul><li><p><b>Large model artifacts:</b> Unexplained storage consumption by .gguf or .pt files.</p></li><li><p><b>Local inference servers:</b> Processes listening on ports like 11434 (Ollama).</p></li><li><p><b>GPU utilization patterns:</b> Spikes in GPU usage while offline or disconnected from VPN.</p></li><li><p><b>Lack of model inventory:</b> Inability to map code outputs to specific model versions.</p></li><li><p><b>License ambiguity:</b> Presence of &quot;non-commercial&quot; model weights in production builds.</p></li></ul><p>Shadow AI 2.0 isn’t a hypothetical future, it’s a predictable consequence of fast hardware, easy distribution, and developer demand. CISOs who focus only on network controls will miss what’s happening on the silicon sitting right on employees’ desks.</p><p>The next phase of AI governance is less about blocking websites and more about controlling artifacts, provenance, and policy at the endpoint, without killing productivity.</p><p><i>Jayachander Reddy Kandakatla is a senior MLOps engineer.</i></p>]]></description>
            <category>Security</category>
            <category>DataDecisionMakers</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/pAoHef9hMVI3aHoyHfluC/f410fef5dc2a910939184a98db76eec4/AI_perimeter.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
        <item>
            <title><![CDATA[AI agent credentials live in the same box as untrusted code. Two new architectures show where the blast radius actually stops.]]></title>
            <link>https://venturebeat.com/security/ai-agent-zero-trust-architecture-audit-credential-isolation-anthropic-nvidia-nemoclaw</link>
            <guid isPermaLink="false">20HlmDVeOH0foEOboDiQiE</guid>
            <pubDate>Fri, 10 Apr 2026 22:26:53 GMT</pubDate>
            <description><![CDATA[<p>Four separate RSAC 2026 keynotes arrived at the same conclusion without coordinating. Microsoft&#x27;s Vasu Jakkal told attendees that zero trust must extend to AI. Cisco&#x27;s Jeetu Patel called for a shift from access control to action control, <a href="https://venturebeat.com/security/rsac-2026-agent-identity-frameworks-three-gaps">saying in an exclusive interview with VentureBeat</a> that agents behave &quot;more like teenagers, supremely intelligent, but with no fear of consequence.&quot; CrowdStrike&#x27;s George Kurtz identified AI governance as the biggest gap in enterprise technology. Splunk&#x27;s John Morgan called for an agentic trust and governance model. Four companies. Four stages. <a href="https://cloudsecurityalliance.org/blog/2026/04/03/every-rsac-keynote-asked-the-same-five-questions-here-s-the-framework-that-answers-them">One problem</a>.</p><p>Matt Caulfield, VP of Product for Identity and Duo at Cisco, put it bluntly in an exclusive VentureBeat interview at RSAC. &quot;While the concept of zero trust is good, we need to take it a step further,&quot; Caulfield said. &quot;It&#x27;s not just about authenticating once and then letting the agent run wild. It&#x27;s about continuously verifying and scrutinizing every single action the agent&#x27;s trying to take, because at any moment, that agent can go rogue.&quot;</p><p>Seventy-nine percent of organizations already use AI agents, according to <a href="https://www.pwc.com/us/en/tech-effect/ai-analytics/ai-agent-survey.html">PwC&#x27;s 2025 AI Agent Survey</a>. Only 14.4% reported full security approval for their entire agent fleet, per the <a href="https://www.gravitee.io/state-of-ai-agent-security">Gravitee State of AI Agent Security 2026 report</a> of 919 organizations in February 2026. A <a href="https://cloudsecurityalliance.org/artifacts/the-state-of-ai-security-and-governance">CSA survey</a> presented at RSAC found that only 26% have AI governance policies. <a href="https://cloudsecurityalliance.org/blog/2026/02/02/the-agentic-trust-framework-zero-trust-governance-for-ai-agents">CSA&#x27;s Agentic Trust Framework</a> describes the resulting gap between deployment velocity and security readiness as a governance emergency.</p><p>Cybersecurity leaders and industry executives at RSAC agreed on the problem. Then two companies shipped architectures that answer the question differently. The gap between their designs reveals where the real risk sits.</p><h2>The monolithic agent problem that security teams are inheriting</h2><p>The default enterprise agent pattern is a monolithic container. The model reasons, calls tools, executes generated code, and holds credentials in one process. Every component trusts every other component. OAuth tokens, API keys, and git credentials sit in the same environment where the agent runs code it wrote seconds ago.</p><p>A prompt injection gives the attacker everything. Tokens are exfiltrable. Sessions are spawnable. The blast radius is not the agent. It is the entire container and every connected service.</p><p>The <a href="https://cloudsecurityalliance.org/press-releases/2026/03/24/more-than-two-thirds-of-organizations-cannot-clearly-distinguish-ai-agent-from-human-actions">CSA and Aembit survey</a> of 228 IT and security professionals quantifies how common this remains: 43% use shared service accounts for agents, 52% rely on workload identities rather than agent-specific credentials, and 68% cannot distinguish agent activity from human activity in their logs. No single function claimed ownership of AI agent access. Security said it was a developer&#x27;s responsibility. Developers said it was a security responsibility. Nobody owned it.</p><p>CrowdStrike CTO Elia Zaitsev, in an exclusive VentureBeat interview, said the pattern should look familiar. &quot;A lot of what securing agents look like would be very similar to what it looks like to secure highly privileged users. They have identities, they have access to underlying systems, they reason, they take action,&quot; Zaitsev said. &quot;There&#x27;s rarely going to be one single solution that is the silver bullet. It&#x27;s a defense in depth strategy.&quot;</p><p>CrowdStrike CEO George Kurtz highlighted ClawHavoc (a supply chain campaign targeting the OpenClaw agentic framework) at RSAC during his <a href="https://venturebeat.com/security/rsac-2026-agentic-soc-agent-telemetry-security-gap">keynote</a>. <a href="https://www.koi.ai/blog/clawhavoc-341-malicious-clawedbot-skills-found-by-the-bot-they-were-targeting">Koi Security</a> named the campaign on February 1, 2026. Antiy CERT confirmed 1,184 malicious skills tied to 12 publisher accounts, according to <a href="https://vpncentral.com/clawhavoc-poisons-openclaw-clawhub-with-1184-malicious-skills/">multiple</a> <a href="https://venturebeat.com/security/rsac-2026-agentic-soc-agent-telemetry-security-gap">independent</a> analyses of the campaign. <a href="https://snyk.io/blog/toxicskills-malicious-ai-agent-skills-clawhub/">Snyk&#x27;s ToxicSkills research</a> found that 36.8% of the 3,984 ClawHub skills scanned contain security flaws at any severity level, with 13.4% rated critical. Average breakout time has dropped to 29 minutes. Fastest observed: 27 seconds. (<a href="https://www.crowdstrike.com/en-us/press-releases/2026-crowdstrike-global-threat-report/">CrowdStrike 2026 Global Threat Report</a>)</p><h2>Anthropic separates the brain from the hands</h2><p><a href="https://www.anthropic.com/engineering/managed-agents">Anthropic&#x27;s Managed Agents</a>, launched April 8 in public beta, split every agent into three components that do not trust each other: a brain (Claude and the harness routing its decisions), hands (disposable Linux containers where code executes), and a session (an append-only event log outside both).</p><p>Separating instructions from execution is one of the oldest patterns in software. Microservices, serverless functions, and message queues. </p><p>Credentials never enter the sandbox. Anthropic stores OAuth tokens in an external vault. When the agent needs to call an MCP tool, it sends a session-bound token to a dedicated proxy. The proxy fetches real credentials from the vault, makes the external call, and returns the result. The agent never sees the actual token. Git tokens get wired into the local remote at sandbox initialization. Push and pull work without the agent touching the credential. For security directors, this means a compromised sandbox yields nothing an attacker can reuse.</p><p>The security gain arrived as a side effect of a performance fix. Anthropic decoupled the brain from the hands so inference could start before the container booted. Median time to first token <a href="https://www.anthropic.com/engineering/managed-agents">dropped roughly 60%</a>. The zero-trust design is also the fastest design. That kills the enterprise objection that security adds latency.</p><p>Session durability is the third structural gain. A container crash in the monolithic pattern means total state loss. In Managed Agents, the session log persists outside both brain and hands. If the harness crashes, a new one boots, reads the event log, and resumes. No state lost turns into a productivity gain over time. Managed Agents include built-in session tracing through the <a href="https://www.anthropic.com/engineering/managed-agents">Claude Console</a>.</p><p>Pricing: $0.08 per session-hour of active runtime, idle time excluded, plus standard API token costs. Security directors can now model agent compromise cost per session-hour against the cost of the architectural controls.</p><h2>Nvidia locks the sandbox down and monitors everything inside it</h2><p><a href="https://github.com/NVIDIA/NemoClaw">Nvidia&#x27;s NemoClaw</a>, released March 16 in early preview, takes the opposite approach. It does not separate the agent from its execution environment. It wraps the entire agent inside four stacked security layers and watches every move. Anthropic and Nvidia are the only two vendors to have shipped zero-trust agent architectures publicly as of this writing; others are in development.</p><p>NemoClaw stacks five enforcement layers between the agent and the host. Sandboxed execution uses Landlock, seccomp, and network namespace isolation at the kernel level. Default-deny outbound networking forces every external connection through explicit operator approval via <a href="https://docs.nvidia.com/nemoclaw/latest/reference/network-policies.html">YAML-based policy</a>. Access runs with minimal privileges. A privacy router directs sensitive queries to locally-running Nemotron models, cutting token cost and data leakage to zero. The layer that matters most to security teams is intent verification: OpenShell&#x27;s policy engine intercepts every agent action before it touches the host. The trade-off for organizations evaluating NemoClaw is straightforward. Stronger runtime visibility costs more operator staffing.</p><p>The agent does not know it is inside NemoClaw. In-policy actions return normally. Out-of-policy actions get a configurable denial.</p><p>Observability is the strongest layer. A real-time Terminal User Interface logs every action, every network request, every blocked connection. The audit trail is complete. The problem is cost: operator load scales linearly with agent activity. Every new endpoint requires manual approval. Observation quality is high. Autonomy is low. That ratio gets expensive fast in production environments running dozens of agents.</p><p>Durability is the gap nobody&#x27;s talking about. Agent state persists as files inside the sandbox. If the sandbox fails, the state goes with it. No external session recovery mechanism exists. Long-running agent tasks carry a durability risk that security teams need to price into deployment planning before they hit production.</p><h2>The credential proximity gap</h2><p>Both architectures are a real step up from the monolithic default. Where they diverge is the question that matters most to security teams: how close do credentials sit to the execution environment?</p><p>Anthropic removes credentials from the blast radius entirely. If an attacker compromises the sandbox through prompt injection, they get a disposable container with no tokens and no persistent state. Exfiltrating credentials requires a two-hop attack: influence the brain&#x27;s reasoning, then convince it to act through a container that holds nothing worth stealing. Single-hop exfiltration is structurally eliminated.</p><p>NemoClaw constrains the blast radius and monitors every action inside it. Four security layers limit lateral movement. Default-deny networking blocks unauthorized connections. But the agent and generated code share the same sandbox. Nvidia&#x27;s privacy router keeps inference credentials on the host, outside the sandbox. But messaging and integration tokens (Telegram, Slack, Discord) are injected into the sandbox as runtime environment variables. Inference API keys are proxied through the privacy router and not passed into the sandbox directly. The exposure varies by credential type. Credentials are policy-gated, not structurally removed.</p><p>That distinction matters most for indirect prompt injection, where an adversary embeds instructions in content the agent queries as part of legitimate work. A poisoned web page. A manipulated API response. The intent verification layer evaluates what the agent proposes to do, not the content of data returned by external tools. Injected instructions enter the reasoning chain as trusted context. With proximity to execution.</p><p>In the Anthropic architecture, indirect injection can influence reasoning but cannot reach the credential vault. In the NemoClaw architecture, injected context sits next to both reasoning and execution inside the shared sandbox. That is the widest gap between the two designs.</p><p>NCC Group&#x27;s David Brauchler, Technical Director and Head of AI/ML Security, <a href="https://www.esecurityplanet.com/artificial-intelligence/rsac-2026-rethinking-trust-in-agentic-ai-security/">advocates for gated agent architectures</a> built on <a href="https://www.nccgroup.com/research/analyzing-secure-ai-architectures/">trust segmentation principles</a> where AI systems inherit the trust level of the data they process. Untrusted input, restricted capabilities. Both Anthropic and Nvidia move in this direction. Neither fully arrives.</p><h2>The zero-trust architecture audit for AI agents</h2><p>The audit grid covers three vendor patterns across six security dimensions, five actions per row. It distills to five priorities:</p><ol><li><p><b>Audit every deployed agent for the monolithic pattern. </b>Flag any agent holding OAuth tokens in its execution environment. The <a href="https://cloudsecurityalliance.org/press-releases/2026/03/24/more-than-two-thirds-of-organizations-cannot-clearly-distinguish-ai-agent-from-human-actions">CSA data</a> shows 43% use shared service accounts. Those are the first targets.</p></li><li><p><b>Require credential isolation in agent deployment RFPs. </b>Specify whether the vendor removes credentials structurally or gates them through policy. Both reduce risk. They reduce it by different amounts with different failure modes.</p></li><li><p><b>Test session recovery before production. </b>Kill a sandbox mid-task. Verify state survives. If it does not, long-horizon work carries a data-loss risk that compounds with task duration.</p></li><li><p><b>Staff for the observability model. </b>Anthropic&#x27;s console tracing integrates with existing observability workflows. NemoClaw&#x27;s TUI requires an operator-in-the-loop. The staffing math is different.</p></li><li><p><b>Track indirect prompt injection roadmaps. </b>Neither architecture fully resolves this vector. Anthropic limits the blast radius of a successful injection. NemoClaw catches malicious proposed actions but not malicious returned data. Require vendor roadmap commitments on this specific gap.</p></li></ol><p>Zero trust for AI agents stopped being a research topic the moment two architectures shipped. The monolithic default is a liability. The 65-point gap between deployment velocity and security approval is where the next class of breaches will start.</p>]]></description>
            <author>louiswcolumbus@gmail.com (Louis Columbus)</author>
            <category>Security</category>
            <enclosure url="https://images.ctfassets.net/jdtwqhzvc2n1/35u1BcqPcGPsUcDOjxBtIh/fb7d22c10100687068f227c644d2e297/zero_trust_hero.png?w=300&amp;q=30" length="0" type="image/png"/>
        </item>
    </channel>
</rss>