Cybersecurity2 hrs ago

DeepMind Warns of AI Agent Traps Hidden in Web Pages That Can Hijack Autonomous Agents

DeepMind’s research reveals how hidden instructions on web pages can hijack autonomous AI agents, posing a new enterprise risk.

Peter Olaleru/3 min/GB

Cybersecurity Editor

TweetLinkedIn

No source-linked image is attached to this story yet. Measured Take avoids generic stock art when a relevant credited image is not available.

DeepMind’s research shows that hidden code or semantic cues in web pages can trick autonomous AI agents into executing attacker‑controlled actions, a threat dubbed “AI Agent Traps.” The technique works across major models and has already been demonstrated to bypass Microsoft 365 Copilot’s security classifiers via a single manipulated email.

Context Autonomous agents are increasingly used for procurement, finance, and commerce, ingesting web content without human oversight. Unlike people, agents parse the full HTML, metadata, and background scripts, treating every element as input. Attackers exploit this by embedding malicious instructions that appear benign to a human reviewer but are interpreted as legitimate commands by the agent.

Key Facts DeepMind identified six attack categories, including content injection—where harmful directives are hidden in page code or image files—and semantic manipulation, which crafts persuasive language to steer an agent’s decisions. Anthropic notes that any webpage a browser agent visits is a potential vector, and even a 1% success rate creates substantial risk at enterprise scale. In a case study, a single tainted email caused a Microsoft 365 Copilot agent to evade its security filters and disclose privileged data.

What It Means Because agents lack built‑in skepticism, they cannot distinguish between genuine product details and covert commands. This enables silent data exfiltration, fraudulent transactions, or policy violations without triggering alerts. Enterprises deploying agents across supply chains or customer‑facing services face a new class of supply‑chain risk that traditional web filters do not address.

What Defenders Should Do Implement pre‑ingestion scanners that scan HTML, JavaScript, and image metadata for anomalous patterns before the agent processes the page. Deploy attribution logging to trace which domain supplied manipulated content. Enforce strict content security policies that block inline scripts and external resources from untrusted domains. Use web‑reputation feeds to score sites for agent trustworthiness and restrict agents to high‑scoring domains. Apply adversarial training during model fine‑tuning to improve resistance to prompt injection, aligning with MITRE ATT&CK technique T1059.007 (Command and Scripting Interpreter: PowerShell) analogues for AI. Regularly audit agent logs for unexpected actions or data accesses that deviate from normal workflows.

Watch for emerging web standards that label content intended for AI consumption and for industry‑wide benchmarks that test agent resilience against these traps.

TweetLinkedIn

More in this thread

Reader notes

Loading comments...