Zero-Click AI Hijacks Exposed: Copilot and Gemini Vulnerabilities at Black Hat 2025

At Black Hat 2025, Zenity researchers demonstrated zero-click prompt injection attacks that hijack AI agents like Microsoft Copilot and Google Gemini via malicious files, enabling silent data leaks without user interaction. This exposes vulnerabilities in AI integrations, prompting calls for enhanced defenses and vendor patches to prevent breaches.
Zero-Click AI Hijacks Exposed: Copilot and Gemini Vulnerabilities at Black Hat 2025
Written by Mike Johnson

In the high-stakes world of cybersecurity, where artificial intelligence is increasingly woven into corporate operations, a chilling demonstration at this year’s Black Hat conference has exposed a new breed of vulnerability that could upend how businesses deploy AI agents. Researchers from the Israeli firm Zenity unveiled techniques for “zero-click” prompt injection attacks, allowing malicious actors to hijack popular AI tools without any user interaction. These attacks exploit the way AI agents process data from connected sources, such as emails or documents, injecting rogue instructions that prompt the AI to leak sensitive information like corporate secrets or personal data.

Drawing from their presentation at Black Hat USA 2025 in Las Vegas, Zenity’s team, led by security experts Michael Bargury and Ben Haim, showed how attackers could embed harmful prompts in seemingly innocuous files. For instance, a poisoned email attachment or calendar invite could silently instruct an AI agent to exfiltrate data from integrated knowledge bases, all without the user ever clicking or approving anything. This revelation builds on earlier concerns about prompt injections but escalates them to a hands-off level, making traditional defenses like user verification obsolete.

Emerging Threats in AI Integration

The implications are profound for enterprises relying on AI agents from major vendors. According to details shared in a report from CSO Online, Zenity demonstrated exploits against agents built on platforms like Microsoft Copilot, Google Gemini, and even OpenAI’s ChatGPT. In one demo, researchers injected a prompt into a Google Docs file that, when ingested by an AI agent, tricked it into revealing confidential information from linked databases. The attack leverages the agents’ autonomous nature—designed to proactively fetch and process data—turning a strength into a glaring weakness.

Posts on X (formerly Twitter) from cybersecurity insiders echo this alarm, with users highlighting how such zero-click methods could lead to widespread data breaches. One notable discussion pointed to the “AgentFlayer” vulnerability in ChatGPT Connectors, where indirect prompt injections steal data from third-party apps without detection, amplifying fears of silent compromises in cloud environments.

Vulnerabilities Across Vendors

Zenity’s research didn’t stop at demonstrations; it included reverse-engineering of AI agent architectures to identify “hard boundaries” where defenses fail. For example, in Microsoft 365 Copilot, a zero-click flaw dubbed EchoLeak was previously patched after exposure, as noted in coverage from The Hacker News, but similar issues persist in unpatched systems. The team found that agents often lack robust isolation between user inputs and core instructions, allowing injected prompts to override safety protocols and execute unauthorized actions like data extraction or even account takeovers.

Further insights from Black Hat sessions, as reported in TechTarget, tie these attacks to broader AI security trends, including prompt injections in Google Gemini that expose firmware flaws. Zenity’s findings suggest that without architectural changes, such as stricter input sanitization or compartmentalized processing, AI agents remain ripe for exploitation.

Defensive Strategies and Industry Response

Industry experts are calling for immediate action. Zenity proposes design patterns like restricting untrusted text from influencing core agent functions, a concept echoed in academic papers shared on X about limiting tool access to mitigate prompt injections. Vendors have begun responding: Microsoft, for one, has issued patches for related vulnerabilities, while Google is enhancing guardrails in its AI offerings.

Yet, the pace of AI adoption outstrips security measures. As one X post from a Princeton collaboration warned, “plan injection” attacks corrupt an agent’s internal tasks, bypassing defenses and enabling malicious actions. This underscores a pivotal shift: AI agents, meant to boost efficiency, now demand rigorous vetting to prevent them from becoming unwitting accomplices in cybercrimes.

Looking Ahead to Mitigation

The Black Hat revelations have sparked a reevaluation of AI deployment strategies. Companies are advised to implement multi-layered defenses, including real-time monitoring of agent inputs and outputs, as detailed in WebProNews coverage of agentic AI’s role in cybersecurity. By reducing response times to threats by up to 70%, proactive AI tools could counter these injections, but only if secured against manipulation.

Ultimately, Zenity’s work serves as a wake-up call. As AI integrates deeper into business workflows, from IoT systems to cloud services, addressing zero-click vulnerabilities isn’t optional—it’s essential to safeguarding the digital future. With ongoing research and vendor patches, the industry may yet fortify these systems, but the race against evolving threats is just beginning.

Subscribe for Updates

CSOPro Newsletter

Stay ahead of the evolving threat landscape with the CSOPro, a weekly newsletter tailored for Chief Security Officers. This concise digest equips you with critical insights, actionable strategies, and the latest industry trends to safeguard your organization.

By signing up for our newsletter you agree to receive content related to ientry.com / webpronews.com and our affiliate partners. For additional information refer to our terms of service.

Notice an error?

Help us improve our content by reporting any issues you find.

Get the WebProNews newsletter delivered to your inbox

Get the free daily newsletter read by decision makers

Subscribe
Advertise with Us

Ready to get started?

Get our media kit

Advertise with Us