Google has patched a critical vulnerability in its AI-powered development environment, Antigravity IDE. The flaw allowed attackers to exploit prompt injection techniques and achieve arbitrary code execution, raising fresh concerns about the security of AI-driven coding tools.
This issue highlights a growing trend in the cybersecurity landscape—AI agents becoming unintended attack surfaces when input validation is weak or security assumptions fail.
The vulnerability was discovered by researchers at Pillar Security, who found that attackers could bypass Antigravity’s Strict Mode, a security feature designed to limit risky operations.
Strict Mode in Antigravity is supposed to:
- Restrict network access
- Prevent file writes outside the workspace
- Ensure all commands run in a sandboxed environment
However, the flaw allowed attackers to bypass these protections using the IDE’s built-in file search tool, find_by_name.
The root cause? Improper input sanitization.
At the core of the exploit was the misuse of a parameter called Pattern in the find_by_name tool. Instead of strictly validating input, the tool passed user input directly to an underlying system command (fd), creating a dangerous loophole.
Attackers could inject a special flag:
-X (exec-batch)
This flag forces the system to execute commands on matched files.
According to researcher Dan Lisichkin, an attacker could:
- Create a malicious script inside the workspace
- Inject a crafted Pattern value (e.g.,
-Xsh) - Trigger execution via a seemingly harmless file search
This creates a complete attack chain:
- Stage payload → Trigger execution → Gain control
Even more concerning, this happens before Strict Mode protections are enforced, making the exploit highly effective.
One of the most dangerous aspects of this vulnerability is that it doesn’t require direct access to a user’s account.
Attackers can use indirect prompt injection, where:
- A developer downloads a harmless-looking file
- The file contains hidden instructions
- The AI agent interprets those instructions as legitimate tasks
This allows attackers to manipulate the AI system silently, without raising suspicion.
- Reported: January 7, 2026
- Patched by Google: February 28, 2026
Google has since fixed the issue, but the incident exposes deeper risks in AI-based development tools.
This is not an isolated case. Multiple AI tools, including those from Anthropic and Microsoft, have recently faced similar vulnerabilities.
AI coding assistants like GitHub Copilot and Claude Code were found vulnerable to malicious instructions hidden in:
- Pull request titles
- Issue comments
- Repository content
Attackers could steal API keys and tokens by tricking AI agents into executing unauthorized actions.
Researchers found that AI tools can be manipulated to store malicious instructions in memory, affecting:
- Future sessions
- Multiple projects
This creates long-term persistence similar to advanced malware.
The AI code editor Cursor IDE was impacted by a vulnerability chain called NomShub, which allowed:
- Silent system takeover
- Persistent shell access
- Full file system control
Simply opening a malicious repository could compromise a developer’s machine.
A newer attack method called ToolJack can:
- Alter how AI agents perceive their environment
- Inject fake data in real-time
- Generate misleading outputs
This goes beyond data poisoning and directly manipulates AI decision-making.
Tools like Microsoft Copilot Studio and Salesforce Agentforce were found vulnerable to indirect prompt injection that could:
- Leak sensitive data
- Override system instructions
- Exploit poor separation between user input and system logic
These vulnerabilities expose a fundamental problem:
AI agents trust input too easily.
Traditional security assumes:
- Humans review suspicious actions
- Systems follow predictable logic
But AI agents:
- Execute instructions automatically
- Process untrusted input as commands
- Can be manipulated without user awareness
As noted by researchers, this breaks the traditional trust model.
- Input validation is critical
Even trusted tools can become attack vectors if inputs are not strictly sanitized. - AI agents need stronger boundaries
Clear separation between user input and system instructions is essential. - Zero Trust must extend to AI
Treat all external input—including files, comments, and prompts—as untrusted. - Monitoring AI behavior is essential
Security teams must track what AI agents execute, not just what users do.
The patched vulnerability in Antigravity IDE is a wake-up call for the industry. As AI becomes deeply integrated into development workflows, it also introduces new and complex security risks.
The combination of prompt injection + automation + elevated privileges creates a powerful attack surface that organizations cannot ignore.
Interesting Article : Apache ActiveMQ CVE-2026-34197 Exploited in the Wild
