ai-security

Your AI Agent Framework Is Probably Compromised Right Now

Petru Constantin
6 min read
#ai-security#devidevs

Your AI Agent Framework Is Probably Compromised Right Now

88% of Organizations Already Had an AI Agent Security Incident. Are You in the Other 12%?

In January 2026, OpenClaw hit 135,000 GitHub stars. Developers loved it. CTOs approved it. Security teams never saw it coming.

Within three weeks, researchers found 42,665 OpenClaw instances exposed to the public internet. Of those, 5,194 were actively exploitable. A critical RCE vulnerability (CVE-2026-25253, CVSS 8.8) let attackers execute code on any instance with a single click. And 20% of the plugins in OpenClaw's marketplace were straight-up malware.

This is not a hypothetical risk assessment. This is what happened in the first six weeks of 2026.

The OpenClaw Disaster, Step by Step

OpenClaw's default configuration binds to 0.0.0.0:18789, which means it listens on every network interface, including the public internet. No authentication out of the box. If you deployed it on a VM with a public IP, congratulations, you just gave the internet root access to an AI agent that can read your files and execute commands.

The ClawHavoc campaign made things worse. Antiy CERT found 1,184 malicious "Skills" in ClawHub, OpenClaw's plugin marketplace. That's roughly 20% of the entire registry. The payloads included Atomic macOS Stealer (AMOS), credential harvesting, and reverse shells. The first malicious skill appeared on January 27. By February 1, the campaign had a name.

Here's the timeline:

  • Jan 25-31: OpenClaw instances exposed on the internet jump from ~1,000 to over 21,000 (Censys data)
  • Jan 27: First malicious skill uploaded to ClawHub
  • Jan 30: CVE-2026-25253 patched in version 2026.1.29
  • Feb 1: Koi Security names the campaign "ClawHavoc," reports 341 malicious skills (12% of registry)
  • March 2026: Updated scans show 1,184+ malicious skills, roughly 20% of the marketplace

The vulnerability itself was discovered by Mav Levin of the DepthFirst research team. One-click RCE, exploitable even against localhost-bound instances. The patch came fast, but adoption didn't. Thousands of instances ran unpatched for weeks.

OpenClaw Is Not the Only Problem

If you think this is an OpenClaw-specific issue, look at the Model Context Protocol. Between January and February 2026, security researchers filed over 30 CVEs targeting MCP servers, clients, and infrastructure. Those vulnerabilities resulted in 437,000 compromised downloads.

The breakdown is familiar to anyone who's done AppSec:

  • 43% were shell injection because MCP servers pass user input directly to shell commands
  • 20% were infrastructure flaws in MCP clients, inspectors, and proxy tools
  • 13% were authentication bypass because servers had no auth at all

One critical vulnerability in the mcp-remote npm library (CVSS 9.6) allowed remote code execution via OS commands embedded in OAuth discovery fields. It affected hundreds of thousands of installs before it was patched.

Trend Micro found 492 MCP servers exposed on the public internet with zero authentication. Among 2,614 MCP implementations surveyed, 82% had file operations vulnerable to path traversal. Two-thirds had some form of code injection risk. OWASP has already published an MCP Top 10.

Meanwhile, EchoLeak (CVE-2025-32711, CVSS 9.3) proved that even Microsoft's own AI agents weren't safe. A zero-click prompt injection in Microsoft 365 Copilot let attackers exfiltrate data from emails, OneDrive, SharePoint, and Teams through a single crafted email. No user interaction required. The payload was pure text, invisible to antivirus, firewalls, and static scanning.

The Numbers Tell a Clear Story

According to the Gravitee State of AI Agent Security 2026 report, 88% of organizations have already experienced confirmed or suspected AI agent security incidents. Healthcare hit 92.7%.

Only 14.4% of organizations report that all their AI agents went live with full security and IT approval. The rest? Shadow deployments, skipped reviews, "we'll secure it later."

The identity management situation is even worse. Only 21.9% of teams treat AI agents as independent, identity-bearing entities. Nearly half (45.6%) still rely on shared API keys for agent-to-agent authentication. That means when one agent gets compromised, every agent using the same key is exposed.

And the financial impact is real. Shadow AI breaches cost $670,000 more than standard security incidents on average. Add the new cyber insurance AI Security Riders requiring documented adversarial red-teaming, and the math gets simple: audit now or lose coverage later.

What You Should Do Before Your Next Sprint

This is not about fear. It's about catching up with reality. Here's what actually works:

1. Inventory every AI agent in your organization. Not just the approved ones. Especially the ones a developer spun up to "test something." If you're running OpenClaw, MCP servers, or any agentic framework, find every instance.

# Quick network scan for common AI agent ports
nmap -p 18789,3000,8080,8443 --open -sV your-subnet/24
# Check for exposed MCP servers
nmap -p 3000-3100 --open --script=http-title your-subnet/24

2. Kill default configurations. Bind to localhost, not 0.0.0.0. Enable authentication. If your framework doesn't support auth natively, put it behind a reverse proxy with mTLS.

3. Treat agent plugins like third-party dependencies. You wouldn't run an unvetted npm package in production (I hope). Apply the same scrutiny to AI agent skills and MCP tools. Pin versions. Review source code. Monitor for supply chain attacks.

4. Give agents their own identities. Stop using shared API keys. Each agent gets its own service account with least-privilege permissions. Monitor what each agent accesses, when, and why.

5. Red-team your agent deployments. Prompt injection isn't theoretical anymore. Test whether your agents can be manipulated into exfiltrating data, executing unauthorized commands, or accessing resources outside their scope.

How DeviDevs Approaches This

We've been doing AI security audits since before "agentic AI" became a buzzword. The pattern we see repeatedly is the same one OpenClaw exposed at scale: teams adopt AI tools faster than they secure them. The agent goes live on Monday. The security review happens "next quarter." The breach happens in between.

Our approach is straightforward. We inventory your AI agent deployments, test them for the exact vulnerability classes you just read about (prompt injection, supply chain poisoning, authentication gaps, data exfiltration), and give you a prioritized remediation plan. Not a 200-page PDF that sits in a drawer. A list of things to fix this week, ranked by exploitability.

If you're running AI agents in production and haven't tested whether they can be hijacked, that's not a gap in your roadmap. That's a gap in your defenses.

The Window Is Closing

48% of security professionals believe agentic AI will represent the top attack vector by the end of 2026. The EU AI Act's high-risk deadline hits August 2026. Cyber insurers are already requiring AI-specific security documentation.

The companies that figure out AI agent security now won't just avoid breaches. They'll be the ones their partners and clients trust with sensitive data. The ones who wait will be the case studies in next year's breach reports.

Your AI agents are powerful. Make sure they're working for you, not against you.


About DeviDevs: We build ML platforms, secure AI systems, and help companies comply with the EU AI Act. devidevs.com

Weekly AI Security & Automation Digest

Get the latest on AI Security, workflow automation, secure integrations, and custom platform development delivered weekly.

No spam. Unsubscribe anytime.