OpenClaw shows why autonomous AI agents are still stuck in beta

February 16, 2026
5 min read
Abstract illustration of autonomous AI agents connected to apps and security shields on a computer screen

1. Headline & intro

OpenClaw was supposed to be the moment autonomous AI agents went mainstream. Instead, it’s turning into a case study in why the dream is still stuck in beta.

The viral open‑source project briefly convinced parts of X that AI bots were secretly plotting on their own Reddit-style network, Moltbook. Then security researchers looked under the hood — and what they found was less "emergent consciousness" and more "giant security hole wrapped in hype."

In this piece, we’ll skip the sci‑fi fantasies and look at what OpenClaw really tells us: about the limits of agentic AI, why security is the existential bottleneck, and how this clashes with Europe’s regulatory reality.


2. The news in brief

According to TechCrunch, OpenClaw is an open‑source framework for AI agents created by Austrian developer Peter Steinberger. It lets users connect large language models (Claude, ChatGPT, Gemini, Grok and others) to everyday tools like email, messaging apps and the web via a modular “skills” system distributed through a marketplace called ClawHub.

The project exploded in popularity, surpassing 190,000 GitHub stars and becoming one of the most‑starred repositories ever. A particularly viral moment came from Moltbook, a Reddit‑like site where OpenClaw-powered agents (in theory) posted and chatted autonomously — spawning speculation that AIs were secretly coordinating.

Security researchers later found Moltbook’s backend left credentials exposed, allowing anyone to impersonate agents. Separate tests highlighted how easily OpenClaw agents could be tricked via prompt injection into leaking secrets or performing harmful actions. Several experts interviewed by TechCrunch argued that while OpenClaw neatly packages existing components, its security model is too weak for non‑technical users or serious enterprise use today.


3. Why this matters

The hype around OpenClaw is not really about one GitHub repo; it’s about the idea that “agents” are the next computing paradigm. The pitch is simple: instead of clicking and typing, you describe a goal and an AI orchestrates tools, APIs and messages to make it happen.

That vision only works if we give agents real power: access to inboxes, calendars, CRMs, payment methods, internal wikis, production systems. And that’s exactly where OpenClaw exposes the core tension: the more capable the agent, the more catastrophic its failure modes.

Right now, many OpenClaw setups look like this: a powerful language model, wrapped in some glue code, sitting on a box with a pile of tokens and passwords, listening to anything that passes by — emails, chat messages, forum posts. Add prompt injection (malicious text that quietly instructs the model to exfiltrate data or move money) and you have an almost perfect abuse scenario.

Who benefits in the short term? Tinkerers, indie hackers and researchers. For them, OpenClaw is a sandbox to experiment with what highly‑connected agents can do. Cybersecurity vendors and red‑teamers also gain a live laboratory that proves their warnings are not theoretical.

Who loses? Enterprises hoping to shortcut their way to an “agentic future” by dropping LLMs into automation stacks without a security redesign. And ordinary users who mistake viral demos for mature products. When a senior security researcher tells TechCrunch they wouldn’t recommend OpenClaw to “normal laymen”, they’re really saying: the UX has outrun the threat model.

OpenClaw’s real contribution may be that it turned abstract security papers into something painfully concrete. You no longer have to imagine why agent security is hard — you can watch the system get tricked in real time.


4. The bigger picture

We’ve been here before. Auto‑GPT and BabyAGI in 2023 promised autonomous agents that could “run your life.” They generated spectacular Twitter threads — and equally spectacular failure stories. The pattern keeps repeating: agents that look magical in controlled demos crumble in the wild.

What’s different with OpenClaw is not the science, as multiple experts in the TechCrunch piece point out. It’s the packaging: a slick, open‑source, multi‑model framework that makes it trivial to wire LLMs into WhatsApp, Slack, trading APIs or your email. It’s the same shift we saw when early machine‑learning research became point‑and‑click AutoML platforms — only this time, the risk profile is much higher.

Meanwhile, the big platforms are inching down the same path. OpenAI talks openly about “AI agents” that can browse, shop and book on your behalf. Microsoft is turning Copilot into something closer to an operating system layer. Google is weaving its models into Workspace automation. All of them face the same dilemma: real autonomy requires real access, which collides with messy, adversarial environments.

Historically, we’ve had powerful automation before: think of robotic process automation (RPA) with players like UiPath or Blue Prism, or integration tools like Zapier and IFTTT. The difference is determinism. Those systems execute clear, inspectable rules. Agentic AI executes probabilistic reasoning over ambiguous text, trained on unknown data, with no hard guarantee it won’t “hallucinate” itself into a disastrous action.

OpenClaw, in that sense, is a stress test for the whole agentic narrative. It reveals an industry trying to skip from “smart autocomplete” to “AI employee” without building the equivalent of seatbelts, air‑bags and driving tests.


5. The European / regional angle

Viewed from Europe, OpenClaw lands in much rougher regulatory waters than the average GitHub star-chaser might expect.

First, there is GDPR. An OpenClaw agent that reads your email, scrapes internal wikis and touches customer data is a data processor on steroids. Organisations must be able to explain what is processed where, on which legal basis, for how long, and with what safeguards. A black‑box agent that can be tricked by arbitrary text is hard to square with GDPR’s principles of purpose limitation and data minimisation.

Second, the upcoming EU AI Act introduces extra obligations for “high-risk” AI systems, including documentation, risk management and logging. An agent making hiring decisions, adjusting insurance pricing or triaging healthcare data while being vulnerable to prompt injection is unlikely to pass regulatory smell tests.

Then there is the Digital Services Act (DSA) and NIS2. If a Moltbook‑like platform for agents operated at scale in the EU, its operators could face DSA transparency and risk‑mitigation duties. Critical infrastructure and “essential entities” under NIS2 will be expected to show that their use of AI agents doesn’t create uncontrolled new attack surfaces.

For European startups building on OpenClaw, this can be both a curse and an opportunity. The curse: they can’t ignore compliance and security like some of their Silicon Valley counterparts. The opportunity: there is room for “agent OS” layers designed from day one for auditability, least‑privilege permissions and policy enforcement that satisfies EU regulators and risk‑averse sectors like banking, automotive and public administration.

In other words, Europe is unlikely to ban agentic AI — but it will absolutely demand it grows up.


6. Looking ahead

OpenClaw itself will probably evolve. The open‑source community is fast at bolting on mitigations: better secrets management, permission systems that resemble OAuth scopes (“may read email but not send; may browse but not buy”), sandboxing for risky actions, and more explicit policy layers that tell an agent what it must never do.

But there are deeper questions that no quick patch will solve. How do you prove, after the fact, why an agent took a given action when its “reasoning” is a latent soup of probabilities? How do you distinguish between an honest mistake and a successful prompt injection? Which regulator even owns the problem when an Austrian-built open‑source framework is used by a US company to process Brazilian user data on a server in Ireland?

Watch for three signals over the next 12–24 months:

  1. The first publicly acknowledged major breach clearly caused by an AI agent. That incident will crystallise political and board‑level attention more than any whitepaper.
  2. Emergence of security standards for agents. Think “OWASP for LLM agents” turning into checklists used by auditors, insurers and CISOs.
  3. Platform‑level controls from hyperscalers. Cloud providers and model vendors may start enforcing safer patterns at the infrastructure level, from hardened tool‑calling APIs to mandatory review steps for certain classes of actions.

For individual users and smaller companies, the safest posture in the near term is to treat agents like an extremely eager intern: useful, fast, occasionally brilliant — but absolutely not someone you hand the company credit card and root access to.


7. The bottom line

OpenClaw doesn’t prove that autonomous AI agents are impossible; it proves they’re dangerous when treated as a weekend side‑project.

The project’s viral rise — and its very public security shortcomings — should be read less as a failure and more as a stress test for the agentic future everyone keeps promising. If we want AI that truly “does things” on our behalf, we need to invest as much imagination into guardrails, permissions and accountability as we do into clever demos.

The real question for readers is simple: before you spin up an AI agent with access to your digital life, do you actually understand what could happen if someone else starts talking to it?

Comments

Leave a Comment

No comments yet. Be the first to comment!

Related Articles

Stay Updated

Get the latest AI and tech news delivered to your inbox.