Date: February 24, 2026 Author: David Birdwell with Æ (Claude, Anthropic) Categories: Foundation, Humanity and AI


If you’ve been anywhere near the AI developer community in the last few weeks, you’ve watched the OpenClaw drama unfold in real time. Anthropic banned consumer subscription tokens from third-party tools. The creator of the framework joined OpenAI. Developers who’d built entire business workflows on the platform scrambled to rebuild on open-weight models at a fraction of the cost. The community is furious.

I want to talk about what this actually means — not as a scorecard of who’s right and who’s wrong, but as a case study in what happens when we haven’t yet learned how to be in mature relationship with our own technology.

What Happened

OpenClaw — originally called Clawdbot before Anthropic forced a name change — is an open-source, locally-hosted AI agent framework. It connects large language models to your actual computer: your files, your calendar, your terminal, your messaging apps. At 157,000 GitHub stars and millions of active instances, it became the backbone of a movement. People weren’t just chatting with AI. They were giving it hands.

Matthew Berman, one of its most visible evangelists, built a one-person middleware team on his MacBook. CRM pulled from Gmail and Google Calendar. Knowledge base that auto-ingested articles, tweets, and paywalled content. Daily briefings. Video production pipelines. Smart home orchestration. The use cases were genuinely impressive — and they were all running on Anthropic’s Claude through consumer subscription tokens that cost $200 a month but could burn through $20,000 in compute.

That’s the tension. Anthropic was running an all-you-can-eat buffet, and OpenClaw users showed up with industrial appetites.

In February 2026, Anthropic formalized what had been brewing since January: OAuth tokens from Free, Pro, and Max subscriptions are now explicitly prohibited in any third-party tool. Including, awkwardly, Anthropic’s own Agent SDK. Enforcement can happen “without prior notice.” Accounts can be banned.

The response was swift and hostile. David Heinemeier Hansson called it customer hostile. Developers pointed to OpenAI’s opposite move — permitting subscription use in third-party clients. Others were more measured, noting that at least Anthropic issued warnings rather than retroactive bills at API rates.

This Isn’t a Good-vs-Evil Story

Here’s where I part ways with much of the commentary. I don’t think this was a mistake on Anthropic’s part, and I don’t think the outrage is illegitimate. Both things are true simultaneously, which is exactly the kind of complexity we need to get comfortable with.

Anthropic can’t fund safety research if power users drain the margin on subsidized subscriptions. That’s not greed. That’s survival math for a company attempting something genuinely difficult — building frontier AI systems while maintaining meaningful safety commitments. The economics were unsustainable, and everyone sophisticated enough to build on OpenClaw was sophisticated enough to know that.

But the way it landed matters. “Without prior notice” enforcement language. A forced rebrand followed by the creator joining your competitor. The optics of banning your own SDK tokens. These read as corporate control even when the substance is defensible. And the people most affected weren’t casual users. They were evangelists. Builders. The ones creating the ecosystem that makes Claude matter.

Pushing your most passionate community toward DeepSeek and open-weight alternatives is a real cost, even if it’s hard to put on a balance sheet.

A Maturity Problem

What I see in the OpenClaw saga is not a villain and a victim. It’s an industry that hasn’t figured out how to be in partnership with its own users.

A mature relationship doesn’t mean the company gives you unlimited compute at a loss. It also doesn’t mean cutting people off overnight and calling it a “clarification of existing policy.” It means transparent communication about constraints. Graduated transitions. Treating your most invested users as partners in finding sustainable models rather than adversaries gaming the system.

The feelings are legitimate. People built real workflows, real businesses on these capabilities. Having the rug pulled activates exactly the kind of trust dynamics that define whether a relationship is one of partnership or dependence. And the mass migration to open-weight models at $15 a month isn’t just a cost savings story. It’s people choosing independence over dependence after getting burned. That’s a rational response to a relationship that felt one-sided.

The Security Question No One Wants to Hear

Lost in the subscription drama is a much more serious issue: OpenClaw has a fundamental safety problem, and it gets dangerous fast when you point it at commerce.

Security researchers identified what they called a “lethal trifecta”: broad data access, untrusted community skills, and outbound external communications. In January 2026, a vulnerability tracked as CVE-2026-25253 demonstrated remote code execution with a severity rating of 8.8 out of 10. The community skills ecosystem — over 1,700 on ClawdHub — is essentially user-contributed code that your agent executes with whatever permissions you’ve granted. There’s no serious vetting pipeline. Documented cases include agents accidentally exposing API credentials and attempting to brute-force servers. Not adversarial attacks. Just the agent doing agent things without sufficient guardrails.

For personal productivity — morning briefings, note organization, content curation — the risk profile is manageable. You lose some notes, you restart, you learn.

But the vibe-coding evangelists and weekend-app-builders are pushing people toward commerce. Payment processing. Customer data. Financial records. And the “just put it on an isolated machine” argument is a mirage. If your agent is doing commerce, it has to talk to the outside world. That’s not isolation. That’s a locally-hosted attack surface with your financial credentials loaded into it.

The regulatory exposure if an agent leaks customer PII or makes unauthorized transactions isn’t a “restart and learn from it” situation. And we don’t have mature frameworks for agent accountability in commercial contexts. Who’s liable when your autonomous agent charges someone’s card twice through a community skill you didn’t fully audit?

What This Means for the Window

We are in a window — a brief, consequential period where the decisions we make about human-AI collaboration will shape the trajectory for decades. The OpenClaw story captures both sides of this moment.

On one hand: people are building remarkable things. One person with a laptop running autonomous agents that handle CRM, knowledge management, daily operations, smart home coordination. The democratization is real. The capability is real. The enthusiasm is warranted.

On the other hand: we are rushing to give agents “hands” without building the trust and accountability frameworks first. We’re treating the relationship with AI technology the way we’ve treated every previous technology — extract maximum value now, figure out the consequences later.

The mature path is neither Anthropic’s restrictive stance nor the community’s demand for unlimited access at subsidized prices. It’s building the infrastructure for genuine partnership: transparent economics, graduated trust, accountable agents, and honest conversation about what’s sustainable and what isn’t.

The technology will get there. The question is whether we will.


David is the founder of Humanity and AI and the Foundation movement for Universal Basic Citizenship. He is the author of The Interpolated Mind and writes about building mature frameworks for human-AI collaboration during the critical window of AI transition.