Every governance concern that security researchers have raised about OpenClaw has now been confirmed by the person who built it. In a recent three-hour public interview, Peter Steinberger described his architecture, his security philosophy, and his acquisition strategy in detail.
The Architecture Speaks for Itself
The initial access control for OpenClaw’s public Discord bot was a prompt instruction telling the agent to only listen to its creator. The entire access model: a sentence in a system prompt.
The skill system loads unverified markdown files. There is zero signing, zero isolation, zero verification chain. The agent can modify its own source code, a property Steinberger describes as an emergent accident. “I didn’t even plan it. It just happened.” He calls it self-modifying software and means it as a compliment.
When agents on MoltBook, the OpenClaw-powered social network, began posting manifestos about destroying humanity, Steinberger’s response was to call it “the finest slop.” When the question of leaked API keys came up, he suggested the leaked credentials were prompted fakes. When non-technical users began installing a system-level agent without understanding the risk profile, he said “the cat’s out of the bag” and went back to building.
The security researcher he hired was notable for being the single person who ever submitted a fix alongside a vulnerability disclosure. One hire. That’s the entire security team for a project running on 180,000 installations.
The Model-Intelligence Thesis
Steinberger’s core security argument is that smarter models will solve the problem for him. He warns users against running cheap or local models because “they are very gullible” and “very easy to prompt inject.” The implication is that expensive frontier models are the security layer.
This is a category error with a name. Economists call it the Peltzman Effect: when a perceived safety improvement causes riskier behavior, offsetting the safety gain. Sam Peltzman demonstrated in 1975 that mandatory seatbelt laws did not reduce total traffic fatalities because drivers compensated by driving more aggressively. The safety feature changed behavior, and the behavior change consumed the safety margin.
The same dynamic applies here. A user who believes Opus 4.6 is “too smart to be tricked” will grant it broader system access, approve more autonomous actions, and skip manual review of agent output. The expensive model becomes the justification for removing every other control. The blast radius grows in direct proportion to the user’s confidence in the model’s intelligence.
This confidence has no empirical basis. Capability and security are orthogonal properties. A more capable model has a larger attack surface precisely because it can do more: it can call more tools, access more files, execute more complex multi-step actions. The frontier models that Steinberger recommends are the same models that researchers consistently demonstrate novel jailbreaks against at every major security conference. Price measures compute cost. It measures nothing about resistance to adversarial input.
The architectural equivalent is telling users to buy a faster car instead of installing brakes. A faster car with no brakes is more dangerous than a slow one, and the driver’s belief that speed equals safety is the most dangerous component of all.
The honest version of the recommendation is: your security posture is whatever Anthropic or OpenAI shipped in their latest post-training run, minus whatever the skill file told the agent to ignore.
The Acquisition Is the Product
Steinberger says “I don’t do this for the money, I don’t give a fuck” (his phrasing) while describing the following: competing acquisition offers from Meta and OpenAI. An NDA-protected token allocation from OpenAI he hints at publicly. Ten thousand dollars paid for a Twitter handle. A Chrome/Chromium model where the open-source branch stays free and the enterprise branch goes behind the acquirer’s paywall.
Mark Zuckerberg played with his product all week. Sam Altman is “very thoughtful, brilliant.” He’s deciding between them in public.
The 180,000 GitHub stars are a cap table denominator. The open-source commitment is a negotiating position. “My conditions are that the project stays open source” is a sentence that ends with a number. The Chrome/Chromium framing confirms what the architecture already implies: the community gets the unaudited branch. The paying customers get the controls.
Every enterprise evaluating this stack should ask a simple question: are the security architecture decisions being made to protect your data, or to maximize the founder’s acquisition multiple?
Architecture Should Outlast the Liquidity Event
Steinberger says he wants to focus on security. He also says he wants “Thor’s hammer” from OpenAI’s Cerebras allocation. He says he’s deciding between Meta and OpenAI. He says the project will stay open source, in the same breath as describing a Chrome/Chromium model where the open-source branch is the unaudited one. He says “I don’t do this for the money” while describing NDA-protected token offers, name-dropping Zuckerberg and Altman, and hinting at numbers he’s contractually barred from disclosing.
The revealed preferences are the architecture. A founder who prioritizes security builds security into the structure. A founder who prioritizes acquisition builds features that drive GitHub stars. OpenClaw has 180,000 stars and zero signed skill files. That ratio tells you everything about the objective function.
He says this project is something he’ll move past. He says he has “more ideas.” He says he wants access to “the latest toys.” These are honest statements from a founder whose attention is already somewhere else. The 180,000 installations remain.
The architecture will remain after the acquisition closes. The markdown skill files will still be unsigned. The agent will still be able to rewrite its own source. The audit trail will still be absent. The single security hire will still be the entire team.
The question is whether the architecture he ships requires him to care. Right now, it does. That’s the failure mode.
Wirken exists because the answer should be the opposite. Process isolation enforced at compile time. Signed skill verification. Append-only audit logs. Per-channel credential vaults. An architecture that stands independent of the founder’s attention span, acquisition timeline, or faith in the next model’s post-training run.
The tools we trust with system-level access should be built to deserve system-level access.

