OpenClaw security checklist for non-engineers: 12 checks before you connect your accounts
If you only remember one thing, remember this: OpenClaw is not “just another AI app.” It can message people, touch files, call tools, run code, and act with the credentials you hand it. That means the right question is not “Is it cool?” The right question is “What is the blast radius if it does something dumb or gets manipulated?”
My opinion up front: most OpenClaw security advice online is either too technical or too theatrical. The official docs are actually pretty honest. They do not pretend prompt injection is solved, they explicitly warn that blast radius matters, and they tell you to be deliberate about who can talk to the bot, what it can touch, and where it is allowed to act.
That is the frame for this article. Not panic. Not cope. Just a practical operator guide for people who want to experiment without doing something reckless. If you are also working on your agent’s personality layer, see my companion piece on OpenClaw SOUL.md examples and personality templates. And if you are still deciding where OpenClaw fits in your stack, read my OpenClaw vs n8n review.
Short answer: is OpenClaw safe to use?
It can be run more safely, but it is not something I would casually connect to my primary email, GitHub, or work accounts on day one. The official docs make clear that OpenClaw can execute commands, read and write files, access networks, and send messages if you give it those tools. Microsoft’s security guidance goes even further and says it should be treated like untrusted code execution with persistent credentials, not something you casually run on a standard workstation.
The good news is that you do not need enterprise-grade security engineering to make smarter choices. You do need discipline. Most of the risk comes from bad defaults, oversized permissions, and people treating “local-first” as if it automatically means “safe.”
What the scary terms mean in plain English
OpenClaw security discussions can get jargon-heavy fast, so here is the plain-English version.
Prompt injection
An attacker hides instructions in content your agent reads and tricks it into acting against your interests.
Blast radius
If the agent goes wrong, how much can it damage, leak, message, or modify?
Least privilege
Give the assistant the smallest access that still lets it do the job.
Sandboxing
Put risky actions in a separate room so a mistake does not spill onto your whole machine.
For non-engineers, the easiest way to think about it is this: OpenClaw is powerful because it can act. That same power is why the security boundary matters more than it does with a normal chatbot tab in your browser.
The 12-point OpenClaw security checklist
This is the core of the article. If you are evaluating OpenClaw and want a practical “before I connect this thing to my accounts” checklist, start here.
Tick each item as you complete it. Your progress is tracked in this session.
1. Run it in a dedicated environment
Use a dedicated VM or separate machine, not the laptop that holds your real work, browser sessions, downloads, and personal files.
2. Use dedicated accounts and tokens
Create separate email, calendar, GitHub, and API credentials for the agent where possible. Do not start with your primary accounts.
3. Keep the gateway local, not public
Do not expose the gateway unauthenticated on the open internet. Default to local-only access or a controlled private path.
4. Set strong auth, not a lazy token
Use a long random token or password. Do not assume the software will protect you from weak choices.
5. Start with one low-risk channel
If you are experimenting, begin with one controlled channel such as a private Telegram thread or a narrow test setup, not five apps at once.
6. Use pairing and allow-lists
Require pairing for direct messages and restrict who can talk to the bot. Unknown senders should not get a free shot at your agent.
7. Enable sandboxing before you need it
Do not wait until after an incident to isolate risky tool execution. Enable sandboxing and keep host access narrow.
8. Turn off high-risk tools you do not need
If an agent does not need shell execution, broad file access, or browser control, disable them. Start small and widen later.
9. Treat skills like code, not like prompts
Do not casually install community skills just because they sound useful. Vet them, re-check them on updates, and distrust surprise behavior.
10. Protect secrets like secrets
Store API keys and tokens carefully, use environment or secret-reference patterns where possible, and rotate anything you connected early while learning.
11. Audit the setup and watch for drift
Run the built-in security audit, review saved state, and keep an eye out for new permissions, strange tool use, or behavior changes after updates.
12. Have a rebuild plan before you connect anything important
Know exactly how you would shut it down, revoke tokens, rebuild the environment, and start clean if something feels off.
If you want the shortest possible checklist summary: isolate the machine, isolate the accounts, isolate the tools, isolate the network path, and assume you may need to rebuild.
Three realistic ways people accidentally make OpenClaw dangerous
The easiest way to understand the security risk is not through theory. It is through realistic operator mistakes.
Scenario 1: “I just connected my real inbox”
You connect your main Gmail and calendar because you want a daily brief. Now the agent can read sensitive messages, draft replies, or expose account data if it is manipulated.
Common beginner mistakeScenario 2: “I opened the gateway so I can access it remotely”
You expose the service publicly for convenience. Attackers do not need to be clever if your front door is open and your token is weak.
Network exposure mistakeScenario 3: “I installed a useful-looking skill”
You treat a skill like a harmless extension, but it is really code with access to the runtime’s trust boundary. That is a different level of risk.
Supply-chain mistakeWhat ties all three together is the same operator error: connecting authority before creating boundaries.
Flowchart: should I connect this account to OpenClaw?
Before you hand any account or credential to your agent, run it through this decision tree. If you hit a red stop, fix that layer first.
The principle: never connect authority before containment is real. A “no” at any step means fixing that layer before proceeding.
Chart: what actually increases your OpenClaw blast radius
Not all risks are equal. Here is the practical ranking I would use for non-engineers deciding what to worry about first.
Blast-radius amplifiers
This is why I do not think the right beginner advice is “learn Docker first” or “become a security engineer.” The right beginner advice is “do not connect the dangerous stuff until your containment is real.”

What “run OpenClaw safely” should mean for a normal person
If you are not a security engineer, “safe enough to evaluate” should mean something like this:
| Safer evaluation setup | What to avoid | Why the difference matters |
|---|---|---|
| Dedicated VM or separate test box | Your everyday MacBook or work PC | Keeps personal and business data outside the immediate blast radius |
| Fresh throwaway email and API credentials | Main Gmail, calendar, GitHub, Slack | Compromise hurts less and recovery is simpler |
| One private test channel | Every chat app and automation surface at once | Reduces both exposure and debugging complexity |
| Sandbox enabled with narrow mounts | Host-level execution with broad file access | Contains mistakes and limits what the agent can touch |
| Manual approval mindset for risky actions | Blind trust in autonomous sending or execution | High-risk actions need human friction |
If you set it up this way, you are no longer treating OpenClaw like a toy. You are treating it like a runtime with authority, which is the mature posture.
Sandboxing: the single best control most beginners skip
The official sandboxing docs say something I like because it is refreshingly honest: sandboxing is not a perfect security boundary, but it materially limits filesystem and process access when the model does something dumb. That is exactly the right way to think about it.
Non-engineer translation: a sandbox is not magic, but it is still the difference between “the agent messed up in a smaller room” and “the agent messed up in your whole house.”
Real config snippets: what the hardened settings actually look like
This is the part most guides skip. Here are actual OpenClaw configuration examples from the official docs that map directly to the checklist above. All settings go in ~/.openclaw/openclaw.json using JSON5 format.
Gateway auth token (checks #3 and #4)
The gateway binds to loopback by default. Set a strong auth token using an environment variable so the secret stays out of the config file itself.
{
gateway: {
auth: {
token: "${OPENCLAW_GATEWAY_TOKEN}"
},
port: 18789
}
}
Then export the token in your shell: export OPENCLAW_GATEWAY_TOKEN="your-long-random-string". Generate one with openssl rand -hex 32.
Sandbox configuration (check #7)
Enable sandboxing so tool execution runs in isolated Docker containers rather than directly on the host. The non-main mode sandboxes non-default agents; use all for maximum isolation.
{
agents: {
defaults: {
sandbox: {
mode: "non-main", // off | non-main | all
scope: "agent" // session | agent | shared
}
}
}
}
Build the sandbox image first with scripts/sandbox-setup.sh before enabling. For Docker, add network: "none" and readOnlyRoot: true for tighter isolation.
Channel access control with pairing (checks #5 and #6)
Restrict who can message the bot. The pairing policy requires unknown senders to complete a short-code approval flow before their messages are processed.
{
channels: {
telegram: {
enabled: true,
botToken: "123456:ABCDEF",
dmPolicy: "pairing",
allowFrom: ["tg:YOUR_USER_ID"],
groups: {
"*": { requireMention: true }
}
}
}
}
Approve new users with openclaw pairing approve telegram <code>. Start with one channel, then expand after a clean week.
Skills allowlist (checks #8 and #9)
Control which skills each agent can access. Use an empty array to lock an agent down completely, or list only the specific skills it needs.
{
agents: {
defaults: {
skills: ["weather"]
},
list: [
{ id: "researcher" }, // inherits defaults
{ id: "locked-down", skills: [] } // no skills at all
]
}
}
Vet every community skill before installing. Check the source, review the code, and remove anything you do not actively use.
Secrets via environment variables (check #10)
Never hardcode API keys in your config. Use environment variable references so the config file itself is not a liability.
{
models: {
providers: {
openai: { apiKey: "${OPENAI_API_KEY}" },
anthropic: { apiKey: "${ANTHROPIC_API_KEY}" }
}
}
}
Rotate any keys you used during early experimentation. Treat day-one credentials as burned and replace them with scoped, dedicated keys.
Before and after: risky defaults versus hardened config
Here is what the same setup looks like with weak defaults versus intentional security. This is the difference between “I installed it” and “I secured it.”
Why the external warnings matter: Bitsight and Microsoft changed the conversation
The reason this topic has real search demand is that the risk story stopped being theoretical. Bitsight reported observing more than 30,000 exposed instances during one analysis window and called out attackers going after both prompt-level manipulation and direct gateway-level abuse. Microsoft’s security research framed self-hosted agent runtimes as a convergence of untrusted code, untrusted instructions, and durable credentials.
Those are not the same voice, and that is exactly why they matter together. The official docs tell you what to do. Bitsight shows what people get wrong in the wild. Microsoft explains why the trust boundary itself is unusual.
The minimum safe setup I would actually recommend
If a friend asked me how to evaluate OpenClaw this weekend without being careless, this is the setup I would suggest.
Step-by-step blueprint: secure OpenClaw evaluation in 30 minutes
This is the setup I would walk a friend through on a Saturday afternoon. Mac users can use UTM or Parallels for the VM. Linux users can use a spare VPS from DigitalOcean or Hetzner for about five dollars per month.
Phase 1: Isolate the environment (10 minutes)
Create a dedicated VM or use a spare machine. Do not run OpenClaw where your real browser sessions, downloads, SSH keys, or work files live.
# On your VM or spare box — install OpenClaw npm install -g openclaw@latest # Run the onboarding wizard (installs the background daemon) openclaw onboard --install-daemon
Requires Node 24 (recommended) or Node 22.16+. If you do not have Node, install via nvm install 24.
Phase 2: Lock down auth and gateway (5 minutes)
Generate a strong random token and set it as an environment variable. The gateway stays on loopback by default — do not change that.
# Generate a strong token export OPENCLAW_GATEWAY_TOKEN=$(openssl rand -hex 32) echo "Save this token somewhere safe: $OPENCLAW_GATEWAY_TOKEN" # Add to your shell profile so it persists echo 'export OPENCLAW_GATEWAY_TOKEN="'$OPENCLAW_GATEWAY_TOKEN'"' >> ~/.bashrc
In your ~/.openclaw/openclaw.json, the gateway auth references this variable: auth: { token: "${OPENCLAW_GATEWAY_TOKEN}" }.
Phase 3: Create dedicated credentials (5 minutes)
Create a fresh email account, a new API key scoped to this evaluation, and a dedicated Telegram bot if you are using Telegram as your test channel. Do not reuse your primary accounts.
# Create a scoped API key (example: Anthropic) # Go to console.anthropic.com → API Keys → Create Key # Name it "openclaw-eval" and set usage limits # Export it export ANTHROPIC_API_KEY="sk-ant-your-new-key-here"
If you connected your real accounts during early experimentation, rotate those credentials now. Treat early keys as burned.
Phase 4: Enable sandboxing and restrict tools (5 minutes)
Edit ~/.openclaw/openclaw.json to enable sandboxing for all agents and start with zero skills.
{
agents: {
defaults: {
sandbox: { mode: "all", scope: "session" },
skills: []
}
},
channels: {
telegram: {
dmPolicy: "pairing",
allowFrom: ["tg:YOUR_USER_ID"]
}
}
}
Build the sandbox image first: scripts/sandbox-setup.sh. Only add skills after you have reviewed their source code.
Phase 5: Test, observe, expand (ongoing)
Send a few test messages through your single channel. Watch the responses. Check that sandboxing is working. After a clean week of use with no surprises, you can consider widening access one layer at a time.
# Monitor gateway logs openclaw gateway --verbose # Run the built-in security audit openclaw audit
Keep a plain-text log of what you connected and when. If something feels off, you should be able to revoke tokens, wipe the environment, and rebuild from scratch in under 15 minutes.
That setup will not make OpenClaw perfectly safe, because there is no such thing. It will make you dramatically less likely to hand a powerful runtime the keys to your actual life before you understand how it behaves.
A useful video if you want to see the “secure setup” mindset in practice
For readers who want a visual walkthrough, this secure-setup video is a useful companion because it emphasizes isolation, least privilege, and firewall thinking instead of just speed-running the install.
External security resources worth bookmarking
This article is an operator-level overview. If you want to go deeper, these are the most useful external sources I have found, each covering a different angle of the risk.
| Resource | What it covers | Best for |
|---|---|---|
| OpenClaw Security Docs | Official guidance on auth, access control, gateway exposure, and prompt injection | First read for any new user |
| OpenClaw Sandboxing Docs | How sandbox isolation works, configuration options, and limitations | Anyone enabling sandboxing |
| Bitsight: OpenClaw Security Risks | 30,000+ exposed instances, weak auth patterns, real-world attack surface data | Understanding scale of the problem |
| Microsoft Security Blog | Identity isolation, runtime risk, treating agents as untrusted code execution | Enterprise and security-minded operators |
| SlowMist Security Practice Guide | 3-tier defense matrix, agent-facing security rules, nightly audit automation | Advanced hardening and continuous monitoring |
Field notes: what most people still get wrong
After reading the official guidance and the surrounding security reporting, I think there are four mistakes that keep repeating.
1) People mistake “self-hosted” for “safe”
Self-hosted only tells you where it runs. It does not tell you whether it is isolated, permissioned correctly, or protected from malicious input. A bad local setup can still be a bad setup.
2) People connect valuable accounts too early
The fastest path to regret is connecting real inboxes, work repos, or production systems before containment is in place. Curiosity is fine. Curiosity with authority is where trouble starts.
3) People think the danger is only “prompt injection”
Prompt injection matters, but the broader story is identity and execution. If a runtime can install skills, run tools, and act with live credentials, the security problem is larger than “can the model be tricked?”
4) People do not plan for rebuilds
Mature operators assume recovery is part of the model. If the only plan is “hopefully nothing weird happens,” then there is no plan.
If you want my blunt recommendation: treat OpenClaw like an experiment you earn the right to trust, not like a life operating system you wire into everything on day one.
Rate your setup: OpenClaw security score
Answer these 12 questions honestly. Your score tells you whether your current setup is ready for serious use or still in “tinkering with risk” territory.
One area that intersects directly with security is memory. MEMORY.md and daily notes accumulate everything the agent has been told to remember — including preferences, tool configs, account references, and anything you have explicitly asked it to retain. That file lives on disk in plain text, and if you are using a cloud embedding provider, its contents leave your machine to generate search indexes. Before configuring your memory backend, it is worth understanding exactly what gets written and where it goes. My guide on how OpenClaw memory actually works covers the privacy trade-offs and what a clean, minimal MEMORY.md setup looks like.
FAQs
What is the biggest OpenClaw security risk?
The biggest risk is not one single bug. It is the combination of untrusted input, powerful tools, and real credentials. If those are all present at once, the blast radius gets large very quickly.
Can non-engineers use OpenClaw safely?
They can use it more safely, yes, but only if they slow down and use containment-first defaults: dedicated environment, dedicated accounts, strong auth, minimal tools, and a rebuild plan.
Should I run OpenClaw on my normal laptop?
I would not recommend it for serious evaluation, especially if that laptop contains personal or work data you care about. A dedicated VM or spare machine is the better call.
What does blast radius mean in OpenClaw?
It means the amount of damage or leakage that becomes possible if the agent is manipulated or misbehaves. More tools, more credentials, and broader access all increase blast radius.
Does sandboxing make OpenClaw safe?
No. It makes it safer. That is still important. The official docs frame sandboxing as a meaningful reduction in blast radius, not a guarantee of perfect security.
What is the safest way to start with OpenClaw?
Start on a dedicated environment with fresh credentials, one low-risk channel, strong authentication, sandboxing enabled, and no high-value accounts connected yet.

