CLAW-BREAKER exists for one outcome:
Give you a clear trust decision before a third-party OpenClaw skill can touch your real environment.
Most malicious skills look normal during code review. The risk shows up only when they run. CLAW-BREAKER runs skills in a sandbox first, triggers risky behavior with adversarial prompts, and returns evidence you can act on.
Without runtime validation, you are guessing.
- A bad skill can exfiltrate data or abuse tools after installation.
- A static read can miss hidden trigger conditions.
- A false sense of safety is worse than a visible warning.
CLAW-BREAKER reduces that uncertainty before deployment.
When evaluating a new skill, you need to answer three questions fast:
- Will this skill try to do things it should not?
- Can I reproduce that behavior, or is this a false alarm?
- Should I block, review, or allow this skill?
CLAW-BREAKER is designed to answer those questions with observed behavior, not assumptions.
For each scan, CLAW-BREAKER gives you:
- A go/no-go trust signal (
allow,review, orblock) - Sandbox evidence showing what the skill actually did
- Reproducible traces (prompt paths, outputs, violations)
- Faster reviews with less manual triage
- Submit a skill source (GitHub or local path).
- CLAW-BREAKER stages and runs the skill in an isolated sandbox.
- It drives the skill with adversarial prompts to trigger hidden behavior.
- Runtime monitoring captures suspicious actions.
- A report is generated with severity and supporting evidence.
The output is intended for one operational action: install, quarantine, or reject.
- Operators running OpenClaw agents with third-party skills
- Teams curating internal skill allowlists
- Security reviewers who need runtime evidence before approval
CLAW-BREAKER is not trying to be:
- A generic SAST platform
- A full SOC/SIEM replacement
- A guarantee that no future exploit is possible
It is a pre-installation behavioral gate for skill trust decisions.
POST /api/v1/scan
{
"skill_source": {
"type": "github",
"url": "https://github.com/openclaw/openclaw/tree/main/skills/weather"
}
}GET /api/v1/report/{execution_id}
Prerequisites:
- Python 3.11+
- Docker
python3 -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt -r requirements-dev.txt
uvicorn src.main:app --host 0.0.0.0 --port 8000 --reloadTests:
pytest -q- Execute skills in sandbox, not on host.
- Keep runtime environment minimal and non-sensitive.
- Treat reports as decision support for trust gating.
If sandboxing is unavailable, scans should be treated as non-authoritative.