Guardrails

ReviewAudited by ClawScan on May 10, 2026.

Overview

Guardrails is a coherent security-helper skill, but it scans workspace context and may send that context plus your answers to OpenAI or Anthropic to generate policies.

This skill appears purpose-aligned and not malicious. Before installing, be comfortable with local workspace scanning and with sending discovery data/interview answers to your chosen LLM provider; review the generated GUARDRAILS.md before allowing it to be written.

Findings (5)

Artifact-based informational review of SKILL.md, metadata, install specs, static scan signals, and capability signals. ClawScan does not execute the skill or run runtime probes.

What this means

Workspace details and your interview answers may leave your machine and be processed by OpenAI or Anthropic.

Why it was flagged

The guardrails generator sends workspace discovery data, risk classification, and interview answers to an external LLM provider when configured. This is disclosed and purpose-aligned, but it is still a sensitive data flow.

Skill content
Reads {discovery, classification, answers} JSON from stdin, uses an LLM to generate a full GUARDRAILS.md document... OPENAI_API_KEY -> OpenAI API... ANTHROPIC_API_KEY -> Anthropic API
Recommendation

Use this only with a provider and API key you trust, and avoid putting secrets in interview answers or workspace files that will be scanned.

What this means

Private workspace notes or prior agent context could be summarized into discovery output and used when generating guardrails.

Why it was flagged

Discovery reads previews from persistent workspace context files. This supports guardrail generation, but those files may contain private notes or instructions that influence generated policy.

Skill content
important_files=("USER.md" "MEMORY.md" "AGENTS.md" "GUARDRAILS.md" "TOOLS.md" "SOUL.md" "BOOTSTRAP.md" "HEARTBEAT.md") ... preview=$(read_first_lines "$filepath" 20)
Recommendation

Review sensitive workspace files before setup, and keep secrets out of USER.md, MEMORY.md, and similar agent context files.

What this means

Your OpenAI or Anthropic account may be used and billed for question or policy generation.

Why it was flagged

The skill uses provider API keys from the environment for LLM generation. There is no evidence of hardcoded keys or unrelated credential use.

Skill content
openai_key = os.environ.get("OPENAI_API_KEY") ... anthropic_key = os.environ.get("ANTHROPIC_API_KEY")
Recommendation

Set only the provider key you intend to use, monitor provider usage, and remove the key from the environment when no longer needed.

What this means

Installing and running setup/review can change the policy files that guide future agent behavior.

Why it was flagged

The skill mutates persistent workspace files, but the instructions require user review and confirmation before writing.

Skill content
Present the generated GUARDRAILS.md for review ... Ask for confirmation before writing to workspace ... Write GUARDRAILS.md to workspace root ... Save guardrails-config.json to workspace root
Recommendation

Read the generated GUARDRAILS.md carefully before approving the write.

What this means

Setup may fail if dependencies are missing, and package installation depends on the user's Python package source.

Why it was flagged

The skill has manual runtime prerequisites and an unpinned pip install recommendation, while there is no install spec to enforce or pin them.

Skill content
Prerequisites... Python 3... Bash... jq... Python scripts that call LLM providers require `requests`: `python3 -m pip install requests`
Recommendation

Install dependencies from trusted package repositories and consider pinning versions in your own environment.