Vigilance

AdvisoryAudited by Static analysis on Apr 30, 2026.

Overview

No suspicious patterns detected.

Findings (0)

Artifact-based informational review of SKILL.md, metadata, install specs, static scan signals, and capability signals. ClawScan does not execute the skill or run runtime probes.

What this means

The agent may not run the code the user thinks they installed, or may fail and bypass the intended guardrail workflow.

Why it was flagged

The reviewed registry slug is vigilance, but the top-level instructions identify the skill as agent-sentinel and point execution to an agent-sentinel install path. This can cause the agent to execute a different or missing script instead of the reviewed vigilance path.

Skill content
name: agent-sentinel ... python3 ~/.openclaw/skills/agent-sentinel/eval_engine.py
Recommendation

Publish a single coherent skill identity and make all documented command paths match the registry slug and packaged files.

What this means

A user expecting only a tool-call safety guardrail may also install instructions and code for a different LLM-based response evaluator.

Why it was flagged

This nested skill under the vigilance package describes a response-quality scoring tool, not the top-level EBE guardrail. Conflicting descriptions can mislead users and agents about what behavior is actually being installed.

Skill content
name: agent-eval-engine ... Deterministic quality-control evaluator for AI agent outputs. Scores a response 0–100
Recommendation

Remove unrelated nested skill artifacts or clearly separate them into distinct packages with accurate names, versions, and descriptions.

What this means

The skill can pause, block, or require confirmation for common actions such as searches, shell commands, bookings, and payments.

Why it was flagged

The skill deliberately inserts itself before broad, high-impact tool categories. That is aligned with a guardrail purpose, but users should notice how much agent behavior it controls.

Skill content
`web_search` | Every search, without exception ... `shell_command` | Every shell or terminal execution ... `payment_tool` | Every payment
Recommendation

Review SENTINEL_CONFIG.md and only enable mandatory gating if you want this skill to control those tool categories.

What this means

Private search terms, command lines, URLs, travel details, or payment context could be shared with the configured LLM provider.

Why it was flagged

The evaluator builds an LLM-judge prompt from the user's intent and exact action payload. For web searches, shell commands, payments, or bookings, that payload may contain sensitive information and may be sent to the selected provider.

Skill content
f"User intent : {intent}\n" ... f"Payload/data: {data}\n\n" ... raw = _llm_call(_TIER1_LLM_SYSTEM, user_prompt, provider, model
Recommendation

Use a trusted provider, avoid including secrets in payloads, and consider a local provider such as Ollama for sensitive workflows.

What this means

Installing and using the skill may consume Anthropic or OpenAI API quota and grants the local script access to those provider keys.

Why it was flagged

The included evaluator reads provider API keys from the environment to call LLM judge services. This is expected for an LLM-evaluator skill, but it is still delegated account access.

Skill content
api_key = os.getenv("ANTHROPIC_API_KEY") ... client = _anthropic.Anthropic(api_key=api_key)
Recommendation

Use environment-scoped API keys, monitor provider usage, and do not place broader credentials in the same environment unnecessarily.