Back to skill
Skillv1.0.1
ClawScan security
LegalFrance · ClawHub's context-aware review of the artifact, metadata, and declared behavior.
Scanner verdict
BenignFeb 13, 2026, 6:44 AM
- Verdict
- benign
- Confidence
- high
- Model
- gpt-5-mini
- Summary
- The skill's files and instructions are coherent with its stated purpose (a French RAG legal assistant); it downloads and indexes a public LEGI corpus and builds local search indexes, but be aware of large downloads, disk writes, and the included system prompt.
- Guidance
- This skill appears to do what it says: build a local RAG assistant over the LEGI dataset and produce SYSTEM+USER prompts for an LLM. Before installing/running: (1) review the scripts yourself (they are included) and run them in an isolated environment/VM if you are concerned; (2) ensure you have ~2+ GB free and sufficient disk space for indexes; (3) be prepared to install Python packages (datasets, chromadb, sentence_transformers) and possibly provide a HuggingFace token if a model requires authentication; (4) note the skill will write persistent indexes under data/, so back up or choose an appropriate working directory; (5) understand the skill supplies a strict system prompt that will control downstream LLM outputs — this is normal for RAG but worth reviewing if you plan to run the LLM with different safety settings.
- Findings
[system-prompt-override] expected: ask.py intentionally defines a detailed SYSTEM_PROMPT to constrain the LLM's behaviour for legal answers; this matches the skill's RAG goal and explains the pre-scan flag.
Review Dimensions
- Purpose & Capability
- okName/description match the code and scripts: ingestion from AgentPublic/legi, ChromaDB + SQLite FTS search, RAG prompt builder and helpers. No unrelated env vars, binaries, or external services are requested.
- Instruction Scope
- noteSKILL.md instructs running ingest/search/one_shot scripts and requires user confirmation before the ingest (which downloads ~2 GB). The code builds a strong SYSTEM_PROMPT for the LLM (this triggered a 'system-prompt-override' pattern). That is expected for a RAG assistant, but you should be aware the skill supplies explicit system-level instructions that will guide any LLM used with the prompts.
- Install Mechanism
- noteThere is no automated install spec (instruction-only), which lowers automated install risk. Running the scripts requires third-party Python packages (datasets, chromadb, sentence_transformers) and will download a large embedding model (BAAI/bge-m3) and the HF dataset. Downloads come from known hosts (HuggingFace, model hub) rather than arbitrary URLs, but expect heavy network and disk activity.
- Credentials
- noteThe skill declares no required environment variables or credentials (good). One caveat: downloading some models or private HF resources can require a HuggingFace token (HF_TOKEN) or similar not declared here; this may cause runtime prompts or failures but is not evidence of hidden credential demands.
- Persistence & Privilege
- okThe skill writes indexes to a local data/ directory (chroma_db and fts_index.db) as expected for a local RAG assistant. It does not request always:true or modify other skills or system-wide settings.
