Back to skill
Skillv1.0.0

ClawScan security

Three Tier Memory · ClawHub's context-aware review of the artifact, metadata, and declared behavior.

Scanner verdict

SuspiciousFeb 16, 2026, 7:03 AM
Verdict
suspicious
Confidence
medium
Model
gpt-5-mini
Summary
Skill generally implements a three-tier memory manager that matches its description, but there are several inconsistencies in its instructions vs. code (config format, LLM integration), and the script reads/writes a workspace path via an undeclared WORKSPACE_DIR environment variable — review before installing.
Guidance
This skill appears to implement the advertised three-tier memory system, but there are several mismatches between the documentation and the code you should review before installing: (1) The docs ask for a YAML config but the script uses a JSON config file; (2) The docs mention automatic summarization via LLMs, yet the script currently uses a local placeholder summary routine (no LLM network calls) — if you expect integrated LLM summaries you must inspect/modify code to provide the intended API hooks and ensure credentials are handled safely; (3) The script writes files into WORKSPACE_DIR (default /Users/scott/.openclaw/workspace) but the SKILL.md does not declare or highlight this environment variable — set WORKSPACE_DIR to an isolated directory or inspect the default path before running; (4) The long-term store uses chromadb if installed; installing third-party Python packages should be done in a virtualenv and reviewed. Recommendation: review the included scripts/memory_manager.py source yourself (or run it in an isolated environment), confirm where files will be written, and only enable LLM/network integrations after verifying how credentials would be provided and stored. If you need higher assurance, request a version that actually integrates with your intended LLM backend and documents required env vars and install steps.

Review Dimensions

Purpose & Capability
noteThe code implements short/medium/long-term memory (sliding-window JSON, summaries, and a local ChromaDB vector store) which matches the skill's stated purpose. Using a local vector DB (Chroma) and local files is reasonable for this purpose.
Instruction Scope
concernSKILL.md and references instruct running the included Python script and mention YAML config and external LLM models, but the script: (a) actually saves config as JSON (not YAML), (b) implements a placeholder local summarize function instead of calling an LLM, and (c) the behavior writes files to a workspace directory — these mismatches mean the runtime behavior may differ from user expectations. The SKILL.md also suggests using specific models (e.g., 'glm-4-flash', 'gpt-3.5-turbo') though the script does not perform real LLM calls.
Install Mechanism
okNo install spec is provided (instruction-only + included script). That is low-risk in terms of install mechanism because nothing is fetched during install; code is shipped with the skill.
Credentials
noteThe skill declares no required env vars, but the script reads WORKSPACE_DIR (defaulting to '/Users/scott/.openclaw/workspace') to determine where it writes memory files. This environment dependency is not documented in SKILL.md. No credentials or secret env vars are requested, which is proportionate.
Persistence & Privilege
okThe skill does not request always: true and does not modify other skills or system-wide settings. It persists data under a workspace directory (creates files and directories), which is expected behavior for a memory manager.