Clawhub Release Auditor

v0.2.1

Validate, package, and verify ClawHub skills before and after publishing. Use when creating or updating a ClawHub skill, preparing a release, diagnosing repe...

0· 150·0 current·0 all-time
bywuu Dao@daowuu

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for daowuu/clawhub-release-auditor.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "Clawhub Release Auditor" (daowuu/clawhub-release-auditor) from ClawHub.
Skill page: https://clawhub.ai/daowuu/clawhub-release-auditor
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Required binaries: clawhub, openclaw
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Bare skill slug

openclaw skills install clawhub-release-auditor

ClawHub CLI

Package manager switcher

npx clawhub@latest install clawhub-release-auditor
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
Name/description match the included scripts and SKILL.md: the scripts implement preflight checks, packaging/verify helpers, history analysis, and release-diff checks. Required binaries (clawhub, openclaw) are reasonable for these operations.
Instruction Scope
SKILL.md instructs running included scripts that read a local skill directory, run packaging validation, and call 'clawhub inspect' for remote verification — all in-scope for a publishing auditor. The scripts scan source files for undeclared env vars/binaries but do not exfiltrate secrets or make unexpected external network calls beyond the expected 'clawhub' CLI usage.
Install Mechanism
No install spec is provided (instruction-only with bundled scripts). No downloads or archive extraction are performed by the skill itself. Scripts are shipped with the skill and executed locally.
Credentials
The skill declares no environment variables and requests no credentials. The code inspects source files to detect env usage but does not itself read or require secrets. No unrelated credentials are requested.
Persistence & Privilege
always:false and normal autonomous invocation settings. The skill does not request permanent presence, does not modify other skills, and does not change system-wide agent configuration.
Assessment
This skill appears to implement exactly what it claims: local preflight checks, packaging validation, and post-publish verification via the 'clawhub' CLI. Before installing or running it, note a few practical cautions: (1) The scripts invoke your local packaging script at ~/project/openclaw/skills/skill-creator/scripts/package_skill.py — that path is an environment assumption and may not exist on your machine; verify or adjust the path before running. (2) The tools will run subprocesses (clawhub inspect, package_skill), which will interact with the network and run whatever logic those CLIs/scripts perform — review package_skill.py and ensure you trust it in your environment. (3) The preflight scanner reads files under the skill dir to detect undeclared env vars and binaries; it does not exfiltrate data, but it will report what it finds. If you want extra caution, run the scripts in a restricted or sandboxed environment and inspect the included Python files (they are short and readable) before use.

Like a lobster shell, security has layers — review code before you run it.

Runtime requirements

Binsclawhub, openclaw
latestvk977qe4e79vydz1gk11w5ha7jh83mcyt
150downloads
0stars
4versions
Updated 1mo ago
v0.2.1
MIT-0

ClawHub Release Auditor

Run a strict preflight before any publish. Prefer stopping with a precise explanation over guessing. Treat repeated versions as a signal that the workflow needs diagnosis, not just another upload.

Workflow

  1. Preflight

    • Run python3 scripts/preflight.py <skill-dir>.
    • Fix all hard errors before continuing.
    • Read warnings carefully; they often explain why a skill ends up suspicious.
  2. Package locally

    • Run python3 ~/project/openclaw/skills/skill-creator/scripts/package_skill.py <skill-dir> [output-dir].
    • If packaging fails, stop and explain the exact validation error.
  3. Confirm before publish

    • Show the skill path, intended version, and any remaining warnings.
    • Do not publish without explicit user confirmation.
  4. Publish

    • Publish from the skill folder, not the .skill archive.
    • After publish, record the exact version that was attempted.
  5. Verify post-publish state

    • Run python3 scripts/verify_publish.py <skill-slug> --expected-version <version>.
    • If latest/version visibility is inconsistent, say so clearly.
    • If scan results matter, check the web page separately and explain whether the issue is pending, version mismatch, or a likely metadata/code mismatch.

What to check during preflight

  • Frontmatter only uses supported keys.
  • name and description are present and sane.
  • Placeholder text is not leaking into examples.
  • Declared metadata.openclaw.requires roughly matches real script usage.
  • Homepage/source metadata exists when possible.
  • Publish path points to the skill directory, not the packaged archive.
  • Local package validation passes before any publish attempt.

Common failure patterns

Frontmatter mismatch

If validation complains about unsupported keys, trust the validator. Do not invent alternate formats from memory.

Metadata drift

If scripts use env vars or binaries that the skill does not declare, expect suspicious scan results. Fix the declaration or the code.

Placeholder leakage

If docs contain example paths like /path/to/..., make sure they are clearly examples and not presented as real files.

Repeated publish loops

If many versions are being published quickly, pause and diagnose:

  • Did packaging actually succeed?
  • Did latest move?
  • Is scan still reading an older version?
  • Is the same metadata mismatch still present?

Scripts

scripts/preflight.py

Checks a skill directory for:

  • frontmatter problems
  • placeholder text
  • likely undeclared env vars and binaries
  • external execution hints
  • package validation failures
  • a simple verdict: do-not-publish, review-before-publish, or ready-to-package

scripts/verify_publish.py

Checks published version state with clawhub inspect and compares it to an expected version.

scripts/analyze_history.py

Inspects recent version history for a public skill and groups releases into rough categories such as docs, metadata, bugfix, and feature work. Use it to study repeated publish loops and sharpen the skill's heuristics.

scripts/failure_buckets.py

Classifies likely publish problems into practical buckets such as frontmatter-invalid, package-validation-failed, latest-not-updated, or no-hard-failure-detected.

scripts/release_worthiness.py

Compares a local skill directory against the latest published version and flags when there is no material diff. Use it to avoid unnecessary republish loops.

Publishing tips

SKILL.md body must have substantial content

ClawHub checks for "Skill content is too thin or templated." This evaluates the SKILL.md body text (markdown below frontmatter), not just the description field.

Why this matters:

  • The description field is only used for UI/search summaries
  • The SKILL.md body is what gets embedded and evaluated for the thin-content check
  • If SKILL.md has only frontmatter and no body text, it will fail even with a perfect description

How to avoid:

  • Always include substantive body content in SKILL.md (at least 300-500 words of meaningful guidance)
  • Include real workflow guidance, usage examples, and operational notes in the body
  • The more comprehensive the SKILL.md body, the less likely it triggers "templated" detection

Other common pitfalls

  • homepage field: Include a valid URL to avoid warnings
  • Empty directories: Remove any empty scripts/, references/, or other directories before packaging
  • Symlinks: These are rejected by the packager and cause failures

References

  • Read references/checklist.md for the release checklist.
  • Read references/research-notes.md when designing heuristics for repeated publish loops and common failure modes.
  • If the skill format or server behavior is unclear, read the official ClawHub skill format docs before guessing. Prefer current docs plus validator output over old habits.

Comments

Loading comments...