Dead Link Scanner

v1.0.0

Scan websites, markdown files, and HTML files for broken links (dead links). Use when checking a website for 404s, validating links in documentation or READM...

0· 158·0 current·0 all-time
byJohn Wang@johnnywang2001
MIT-0
Download zip
LicenseMIT-0 · Free to use, modify, and redistribute. No attribution required.
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
Name/description match the included Python implementation: the code crawls HTML pages, extracts links from HTML and Markdown, and checks links' HTTP status. It does not request unrelated capabilities (no cloud creds, no extra binaries).
Instruction Scope
SKILL.md instructs the agent to run the included script to scan a URL or local files. The script only reads files explicitly passed to the 'file' command and performs network requests for discovered links; it does not attempt to read arbitrary system state or extra environment variables.
Install Mechanism
There is no install spec; this is instruction+code only. The Python script is bundled with the skill and nothing is downloaded or executed from external installers.
Credentials
The skill requires no environment variables, credentials, or config paths. The single notable hard-coded value is a USER_AGENT string (dead-link-scanner/1.0 (+https://clawhub.com)), which is benign but reveals an identifying User-Agent to servers it contacts.
Persistence & Privilege
Skill does not request persistent or elevated privileges. always is false and it does not modify other skills or system configurations. It can be invoked autonomously by the agent (default), which is expected behavior for a skill.
Assessment
This skill is coherent and implements a dead-link scanner in Python. Before installing or running it, be aware that: (1) it will make outbound HTTP(S) requests to every discovered URL (including internal/intranet addresses you point it at), which can trigger server logs or expose your agent's IP; (2) it reads only files you pass on the command line — avoid scanning sensitive files you don't intend to expose; (3) it does not appear to obey robots.txt and will crawl pages according to the provided depth/max-urls/delay settings, so use polite settings (delay, max-urls) when crawling third-party sites; and (4) the User-Agent includes an identifying URL, so contacted servers will see that string. If you need autonomous scanning across private networks or sensitive repos, review the script and enforce access controls and network restrictions first.

Like a lobster shell, security has layers — review code before you run it.

latestvk977920p3nfd9t7pbgbvfyqxa182t5bzlinksvk977920p3nfd9t7pbgbvfyqxa182t5bzqavk977920p3nfd9t7pbgbvfyqxa182t5bzseovk977920p3nfd9t7pbgbvfyqxa182t5bzwebvk977920p3nfd9t7pbgbvfyqxa182t5bz

License

MIT-0
Free to use, modify, and redistribute. No attribution required.

Comments