Korean Gov Programs

v1.0.8

Collect Korean government support programs (TIPS, Small Business, R&D grants) into structured JSONL files. Supports incremental collection with checkpoints.

0· 705·1 current·1 all-time
byTomas@lifeissea
MIT-0
Download zip
LicenseMIT-0 · Free to use, modify, and redistribute. No attribution required.
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
medium confidence
Purpose & Capability
The name/description describe collecting Korean government program listings; the repo contains scraping code (scripts/collect.py) and a stats helper (scripts/stats.sh) that align with that purpose. There are no unrelated credentials, binaries, or external packages requested.
Instruction Scope
Runtime instructions and code perform HTTP GETs against government sites, parse HTML, and append structured JSONL records to an output directory, storing progress in .checkpoint.json. The code only reads/writes files under the user-specified output directory and does not access unrelated system files or secrets in the visible portion.
Install Mechanism
No install spec; this is instruction+script only and uses only Python standard libraries and a bash helper. Nothing is downloaded from arbitrary URLs or installed automatically.
Credentials
No required environment variables or credentials are declared. The code optionally reads GOV_SCRAPER_UA to override the User-Agent, which is benign but not documented in SKILL.md (optional only). No sensitive credentials are requested.
Persistence & Privilege
Skill does not request always:true and does not modify other skills or system-wide configs. It creates/updates only its own output files and .checkpoint.json in the user-specified directory (append-only behavior is implemented).
Assessment
This appears to be a straightforward scraper: it will perform HTTP requests to government sites and create files in whatever output directory you specify (including a .checkpoint.json). Before running, consider: (1) run it in an isolated directory you control to avoid accidental file placement; (2) respect target sites' robots.txt and rate limits (the script already sleeps 0.8s between requests), and be aware some sources require Selenium/Playwright (skipped by this script); (3) the code optionally reads GOV_SCRAPER_UA to set a User-Agent — harmless but undocumented; (4) the provided collect.py in the review was partially truncated, so if you want higher assurance, open and inspect the entire collect.py for any additional network endpoints or behaviors before executing. If comfortable, run it with a local Python environment only (no elevated privileges) and point --output to a directory you control.

Like a lobster shell, security has layers — review code before you run it.

latestvk978txbeqq3v28g0ajy6z5f6kh81vcqx

License

MIT-0
Free to use, modify, and redistribute. No attribution required.

Comments