News Crawler
v1.0.0新闻自动爬取与总结工具。用于抓取指定网站或RSS源的新闻内容,并生成摘要报告。 使用场景: 1. 用户要求"获取今日新闻"、"爬取某网站内容" 2. 用户需要"总结新闻"、"生成日报" 3. 用户指定具体URL要求抓取内容 4. 需要监控特定新闻源的最新动态
⭐ 0· 724·6 current·6 all-time
MIT-0
Download zip
LicenseMIT-0 · Free to use, modify, and redistribute. No attribution required.
Security Scan
OpenClaw
Benign
high confidencePurpose & Capability
Name/description align with the provided artifacts: scripts/rss_fetcher.py fetches RSS feeds and scripts/crawl.py fetches and extracts page text. The SKILL.md describes using an LLM to summarize (LLM calls are expected to be performed by the agent, not the shipped scripts), which is consistent with an instruction-only skill.
Instruction Scope
Runtime instructions only tell the agent to run the included Python scripts and then use an LLM to summarize results. The scripts fetch arbitrary user-supplied URLs (no host whitelisting), so while this is expected for a crawler, it means the agent could be used to fetch internal/private endpoints or other sensitive URLs if directed to — the SKILL.md does advise respecting robots.txt and rate limits.
Install Mechanism
No install spec is provided (instruction-only plus included Python scripts). Nothing is downloaded from external URLs or added to disk by an installer, so install risk is low.
Credentials
The skill declares no required environment variables, credentials, or config paths. The code does not access secrets or environment variables, so requested privileges are proportionate.
Persistence & Privilege
always is false and the skill does not request persistent/always-on presence or modify other skills. Autonomous invocation is allowed (platform default) but not combined with other privilege concerns.
Assessment
This skill appears coherent and does what it says: fetch RSS feeds or web pages and produce text the agent can summarize. Before installing or running it, consider: 1) The scripts will fetch any URL you or the agent supplies — avoid pointing it at internal services or private endpoints (risk of exposing internal data). 2) The fetched article text will be sent to your agent/LLM for summarization, so don't feed paywalled or confidential pages unless you accept that exposure. 3) Review rate limiting and robots.txt compliance for targets you crawl to avoid abuse. 4) The HTML/text extraction is basic; for complex sites you may need to review or harden parsing. If you plan to use this skill in production, run it in a sandbox, audit the code, and consider adding host whitelists, request throttling, and logging controls.Like a lobster shell, security has layers — review code before you run it.
latestvk97771qa4jgfeh96ecfm9zv4px82tqjq
License
MIT-0
Free to use, modify, and redistribute. No attribution required.
