Decodo Web Scraper

v1.1.0

Search Google, scrape web pages, Amazon product pages, YouTube subtitles, or Reddit (post/subreddit) using the Decodo Scraper OpenClaw Skill.

1· 697·1 current·1 all-time

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for donatasdecodo/decodo-scraper-skill.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "Decodo Web Scraper" (donatasdecodo/decodo-scraper-skill) from ClawHub.
Skill page: https://clawhub.ai/donatasdecodo/decodo-scraper-skill
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Bare skill slug

openclaw skills install decodo-scraper-skill

ClawHub CLI

Package manager switcher

npx clawhub@latest install decodo-scraper-skill
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
Name/description, README, SKILL.md and the included tools/scrape.py are all aligned: the skill calls Decodo's scraping API for Google, universal URLs, Amazon, YouTube, and Reddit. No unrelated capabilities or credentials are requested by the code.
Instruction Scope
SKILL.md and the CLI script focus only on building requests to Decodo's scraper API. The runtime instructions ask the agent to set DECODO_AUTH_TOKEN or put it in a .env file; the script reads only that token and does not direct the agent to read other system files or exfiltrate additional environment variables.
Install Mechanism
This is an instruction-only skill with a small Python helper and a requirements.txt (requests, python-dotenv). There is no download-from-arbitrary-URL or install script; risk from install mechanism is low.
Credentials
The skill legitimately requires a single DECODO_AUTH_TOKEN (Basic auth) to call Decodo's API, which is proportionate. however, registry metadata at the top claims 'Required env vars: none' while SKILL.md and the code require DECODO_AUTH_TOKEN — this metadata mismatch should be fixed. Treat the token as sensitive and only provide one obtained from Decodo's dashboard.
Persistence & Privilege
The skill does not request always:true, does not modify other skills or system configs, and does not attempt privileged or persistent system changes. It uses the agent process to make outbound HTTPS calls to scraper-api.decodo.com.
Assessment
This skill appears to do what it says: it calls Decodo's scraping API and returns results. Before installing, confirm the following: 1) the DECODO_AUTH_TOKEN is required (SKILL.md and tools/scrape.py use it) even though registry metadata omitted that — only provide a token from Decodo's dashboard and store it securely (e.g., local .env, secret manager), 2) the tool makes outbound requests to https://scraper-api.decodo.com — ensure your environment/network policy allows that and that you trust Decodo, 3) scraping content may have legal/ToS implications for sites you target (Amazon, Google, Reddit, YouTube); ensure you have the right to scrape and use scraped data, and 4) if you need stronger assurance, verify the repo origin (the README points to a GitHub repo) and check that the hosted homepage and dashboard domain match your expectations before providing credentials.

Like a lobster shell, security has layers — review code before you run it.

latestvk975z0j4sdfp68hnem63p1e89981eax9
697downloads
1stars
1versions
Updated 2mo ago
v1.1.0
MIT-0

Decodo Scraper OpenClaw Skill

Use this skill to search Google, scrape any URL, or fetch YouTube subtitles via the Decodo Web Scraping API. Search outputs a JSON object of result sections; Scrape URL outputs plain markdown; Amazon and Amazon search output parsed product-page or search results (JSON). Amazon search uses --query. YouTube subtitles outputs transcript/subtitles. Reddit post and Reddit subreddit output post/listing content (JSON).

Authentication: Set DECODO_AUTH_TOKEN (Basic auth token from Decodo Dashboard → Scraping APIs) in your environment or in a .env file in the repo root.

Errors: On failure the script writes a JSON error to stderr and exits with code 1.


Tools

1. Search Google

Use this to find URLs, answers, or structured search results. The API returns a JSON object whose results key contains several sections (not all may be present for every query):

SectionDescription
organicMain search results (titles, links, snippets).
ai_overviewsAI-generated overviews or summaries when Google shows them.
paidPaid/sponsored results (ads).
related_questions“People also ask”–style questions and answers.
related_searchesSuggested related search queries.
discussions_and_forumsForum or discussion results (e.g. Reddit, Stack Exchange).

The script outputs only the inner results object (these sections); pagination info (page, last_visible_page, parse_status_code) is not included.

Command:

python3 tools/scrape.py --target google_search --query "your search query"

Examples:

python3 tools/scrape.py --target google_search --query "best laptops 2025"
python3 tools/scrape.py --target google_search --query "python requests tutorial"

Optional: --geo us or --locale en for location/language.


2. Scrape URL

Use this to get the content of a specific web page. By default the API returns content as Markdown (cleaner for LLMs and lower token usage).

Command:

python3 tools/scrape.py --target universal --url "https://example.com"

Examples:

python3 tools/scrape.py --target universal --url "https://example.com"
python3 tools/scrape.py --target universal --url "https://news.ycombinator.com/"

3. Amazon product page

Use this to get parsed data from an Amazon product (or other Amazon) page. Pass the product page URL as --url. The script sends parse: true and outputs the inner results object (e.g. ads, product details, etc.).

Command:

python3 tools/scrape.py --target amazon --url "https://www.amazon.com/dp/PRODUCT_ID"

Examples:

python3 tools/scrape.py --target amazon --url "https://www.amazon.com/dp/B09H74FXNW"

4. Amazon search

Use this to search Amazon and get parsed results (search results list, delivery_postcode, etc.). Pass the search query as --query.

Command:

python3 tools/scrape.py --target amazon_search --query "your search query"

Examples:

python3 tools/scrape.py --target amazon_search --query "laptop"

5. YouTube subtitles

Use this to get subtitles/transcript for a YouTube video. Pass the video ID (e.g. from youtube.com/watch?v=VIDEO_ID) as --query.

Command:

python3 tools/scrape.py --target youtube_subtitles --query "VIDEO_ID"

Examples:

python3 tools/scrape.py --target youtube_subtitles --query "dFu9aKJoqGg"

6. Reddit post

Use this to get the content of a Reddit post (thread). Pass the full post URL as --url.

Command:

python3 tools/scrape.py --target reddit_post --url "https://www.reddit.com/r/SUBREDDIT/comments/ID/..."

Examples:

python3 tools/scrape.py --target reddit_post --url "https://www.reddit.com/r/nba/comments/17jrqc5/serious_next_day_thread_postgame_discussion/"

7. Reddit subreddit

Use this to get the listing (posts) of a Reddit subreddit. Pass the subreddit URL as --url.

Command:

python3 tools/scrape.py --target reddit_subreddit --url "https://www.reddit.com/r/SUBREDDIT/"

Examples:

python3 tools/scrape.py --target reddit_subreddit --url "https://www.reddit.com/r/nba/"

Summary

ActionTargetArgumentExample command
Searchgoogle_search--querypython3 tools/scrape.py --target google_search --query "laptop"
Scrape pageuniversal--urlpython3 tools/scrape.py --target universal --url "https://example.com"
Amazon productamazon--urlpython3 tools/scrape.py --target amazon --url "https://www.amazon.com/dp/B09H74FXNW"
Amazon searchamazon_search--querypython3 tools/scrape.py --target amazon_search --query "laptop"
YouTube subtitlesyoutube_subtitles--querypython3 tools/scrape.py --target youtube_subtitles --query "dFu9aKJoqGg"
Reddit postreddit_post--urlpython3 tools/scrape.py --target reddit_post --url "https://www.reddit.com/r/nba/comments/17jrqc5/..."
Reddit subredditreddit_subreddit--urlpython3 tools/scrape.py --target reddit_subreddit --url "https://www.reddit.com/r/nba/"

Output: Search → JSON (sections). Scrape URL → markdown. Amazon / Amazon search → JSON (results e.g. ads, product info, delivery_postcode). YouTube → transcript. Reddit → JSON (content).

Comments

Loading comments...