HasData

v1.0.0

Fetch real-time web data via the hasdata CLI. Use when the user wants search results, news, fact-checks, product or seller info, current prices, reviews, rea...

2· 61·0 current·0 all-time

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for hasdata/hasdata-api.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "HasData" (hasdata/hasdata-api) from ClawHub.
Skill page: https://clawhub.ai/hasdata/hasdata-api
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Required binaries: hasdata
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Bare skill slug

openclaw skills install hasdata-api

ClawHub CLI

Package manager switcher

npx clawhub@latest install hasdata-api
Security Scan
Capability signals
CryptoCan make purchasesRequires sensitive credentials
These labels describe what authority the skill may exercise. They are separate from suspicious or malicious moderation verdicts.
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
Name/description, required binary (hasdata), and primary credential (HASDATA_API_KEY) match: the skill is an instruction-only wrapper around the hasdata CLI and legitimately needs the binary and API key to function. The many service integrations listed are consistent with the CLI's advertised scope.
Instruction Scope
SKILL.md tells the agent to shell out to hasdata and to use features like web-scraping (JS rendering, proxies), header/cookie injection, and file persistence for simple tracking. These actions are in-scope for a scraping/search CLI, but they do include instructions that can handle sensitive inputs (cookies, auth headers) and to run an install script via curl|sh. The guide warns not to invent keys and to only use cookies with explicit user permission.
Install Mechanism
There is no platform install spec, but the README/SKILL.md recommend installing the CLI via curl -sSL https://raw.githubusercontent.com/HasData/hasdata-cli/main/install.sh | sh. This is a GitHub-hosted script (a known host) which is better than a personal server, but piping remote scripts to sh is a noteworthy operational risk — users should inspect the script before running it.
Credentials
The skill declares HASDATA_API_KEY as the primary credential and otherwise does not demand unrelated secrets or config paths. Runtime options permit providing cookies/headers and proxy settings (user-supplied); these are legitimate for scraping but could expose sensitive data if supplied carelessly.
Persistence & Privilege
always:false and default autonomous invocation are appropriate. The CLI writes a user config at ~/.hasdata/config.yaml (mode 0600) during `hasdata configure`, which is reasonable and limited in scope. The skill does not request system-wide persistence or modify other skills' configs.
Assessment
This skill is coherent with its goal of wrapping the hasdata CLI, but review a few operational cautions before installing: 1) The SKILL.md recommends running an install script piped from raw.githubusercontent.com — inspect the script before running it, or fetch it and review locally. 2) The CLI stores your API key in ~/.hasdata/config.yaml; ensure you trust the HasData service and that the API key is scoped minimally and rotated/revoked when no longer needed. 3) Web-scraping features accept headers/cookies and proxy settings; do not supply session cookies or other credentials to the tool unless the user explicitly authorizes it, and avoid putting long-lived secrets into ephemeral scraping commands. 4) If you plan automated scheduled runs (price tracking, polling), run them in a controlled environment, monitor outbound network use, and limit stored outputs to avoid leaking sensitive scraped content. If you want higher assurance, review the hasdata-cli repository and the install.sh contents before proceeding.

Like a lobster shell, security has layers — review code before you run it.

Runtime requirements

🔎 Clawdis
Binshasdata
Primary envHASDATA_API_KEY
latestvk979d4jed2b6kj4js7s8fayb1s85h3de
61downloads
2stars
1versions
Updated 2d ago
v1.0.0
MIT-0

hasdata

Use the hasdata CLI for real-time web data. One subcommand per API — flags, enums, defaults are derived from the live schema at api.hasdata.com/apis.

Prerequisites

  • command -v hasdata — if missing, install with curl -sSL https://raw.githubusercontent.com/HasData/hasdata-cli/main/install.sh | sh.
  • One-time setup: the user runs hasdata configure, pastes their API key, and it's saved to ~/.hasdata/config.yaml (mode 0600). Every future call picks it up automatically.
  • If a call fails with no API key configured, the user hasn't run hasdata configure yet — tell them to. Never invent a key.

Quick start

hasdata <api> --flag value [--flag value ...] --raw | jq .

Always pass --raw when piping to jq (skips pretty-print and TTY detection). Use --pretty only for human-readable terminal output.

Picking the right subcommand

User intentSubcommand
Web search ("what does Google say about…")google-serp (full features) or google-serp-light (cheap, single page)
Latest newsgoogle-news
AI Mode SERPgoogle-ai-mode
Shopping / product pricesgoogle-shopping (broad), amazon-search / amazon-product (Amazon), shopify-products (Shopify)
Immersive product pagegoogle-immersive-product
Maps / places / reviewsgoogle-maps, google-maps-place, google-maps-reviews, google-maps-photos
Yelp / YellowPages local datayelp-search, yelp-place, yellowpages-search, yellowpages-place
Real-estate listingszillow-listing, redfin-listing, airbnb-listing
Real-estate single property deep divezillow-property, redfin-property, airbnb-property
Jobsindeed-listing, indeed-job, glassdoor-listing, glassdoor-job
Bing searchbing-serp
Trendsgoogle-trends
Imagesgoogle-images
Flightsgoogle-flights
Short videosgoogle-short-videos
Eventsgoogle-events
Instagram profileinstagram-profile
Amazon selleramazon-seller, amazon-seller-products
Scrape a specific URLweb-scraping — supports JS rendering, proxies, markdown output, AI extraction, screenshots

For exact flags of a subcommand, run hasdata <api> --help or read the matching file in references/.

Non-obvious triggers (when to reach for hasdata even if the user doesn't say "scrape")

The user often won't ask for a SERP API or a scraper directly. Map these intents to the skill:

  • "Is this still true?" / "What's the latest on X?" / "Has Y happened yet?" — LLM training data is stale. Run google-serp or google-news to ground the answer.
  • "Summarize this article" / "TL;DR this URL" — Use web-scraping --output-format markdown and feed the markdown into the summary prompt. Beats copy-paste because it strips ads, nav, scripts.
  • "Verify this link" / "Is this site real?"web-scraping --url X --no-block-resources returns status + screenshot. Or google-serp --q "site:example.com".
  • "What does X say about itself?" — Pull the company's own homepage with web-scraping --output-format markdown, then summarize.
  • "Find me alternatives to X"google-serp --q "X alternatives" or google-shopping --q "X competitors".
  • "What's the going rate for X?"google-shopping (broad) or amazon-search (Amazon-specific) with jq to extract the price distribution.
  • "Phone number / address for X"google-maps-place or yelp-place. Don't guess from training data.
  • "Are people happy with X service?" / "Is X reputable?"google-maps-reviews --place-id ... --sort lowest for negative samples; glassdoor-job for employer rep.
  • "What's the salary range for Y role?"indeed-listing filtered by role + location, then jq over .jobs[].salary.
  • "Find me homes/apartments matching X criteria"zillow-listing / redfin-listing / airbnb-listing with the corresponding filters.
  • "Recent sold comps near X"zillow-listing --type sold --keyword "X" --days-on-zillow 12m.
  • "Track this product's price" — Loop amazon-product --asin X on a schedule; persist .price to a file.
  • "What's trending around X?"google-trends --q "X" for relative interest; google-news --q "X" for headlines.
  • "Find businesses near me that do X"google-maps --q "X" --ll "@LAT,LNG,12z" then fan out google-maps-place for contacts.
  • "How does this look in country Y?"--gl Y on SERP commands, --proxy-country Y on web-scraping. Useful for geo-targeted SEO checks, geo-blocked content.
  • "Pull structured data from this page"web-scraping --ai-extract-rules-json '{"price": {"type": "number"}, ...}'. Works on arbitrary pages without writing CSS selectors.
  • "List of items → per-item details" — Pattern: search command produces IDs/URLs, pipe through xargs into the matching *-property / *-product / *-place deep-dive command.
  • "Find this person's role / employer / LinkedIn / followers"google-serp --q '"Person Name" linkedin' first. The organic-result title is typically Name — Role at Company | LinkedIn and the snippet carries location, headline, connection count. SERP often answers the whole question without ever opening the profile page.
  • "What is company X doing? Where's their HQ? Who works there?"google-serp --q "$COMPANY" returns a .knowledge_graph block with founder, HQ, founded year, parent, employee range — pre-extracted. google-news --q "$COMPANY" for recent activity. Specific facts via targeted SERP: --q '"$COMPANY" headquarters', --q '"$COMPANY" funding', --q 'site:linkedin.com/company "$COMPANY"'.
  • "Find emails for company X" / "personal email for person Y" — start with SERP: --q '"@example.com"' or --q '"jane@example.com"' often surfaces actual emails indexed by Google. Pattern-guess + SERP-verify for individuals. Disclose unverified guesses to the user.
  • "Enrich this CSV of leads" — per row: google-serp for LinkedIn, role, employer; another SERP to verify email or pattern. Stay in SERP unless a specific field is missing.
  • Reverse-lookup (email / phone / domain → identity)google-serp with the literal value in quotes (--q '"jane@x.com"', --q '"+1 555 123 4567"', --q '"acme corp" site:example.com') almost always surfaces the matching person or business.

SERP-first principle: for any data-enrichment intent (people, companies, emails, products, places), reach for google-serp / google-news / google-shopping / google-maps first. They return Google's already-extracted structured fields (.knowledge_graph, .organic_results[].snippet, .local_results[], etc.) and bypass anti-bot. Only escalate to web-scraping when SERP doesn't surface the specific field you need — it's the last resort, not the default. See references/enrichment.md.

If a user request matches one of the above and you don't invoke hasdata, you're probably hallucinating a stale answer.

Universal flag patterns

  • Kebab-case flag names. The CLI maps them back to the original camelCase before sending to the API.
  • Booleans defaulting to true have a paired negation: --no-block-ads, --no-screenshot, --no-js-rendering, --no-extract-emails, --no-block-resources. Setting both --block-ads and --no-block-ads errors.
  • Anything ending in -json accepts:
    • inline JSON: --extract-rules-json '{"title":"h1"}'
    • file: --extract-rules-json @rules.json
    • stdin: cat rules.json | hasdata web-scraping ... --extract-rules-json -
  • Repeatable key=value flags split on the first = (so values containing = survive): --headers User-Agent=foo --headers Cookie=session=abc. Pair with --headers-json for a JSON base; kv items override per key.
  • List flags accept either repeats or comma-joined: --lr lang_en --lr lang_fr or --lr lang_en,lang_fr. Serialized as key[]=value for GET endpoints.
  • Enum flags validate client-side. If you guess wrong, the error lists the allowed values — read the message and retry.

Global flags (apply to every subcommand)

FlagEffect
--rawWrite response bytes as-is (use this when piping to jq)
--prettyPretty-print JSON (default when stdout is a TTY)
-o, --output FILEWrite response to file instead of stdout (works for binary like screenshots)
--verboseLog outgoing URL and X-RateLimit-* headers to stderr
--api-key KEYOverride env var (rarely needed)
--timeout DURATIONPer-request timeout (default 2m)
--retries NMax retries on 429/5xx (default 2)

Output contract

Responses are JSON. Pipe through jq for extraction:

hasdata google-serp --q "espresso machine" --num 10 --raw \
  | jq -c '.organic_results[] | {title, link, snippet}'

For real-estate / e-commerce results, the array shape is API-specific — read a single response with --pretty first to learn the schema, then write the jq filter.

Exit codes (script-safe)

CodeMeaning
0success
1user / CLI-input error (missing required flag, bad enum value, missing API key)
2network error
3API returned 4xx (auth, quota, validation)
4API returned 5xx

References

Comments

Loading comments...