PulpMiner Web Scraper - Convert Any Webpage to Realtime JSON API

v1.0.1

Convert any webpage into structured JSON data using AI. Scrape websites, extract data into custom JSON schemas, and call saved APIs programmatically. Useful for web scraping, data extraction, content monitoring, lead generation, price tracking, and building data pipelines.

4· 1.1k·0 current·0 all-time

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for melvin2016/webscraper-pulpminer.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "PulpMiner Web Scraper - Convert Any Webpage to Realtime JSON API" (melvin2016/webscraper-pulpminer) from ClawHub.
Skill page: https://clawhub.ai/melvin2016/webscraper-pulpminer
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Required env vars: PULPMINER_API_KEY
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Canonical install target

openclaw skills install melvin2016/webscraper-pulpminer

ClawHub CLI

Package manager switcher

npx clawhub@latest install webscraper-pulpminer
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
Name/description describe converting webpages to JSON. The only required secret is PULPMINER_API_KEY and all documented operations call https://api.pulpminer.com — these requirements align with the stated purpose.
Instruction Scope
SKILL.md contains only API usage examples, dashboard/config guidance, and webhook callback examples. It does not instruct the agent to read arbitrary local files, system credentials, or run shell commands. The metadata references storing the API key in config, which matches authentication needs.
Install Mechanism
No install spec or code is provided (instruction-only), so nothing is written to disk or fetched at install time. Low install risk.
Credentials
Only PULPMINER_API_KEY is required. That single credential is appropriate for an API-based scraping service. No unrelated secrets, config paths, or excessive env vars are requested.
Persistence & Privilege
always is false and there is no request to modify other skills or system settings. The skill may be invoked autonomously by the agent (platform default), which is expected for skills that call external APIs.
Assessment
This skill appears coherent: it simply documents how to call PulpMiner's API and needs only your PULPMINER_API_KEY. Before installing, confirm you trust pulpminer.com and are willing to provide that API key (it grants the skill ability to make API calls and incur credit usage). Be cautious when using callback URLs (Zapier/webhooks) because scraped results will be POSTed to those external endpoints — don't send sensitive internal URLs or credentials to the service. If you want to avoid the agent calling the skill autonomously, restrict skill invocation in your agent's policy or only provide the API key when you explicitly call the skill. If you need a deeper assurance, request the skill's code/install spec or a publisher identity (this entry has an unknown source) before granting long-lived credentials.

Like a lobster shell, security has layers — review code before you run it.

Runtime requirements

EnvPULPMINER_API_KEY
latestvk974d08bcrfbs7e6gd3ssrvfc980znqy
1.1kdownloads
4stars
2versions
Updated 2mo ago
v1.0.1
MIT-0

PulpMiner — AI Web Scraping & JSON API

PulpMiner converts any webpage into structured JSON using AI. You provide a URL and optionally a JSON template, and PulpMiner scrapes the page, runs it through an LLM, and returns clean structured data.

Authentication

All API calls require the apikey header:

apikey: <PULPMINER_API_KEY>

Get your API key from https://pulpminer.com/api — click "Regenerate Key" if you don't have one.

Core Workflow

PulpMiner works in two phases:

  1. Create a saved API — Configure a URL, scraper, LLM, and optional JSON template via the PulpMiner dashboard at https://pulpminer.com/api
  2. Call the saved API — Use the external endpoint with your API key to fetch structured JSON

Calling a Saved API

Static API (fixed URL)

curl -X GET "https://api.pulpminer.com/external/<apiId>" \
  -H "apikey: <PULPMINER_API_KEY>"

Returns JSON extracted from the configured webpage.

Dynamic API (URL with variables)

For APIs saved with template URLs like https://example.com/search?q={{query}}&page={{page}}:

curl -X POST "https://api.pulpminer.com/external/<apiId>" \
  -H "apikey: <PULPMINER_API_KEY>" \
  -H "Content-Type: application/json" \
  -d '{"query": "javascript frameworks", "page": "1"}'

The {{variable}} placeholders in the saved URL get replaced with the values you provide.

Response Format

Successful responses return:

{
  "data": { ... },
  "errors": null
}

Error responses return:

{
  "data": null,
  "errors": "Error message describing what went wrong"
}

Caching

  • API responses are cached for 24 hours by default
  • If cache is older than 15 minutes, PulpMiner serves the cached version while refreshing in the background
  • Cache can be disabled per-API in the dashboard settings

Configuration Options (Set in Dashboard)

When creating a saved API at https://pulpminer.com/api, you can configure:

OptionDescription
URLThe webpage to scrape
JSON TemplateOptional JSON structure for the LLM to follow (e.g., {"name": "", "price": ""})
Render JSEnable for SPAs and JS-heavy pages (uses headless browser)
CSS SelectorExtract only a specific part of the page (e.g., .product-list, #main-content)
Extra InstructionsAdditional guidance for the AI (e.g., "Only extract items with prices above $50")
Dynamic URLEnable template variables in the URL with {{variable}} syntax
CacheToggle response caching on/off

Integration with Zapier

For async scraping in Zapier workflows:

# Static API
curl -X POST "https://api.pulpminer.com/external/zapier/get/<apiId>" \
  -H "apikey: <PULPMINER_API_KEY>" \
  -d '{"callbackURL": "https://hooks.zapier.com/..."}'

# Dynamic API
curl -X POST "https://api.pulpminer.com/external/zapier/post/<apiId>" \
  -H "apikey: <PULPMINER_API_KEY>" \
  -d '{"callbackURL": "https://hooks.zapier.com/...", "query": "value"}'

Returns 201 immediately. Sends scraped data to the callback URL when complete.

Integration with n8n

Verify authentication:

curl -X GET "https://api.pulpminer.com/external/n8n/auth" \
  -H "apikey: <PULPMINER_API_KEY>"

Then use the standard /external/<apiId> endpoints for data fetching.

Credits

  • Each API call costs 0.25–0.4 credits depending on the endpoint
  • JavaScript rendering adds 0.1 credits extra
  • New users get 5 free credits
  • Purchase more at https://pulpminer.com/credits

Tips

  • Use CSS selectors to narrow down the scraped content and improve accuracy
  • Provide a JSON template for consistent, predictable output structures
  • Enable JS rendering only when needed — static pages scrape faster and cost fewer credits
  • Use extra instructions to guide the AI (e.g., "Return dates in ISO 8601 format")
  • For monitoring use cases, keep caching enabled to reduce credit usage
  • Use the playground first to verify a URL is scrapable before saving an API config
  • Dynamic APIs are ideal for search pages, paginated content, and parameterized URLs

Links

Comments

Loading comments...