Install
openclaw skills install anakinConvert websites into clean data at scale - scrape URLs, batch scrape, AI search, and autonomous research
openclaw skills install anakinConvert websites into clean data at scale using the anakin-cli. Supports single URL scraping, batch scraping, AI-powered search, and autonomous deep research.
Check status and authentication:
anakin status
Output when ready:
✓ Authenticated
Endpoint: https://api.anakin.io
Account: user@example.com
If not installed: pip install anakin-cli
Always refer to the installation rules in rules/install.md for more information if the user is not logged in.
If not authenticated, run:
anakin login --api-key "ak-your-key-here"
Get your API key from anakin.io/dashboard.
Create a .anakin/ folder in the working directory unless it already exists to store results. Add .anakin/ to the .gitignore file if not already there. Always use -o to write directly to file (avoids flooding context):
mkdir -p .anakin
echo ".anakin/" >> .gitignore
anakin scrape "https://example.com" -o .anakin/output.md
Extract content from a single web page in multiple formats.
When to use:
Basic usage:
# Clean readable text (default markdown format)
anakin scrape "https://example.com" -o output.md
# Structured data (JSON)
anakin scrape "https://example.com" --format json -o output.json
# Full API response with HTML and metadata
anakin scrape "https://example.com" --format raw -o output.json
Advanced options:
# JavaScript-heavy or single-page app sites
anakin scrape "https://example.com" --browser -o output.md
# Geo-targeted scraping (country code)
anakin scrape "https://example.com" --country gb -o output.md
# Custom timeout for slow pages (in seconds)
anakin scrape "https://example.com" --timeout 300 -o output.md
Scrape up to 10 URLs at once for efficient parallel processing.
When to use:
Basic usage:
# Batch scrape multiple URLs (up to 10)
anakin scrape-batch "https://example.com/page1" "https://example.com/page2" "https://example.com/page3" -o batch-results.json
For large lists (>10 URLs):
# First batch (URLs 1-10)
anakin scrape-batch "https://url1.com" ... "https://url10.com" -o batch-1.json
# Second batch (URLs 11-20)
anakin scrape-batch "https://url11.com" ... "https://url20.com" -o batch-2.json
Output format: JSON file with combined results, each URL's status (success/failure), content, metadata, and any errors.
Run intelligent web searches to find pages, answer questions, and discover sources.
When to use:
Basic usage:
# AI-powered web search
anakin search "your search query here" -o search-results.json
Follow-up workflow:
# 1. Search for relevant pages
anakin search "machine learning tutorials" -o search-results.json
# 2. Scrape specific results for full content
anakin scrape "https://result-url-from-search.com" -o page.md
Output format: JSON file with search results including titles, URLs, snippets, relevance scores, and metadata.
Run comprehensive autonomous research that explores the web and returns detailed reports.
When to use:
Basic usage:
# Deep agentic research (takes 1-5 minutes)
anakin research "your research topic or question" -o research-report.json
# With extended timeout for complex topics
anakin research "comprehensive analysis of quantum computing" --timeout 600 -o research-report.json
⏱️ Important: Deep research takes 1-5 minutes and runs autonomously. Always inform the user about this duration before starting.
What it does:
Output format: JSON file with executive summary, detailed report by subtopics, key insights, citations with URLs, confidence scores, and related topics.
Use anakin scrape when:
--browser)Use anakin scrape-batch when:
Use anakin search when:
Use anakin research when:
?, &, # charactersanakin scrape "https://example.com?param=value" not anakin scrape https://example.com?param=value-o <file> to save output to a file rather than flooding the terminal--format json for structured data processing--format raw for full API response with HTML--browser only when standard scrape returns empty or incomplete contentanakin login rather than retrying the same command| Error | Solution |
|---|---|
| HTTP 401 (Unauthorized) | Re-run anakin login --api-key "your-key" |
| HTTP 429 (Rate Limited) | Wait before retrying, do not loop immediately |
| Empty content | Try adding --browser flag for JavaScript-heavy sites |
| Timeout | Increase with --timeout <seconds> for slow pages |
| Batch partial failure | Check output JSON for individual statuses, retry failed URLs with --browser |
| Research fails | Fall back to search + multiple scrape calls manually |
.md.json.jsonanakin scrape "https://blog.example.com/article" -o article.md
anakin scrape-batch "https://store1.com/product" "https://store2.com/product" "https://store3.com/product" -o products.json
# Step 1: Find relevant URLs
anakin search "best coffee shops in Seattle" -o coffee-search.json
# Step 2: Scrape the top results
anakin scrape-batch "url1" "url2" "url3" -o coffee-details.json
anakin research "market trends in electric vehicle adoption 2024-2026" -o ev-research.json
anakin scrape "https://spa-application.com" --browser -o spa-content.md
anakin scrape "https://news-site.com" --country us -o us-news.md
anakin scrape "https://news-site.com" --country gb -o gb-news.md
-o flag to save results to filesanakin status before starting workscrapescrape-batchsearch firstresearch# Check status
anakin status
# Re-authenticate
anakin login --api-key "ak-your-key-here"
--browser flag for JavaScript-heavy sites--timeout 300--country <code>