Install
openclaw skills install stealthy-auto-browseBrowser automation that passes CreepJS, BrowserScan, Pixelscan, and Cloudflare — zero CDP exposure, OS-level input, persistent fingerprints. Use when standard browser skills get 403s or CAPTCHAs.
openclaw skills install stealthy-auto-browseStealth browser in Docker. Camoufox (custom Firefox) — zero CDP signals. OS-level mouse/keyboard via PyAutoGUI — undetectable. Passes Cloudflare, DataDome, PerimeterX, Akamai.
For installation, configuration, and container setup, see references/setup.md.
curl or WebFetchcurlThe API should already be running. Set the base URL:
export STEALTHY_AUTO_BROWSE_URL=http://localhost:8080
Verify: curl $STEALTHY_AUTO_BROWSE_URL/health returns ok.
All commands: POST $STEALTHY_AUTO_BROWSE_URL/ with JSON body {"action": "name", ...params}.
If AUTH_TOKEN is set on the server, include it on every request (except /health):
Authorization: Bearer <key>
Or pass it as a query param: ?auth_token=<key> (useful for MCP clients that can't set headers).
In single-instance mode, requests are serialized automatically — only one runs at a time, the rest queue up.
Every response:
{
"success": true,
"timestamp": 1234567890.123,
"data": { ... },
"error": "only when success is false"
}
Uses PyAutoGUI for real OS-level events. The browser has no idea it's automated.
system_click — move mouse with human-like curve, then click (viewport x,y coords)mouse_move — move mouse without clicking (hover menus, tooltips)mouse_click — click at position or current location (no smooth movement)system_type — type text character-by-character with randomized delayssend_key — press a key or combo (enter, tab, ctrl+a)scroll — mouse wheel scroll (negative = down)Get viewport coordinates from get_interactive_elements.
Uses Playwright's DOM events. Faster, uses CSS selectors/XPath, but detectable.
click — click by selectorfill — set input value instantlytype — type into element character-by-characterclick with a CSS selector first — fast and reliable.
Only use system_click if the site detects DOM events and blocks them.
system_click requires calibrate first or coordinates will be wrong.fill for inputs (fast). system_type for stealth when bot detection is active.click, fill) is fine.calibrate first.goto → load the pageget_text → read what's on the pageget_interactive_elements → find buttons/inputs with selectors and x,y coordsclick (CSS selector) → interact; fall back to system_click only if neededwait_for_element / wait_for_text → wait for resultsget_text → verify{"action": "goto", "url": "https://example.com"}
{"action": "goto", "url": "https://example.com", "wait_until": "networkidle"}
{"action": "goto", "url": "https://example.com", "referer": "https://google.com/search?q=stuff"}
{"action": "refresh"}
{"action": "refresh", "wait_until": "networkidle"}
wait_until: "domcontentloaded" (default), "load", "networkidle".
referer: set HTTP Referer header (for sites that check referrer).
Response: {"url": "...", "title": "..."}
{"action": "system_click", "x": 500, "y": 300}
{"action": "system_click", "x": 500, "y": 300, "duration": 0.5}
{"action": "mouse_move", "x": 500, "y": 300}
{"action": "mouse_click", "x": 500, "y": 300}
{"action": "mouse_click"}
{"action": "system_type", "text": "hello world", "interval": 0.08}
{"action": "send_key", "key": "enter"}
{"action": "send_key", "key": "ctrl+a"}
{"action": "scroll", "amount": -3}
{"action": "scroll", "amount": -3, "x": 500, "y": 300}
{"action": "click", "selector": "#submit-btn"}
{"action": "click", "selector": "xpath=//button[@id='submit']"}
{"action": "fill", "selector": "input[name='email']", "value": "user@example.com"}
{"action": "type", "selector": "#search", "text": "query", "delay": 0.05}
{"action": "get_interactive_elements"}
{"action": "get_interactive_elements", "visible_only": true}
{"action": "get_text"}
{"action": "get_html"}
{"action": "eval", "expression": "document.title"}
{"action": "eval", "expression": "document.querySelectorAll('a').length"}
get_interactive_elements returns all buttons, links, inputs with x, y, w, h, text, selector, visible. Pass x, y directly to system_click.
get_text returns visible page text (truncated to 10,000 chars). Call this first after navigating.
# Browser viewport
curl -s "$STEALTHY_AUTO_BROWSE_URL/screenshot/browser?whLargest=512" -o screenshot.png
# Full desktop
curl -s "$STEALTHY_AUTO_BROWSE_URL/screenshot/desktop?whLargest=512" -o desktop.png
Resize params: whLargest=512 (recommended), width=800, height=300, width=400&height=400.
Via action (for script mode — returns base64 with output_id):
{"action": "save_screenshot"}
{"action": "save_screenshot", "type": "desktop"}
{"action": "save_screenshot", "output_id": "my_screenshot", "whLargest": 512}
{"action": "save_screenshot", "path": "/output/page.png"}
Use these instead of sleep.
{"action": "wait_for_element", "selector": "#results", "state": "visible", "timeout": 10}
{"action": "wait_for_text", "text": "Search results", "timeout": 10}
{"action": "wait_for_url", "url": "**/dashboard", "timeout": 10}
{"action": "wait_for_network_idle", "timeout": 30}
state: "visible" (default), "hidden", "attached", "detached".
{"action": "list_tabs"}
{"action": "new_tab", "url": "https://example.com"}
{"action": "switch_tab", "index": 0}
{"action": "close_tab", "index": 1}
Call handle_dialog BEFORE the action that triggers the dialog. Dialogs are auto-accepted by default.
{"action": "handle_dialog", "accept": true}
{"action": "handle_dialog", "accept": false}
{"action": "handle_dialog", "accept": true, "text": "prompt response"}
{"action": "get_last_dialog"}
{"action": "get_cookies"}
{"action": "get_cookies", "urls": ["https://example.com"]}
{"action": "set_cookie", "name": "session", "value": "abc", "url": "https://example.com"}
{"action": "delete_cookies"}
{"action": "get_storage", "type": "local"}
{"action": "set_storage", "type": "local", "key": "theme", "value": "dark"}
{"action": "clear_storage", "type": "local"}
type: "local" (default) or "session".
{"action": "get_last_download"}
{"action": "upload_file", "selector": "#file-input", "file_path": "/tmp/doc.pdf"}
{"action": "enable_network_log"}
{"action": "get_network_log"}
{"action": "clear_network_log"}
{"action": "getclear_network_log"}
{"action": "disable_network_log"}
Capture console.log, console.error, console.warn, etc. Each entry has type, text, location, timestamp.
{"action": "enable_console_log"}
{"action": "get_console_log"}
{"action": "clear_console_log"}
{"action": "getclear_console_log"}
{"action": "disable_console_log"}
{"action": "scroll_to_bottom", "delay": 0.4}
{"action": "scroll_to_bottom_humanized", "min_clicks": 2, "max_clicks": 6, "delay": 0.5}
scroll_to_bottom uses JS (fast). scroll_to_bottom_humanized uses OS-level mouse wheel (undetectable).
{"action": "calibrate"}
{"action": "get_resolution"}
{"action": "enter_fullscreen"}
{"action": "exit_fullscreen"}
Call calibrate after fullscreen changes.
Run multiple actions as one atomic request. Steps with output_id collect results.
{"action": "run_script", "steps": [
{"action": "goto", "url": "https://example.com", "wait_until": "domcontentloaded"},
{"action": "sleep", "duration": 2},
{"action": "get_text", "output_id": "text"},
{"action": "eval", "expression": "document.title", "output_id": "title"}
]}
Also accepts "yaml": "..." with the same YAML format used in script mode.
on_error: "stop" (default) or "continue".
{"action": "ping"}
{"action": "sleep", "duration": 2}
{"action": "close"}
curl $STEALTHY_AUTO_BROWSE_URL/health # "ok" when ready
curl $STEALTHY_AUTO_BROWSE_URL/state # {"status", "url", "title", "window_offset"}
The browser exposes all actions as MCP tools via Streamable HTTP at /mcp/ on the same port as the HTTP API.
http://localhost:8080/mcp/
Connect any MCP-compatible client to that URL. All actions from the HTTP API are available as tools — goto, screenshot, system_click, system_type, eval_js, get_text, get_cookies, run_script (multi-step), browser_action (generic fallback for everything else), and more.
If AUTH_TOKEN is set, connect to http://localhost:8080/mcp/?auth_token=<key>.
Works in both standalone and cluster mode. In cluster mode, only run_script is available (same restriction as HTTP API).
Run multiple browser instances behind HAProxy with a request queue, sticky sessions, and Redis cookie sync. For setup see references/setup.md.
Entry point is http://localhost:8080 — same API. HAProxy queues requests when all instances are busy instead of returning errors.
Script-only enforcement (v1.0.0+): When NUM_REPLICAS > 1, both the HTTP API and MCP server only allow run_script, ping, and sleep. All other individual actions are rejected. Use run_script to bundle multiple actions into a single atomic request — one request = one routing decision = one browser instance handles the entire sequence. All actions remain available as steps inside run_script.
Sticky sessions: HAProxy sets an INSTANCEID cookie. Send it back on subsequent requests to keep routing to the same browser instance. All browser state (tabs, DOM, JS, local storage) lives on that specific container — only cookies sync via Redis.
Redis cookie sync: Cookies set on any instance propagate to all others instantly via PubSub. Log in once, the whole fleet is authenticated.
Pipe a YAML script via stdin, get JSON results on stdout, container exits. No HTTP server.
cat my-script.yaml | docker run --rm -i \
-e TARGET_URL=https://example.com \
psyb0t/stealthy-auto-browse --script > results.json
name: Scrape Example
on_error: stop # "stop" (default) or "continue"
steps:
- action: goto
url: ${env.TARGET_URL}
wait_until: networkidle
- action: sleep
duration: 2
- action: save_screenshot
output_id: screenshot
- action: get_text
output_id: page_text
- action: eval
expression: "document.title"
output_id: title
{
"name": "Scrape Example",
"success": true,
"steps_executed": 5,
"steps_total": 5,
"duration": 3.42,
"step_results": [ ... ],
"outputs": {
"screenshot": "data:image/png;base64,iVBOR...",
"page_text": { "text": "...", "length": 1234 },
"title": { "result": "Example Domain" }
}
}
output_id on any step collects its result into outputs. Screenshots become base64 data URIs.${env.VAR_NAME} substitutes environment variables.on_error: continue keeps going past failures. stop (default) halts.cat <<'EOF' | docker run --rm -i -e URL=https://example.com \
psyb0t/stealthy-auto-browse --script > results.json
name: Quick Scrape
steps:
- action: goto
url: ${env.URL}
wait_until: networkidle
- action: save_screenshot
output_id: screenshot
whLargest: 1024
- action: get_text
output_id: text
- action: eval
expression: "document.title"
output_id: title
EOF
Mount YAML files to /loaders. When goto hits a matching URL, the loader's steps execute instead of normal navigation. Works in both API and script mode.
docker run -d -p 8080:8080 -v ./my-loaders:/loaders psyb0t/stealthy-auto-browse
In script mode:
cat script.yaml | docker run --rm -i \
-v ./my-loaders:/loaders \
psyb0t/stealthy-auto-browse --script
name: News Site Cleanup
match:
domain: news-site.com # exact hostname (www. stripped)
path_prefix: /articles # path starts with
regex: "article/\\d+" # full URL regex
steps:
- action: goto
url: "${url}" # ${url} = original URL
wait_until: networkidle
- action: eval
expression: "document.querySelector('.cookie-banner')?.remove()"
- action: wait_for_element
selector: "article"
timeout: 10
Match fields are optional but at least one is required. All specified fields must match.
scripts/websearch.py)Multi-engine parallel web search using the browser API. Searches Brave, Google, and Bing, extracts structured results (title, URL, snippet) and AI overviews when available.
pip install requests beautifulsoup4
# Search all engines
STEALTHY_AUTO_BROWSE_URL=http://localhost:8080 \
python scripts/websearch.py "your search query"
# Search specific engines
WEBSEARCH_ENGINES=brave,google python scripts/websearch.py "query"
Output is JSON: [{"engine": "brave", "query": "...", "ai_overview": "...", "search_results": [{"title": "...", "url": "...", "snippet": "..."}]}]
Env vars: STEALTHY_AUTO_BROWSE_URL, WEBSEARCH_ENGINES (default: brave,google,bing), AUTH_TOKEN, USER_AGENT.
In cluster mode, each engine gets its own browser instance for true parallelism. In single mode, requests serialize via the request lock.
get_text or get_html first; screenshots are last resortwhLargest=512 — full resolution wastes tokens; fine detail is rarely neededclick with CSS selector — reliable and fast; use system_click only when site blocks DOM eventscalibrate before system_click — without it, coordinates are wrong and clicks missget_interactive_elements before clicking — gets both selectors and coordinateswait_for_element, wait_for_text, wait_for_urlhandle_dialog BEFORE the trigger — dialogs are auto-accepted otherwisecalibrate after fullscreen — coordinate mapping shifts