Install
openclaw skills install mirroirControl a real iPhone through macOS iPhone Mirroring — screenshot, tap, swipe, type, launch apps, record video, OCR, and run multi-step scenarios. Works with...
openclaw skills install mirroirUse mirroir to control a real iPhone through macOS iPhone Mirroring. Screenshot, tap, swipe, type, launch apps, record video, OCR the screen, and run multi-step automation scenarios — all from the terminal. Works with any app on screen, no source code or jailbreak required.
✅ USE this skill when:
❌ DON'T use this skill when:
imsg skillapple-reminders skillapple-notes skillpeekaboo skillcamsnap skillmirroir status first)After installing, run the setup to configure the helper daemon and Karabiner:
# One-line install (recommended)
/bin/bash -c "$(curl -fsSL https://mirroir.dev/get-mirroir.sh)"
# Or via Homebrew
brew tap jfarcand/tap && brew install iphone-mirroir-mcp
# Or via npx
npx -y iphone-mirroir-mcp install
Approve the Karabiner DriverKit extension if prompted: System Settings > General > Login Items & Extensions — enable all toggles under Karabiner-Elements.
Mirroir is an MCP server. Configure it in your OpenClaw MCP settings:
{
"mirroir": {
"command": "npx",
"args": ["-y", "iphone-mirroir-mcp"]
}
}
Or if installed via Homebrew, use the binary path directly:
{
"mirroir": {
"command": "iphone-mirroir-mcp"
}
}
The typical workflow for any iPhone task:
mirroir status — verify iPhone Mirroring is connectedmirroir describe_screen — OCR the screen to find tap targetsscreenshot — Capture the iPhone screen as PNGdescribe_screen — OCR the screen, returns text elements with exact tap coordinates plus a grid-overlaid screenshotget_orientation — Report portrait/landscape and window dimensionsstatus — Connection state, window geometry, device readinesscheck_health — Full diagnostic: mirroring, helper, Karabiner, screen capturetap x y — Tap at coordinatesdouble_tap x y — Two rapid taps (zoom, text selection)long_press x y — Hold tap for context menus (default 500ms)swipe from_x from_y to_x to_y — Swipe between two pointsdrag from_x from_y to_x to_y — Slow drag for icons, sliderstype_text "text" — Type text via Karabiner virtual keyboardpress_key key [modifiers] — Send special keys (return, escape, tab, arrows) with optional modifiers (command, shift, option, control)shake — Trigger shake gesture (Ctrl+Cmd+Z) for undo/dev menuslaunch_app "AppName" — Open app via Spotlight searchopen_url "https://..." — Open URL in Safaripress_home — Go to home screenpress_app_switcher — Open app switcherspotlight — Open Spotlight searchscroll_to "label" — Scroll until a text element becomes visible via OCRreset_app "AppName" — Force-quit app via App Switcherstart_recording — Start video recording of the mirrored screenstop_recording — Stop recording and return the .mov file pathmeasure action until [max_seconds] — Time a screen transitionset_network mode — Toggle airplane/Wi-Fi/cellular via Settingslist_scenarios — List available YAML automation scenariosget_scenario "name" — Read a scenario fileCoordinates are in points relative to the mirroring window's top-left corner. Always use describe_screen first to get exact tap coordinates via OCR. The grid overlay helps target unlabeled icons (back arrows, gears, stars).
1. launch_app "Messages"
2. describe_screen → find "New Message" button coordinates
3. tap [x] [y] on "New Message"
4. type_text "Alice"
5. describe_screen → find Alice in suggestions
6. tap [x] [y] on Alice
7. tap [x] [y] on the message field
8. type_text "Running 10 min late"
9. press_key return
10. screenshot → confirm sent
1. launch_app "MyApp"
2. describe_screen → find Email field
3. tap [x] [y] on Email
4. type_text "${TEST_EMAIL}"
5. tap [x] [y] on Password
6. type_text "${TEST_PASSWORD}"
7. tap [x] [y] on "Sign In"
8. describe_screen → verify "Welcome" appears
1. launch_app "Waze"
2. describe_screen → read ETA to current destination (e.g. "23 min")
3. press_home
4. launch_app "Slack"
5. describe_screen → find target channel
6. tap [x] [y] on "#standup"
7. tap [x] [y] on message field
8. type_text "Heads up — Waze says 23 min out, be there by 9:25"
9. press_key return
10. screenshot → confirm sent
1. start_recording
2. launch_app "Settings"
3. scroll_to "General"
4. tap [x] [y] on "General"
5. scroll_to "About"
6. tap [x] [y] on "About"
7. stop_recording → returns path to .mov file
Mirroir supports YAML scenario files for multi-step automation flows. Scenarios describe intents, not coordinates — the AI reads the steps and executes them using the MCP tools above, adapting to what's actually on screen.
name: Expo Go Login Flow
app: Expo Go
description: Test the login screen of an Expo Go app with valid credentials
steps:
- launch: "Expo Go"
- wait_for: "${APP_SCREEN:-LoginDemo}"
- tap: "${APP_SCREEN:-LoginDemo}"
- wait_for: "Email"
- tap: "Email"
- type: "${TEST_EMAIL}"
- tap: "Password"
- type: "${TEST_PASSWORD}"
- tap: "Sign In"
- assert_visible: "Welcome"
- screenshot: "login_success"
The step labels (launch, wait_for, tap, type, assert_visible, screenshot) are semantic intents — the AI interprets each one and calls the appropriate MCP tools (launch_app, describe_screen, tap, type_text, screenshot, etc.) to carry them out.
Use list_scenarios to discover available scenarios and get_scenario to load them.
describe_screen before tapping — never guess coordinates.scroll_to "label" to find off-screen elements instead of manual swiping.reset_app before launch_app to ensure a fresh app state.press_key with modifiers (e.g., press_key n [command] for new message in Mail).describe_screen with skip_ocr: true returns only the grid screenshot, letting your vision model identify icons and images OCR can't read.npx iphone-mirroir-mcp setup to reinstall the helper daemonbrew install jfarcand/tap/iphone-mirroir-mcpnpm i -g iphone-mirroir-mcp