TestDino — Test Memory for Agents

v1.0.2

Connect OpenClaw to TestDino for real-time Playwright CI intelligence. Ask about test failures, flaky tests, run history, and CI health in plain English.

0· 132·0 current·0 all-time
byTestDino Devs@testdino-inc

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for testdino-inc/testdino-playwright.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "TestDino — Test Memory for Agents" (testdino-inc/testdino-playwright) from ClawHub.
Skill page: https://clawhub.ai/testdino-inc/testdino-playwright
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Required env vars: TESTDINO_PAT
Required binaries: mcporter, testdino-mcp
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Bare skill slug

openclaw skills install testdino-playwright

ClawHub CLI

Package manager switcher

npx clawhub@latest install testdino-playwright
Security Scan
Capability signals
CryptoCan make purchasesRequires OAuth tokenRequires sensitive credentials
These labels describe what authority the skill may exercise. They are separate from suspicious or malicious moderation verdicts.
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
The name/description (TestDino Playwright CI intelligence) match the declared requirements and instructions: it requires mcporter and testdino-mcp and the TESTDINO_PAT token, and all runtime commands use mcporter call testdino.*. Nothing unrelated (e.g., cloud provider creds or unrelated binaries) is requested.
Instruction Scope
SKILL.md is prescriptive and stays within the TestDino use case: it always runs mcporter call testdino.health first and then specific testdino calls. It also instructs editing ~/.mcporter/mcporter.json and ~/.openclaw/openclaw.json and adding cron jobs. These are expected for this integration, but they do cause persistent local config changes and enable exec-based commands which run local binaries — users should review those config edits and understand exec/cron implications before proceeding.
Install Mechanism
This is instruction-only (no remote archive downloads). The metadata references npm packages (mcporter, testdino-mcp) which is proportionate for a Node-based CLI integration. The included install-check.sh only validates Node, npx/npm reachability, and PAT auth against https://api.testdino.com — no downloads from unknown hosts or obfuscated steps.
Credentials
Only TESTDINO_PAT is requested and it is the declared primary credential — this is appropriate. However the instructions encourage placing the PAT in ~/.openclaw/openclaw.json and ~/.mcporter/mcporter.json (plaintext config), which increases exposure risk. Consider using environment-only injection or a secrets store and create a PAT with minimal scope.
Persistence & Privilege
always:false (normal). The skill asks you to enable the 'coding' tools profile so the exec tool works and to create crons that will run scheduled agent turns (requiring your gateway token and channel destinations). Those behaviors are necessary for digests/alerts but grant ongoing capability to run mcporter calls and post to external channels while the gateway runs — review cron entries, gateway tokens, and allowed exec commands before enabling.
Assessment
What to check before installing: (1) Trust the source — the skill will call mcporter/testdino-mcp and needs your TestDino PAT; confirm testdino-mcp is the official npm package. (2) Secrets handling — the README recommends putting TESTDINO_PAT into ~/.openclaw/openclaw.json and ~/.mcporter/mcporter.json (plaintext). If that concerns you, prefer exporting TESTDINO_PAT in the environment or using a secrets manager and create a PAT with minimal permissions. (3) Exec and mcporter — enabling the 'coding' profile and approving mcporter gives the agent the ability to run local binaries via the exec tool; ensure mcporter is configured only to run the expected testdino-mcp command and you approve mcporter as an allowed exec. (4) Crons and gateway tokens — scheduled jobs require your gateway token and post to external channels; review cron payloads, destination channels, and token scope. (5) Verify packages — if you install globally, verify mcporter and testdino-mcp package integrity (official npm pages, package maintainers) before npm -g install. If these items are acceptable, the skill appears coherent with its stated purpose.

Like a lobster shell, security has layers — review code before you run it.

Runtime requirements

OSLinux · macOS · Windows
Binsmcporter, testdino-mcp
EnvTESTDINO_PAT
Primary envTESTDINO_PAT
latestvk977n3fk4z7tfmj9w41s4se8dd85mmnd
132downloads
0stars
2versions
Updated 20h ago
v1.0.2
MIT-0
Linux, macOS, Windows

TestDino — Playwright CI Intelligence

Use the exec tool to call TestDino. Always use mcporter:

mcporter call testdino.<tool> [params]

Important: Always call health first to get the projectId unless the user has already provided it. Every other tool requires it.


Commands (copy these exactly)

Check connection / health check:

exec: mcporter call testdino.health

Project summary / CI overview / "summarize project X":

exec: mcporter call testdino.health
exec: mcporter call testdino.list_testruns projectId=X by_time_interval=1d limit=5

Use the returned run IDs to fetch details if the user wants more depth.

Show failures / failed tests (any time range):

Resolve TIME from the user's words using the time mapping below, then run:

exec: mcporter call testdino.health
exec: mcporter call testdino.list_testcase projectId=X by_status=failed by_time_interval=TIME limit=10

The response includes totalCount — always use that as the total failure count.

Show failures on a branch / any failures on branch Y? / new failures on branch Y?:

exec: mcporter call testdino.health
exec: mcporter call testdino.list_testcase projectId=X by_branch=BRANCH_NAME by_status=failed limit=10

CI summary / how did CI look / test run stats (any time range):

exec: mcporter call testdino.health
exec: mcporter call testdino.list_testruns projectId=X by_time_interval=TIME

Show recent test runs:

exec: mcporter call testdino.health
exec: mcporter call testdino.list_testruns projectId=X limit=10

⚠️ How to read list_testruns response:

  • The testRuns array is the complete flat list — use it directly, ordered newest first
  • Each run has a counter field — always use this as the run number (e.g. "Run #10")
  • Show ALL runs from testRuns, never skip any
  • rerunMetadata.isRerun: true means this run is a retry — note it as "🔁 Retry of #[parentCounter]"

Why did test run [ID or counter] fail? / Details for a specific test run:

If the user provides a test run ID (e.g. test_run_69c3d36648201d8fe8393e33):

exec: mcporter call testdino.health
exec: mcporter call testdino.get_run_details projectId=X testrun_id=test_run_XXX
exec: mcporter call testdino.list_testcase projectId=X by_testrun_id=test_run_XXX by_status=failed limit=10

If the user provides a counter number (e.g. run #42):

exec: mcporter call testdino.health
exec: mcporter call testdino.get_run_details projectId=X counter=N
exec: mcporter call testdino.list_testcase projectId=X counter=N by_status=failed limit=10

Why is [test name] failing? / Debug [test]:

If the user did not provide a test name, ask: "Which test case name would you like to debug?" and wait for their answer before running anything.

Once you have the real test name, ALWAYS run these commands — even if a previous attempt failed. Never give generic advice instead of running the tool:

exec: mcporter call testdino.health
exec: mcporter call testdino.debug_testcase projectId=X testcase_name=EXACT TEST NAME WITHOUT QUOTES

Show flaky tests / is [test] flaky? / is [test] flaky or a real bug?:

If user names a specific test → use debug_testcase (it returns flakiness patterns). If user asks generally → use:

exec: mcporter call testdino.health
exec: mcporter call testdino.list_testcase projectId=X by_status=flaky by_time_interval=TIME limit=10

Is it safe to merge branch Y?:

If no branch name was given, ask: "Which branch should I check?" and wait before running.

exec: mcporter call testdino.health
exec: mcporter call testdino.list_testcase projectId=X by_branch=Y by_status=failed limit=10

Show timeout failures:

exec: mcporter call testdino.health
exec: mcporter call testdino.list_testcase projectId=X by_error_category=timeout_issues by_time_interval=TIME limit=10

List manual test cases / total count / show [priority] test cases / show [type] test cases:

⚠️ Use list_manual_test_cases — NOT list_testcase. These are completely different tools.

exec: mcporter call testdino.health
exec: mcporter call testdino.list_manual_test_cases projectId=X limit=1000

⚠️ IMPORTANT — verify the count: After the response comes back, check the count field. If count is exactly 10, the API used its default limit and you do NOT have all results. You MUST retry with limit=1000:

exec: mcporter call testdino.list_manual_test_cases projectId=X limit=1000

Only trust the count from a response where you explicitly passed limit=1000.

Add filters if user specifies:

  • priority=critical / priority=high / priority=medium / priority=low
  • type=smoke / type=regression / type=functional
  • status=active / status=draft

Examples:

  • "critical priority test cases" → list_manual_test_cases projectId=X priority=critical limit=1000
  • "smoke test cases" → list_manual_test_cases projectId=X type=smoke limit=1000

If user asks for total count: Read the count field from the response and reply with just the number. If user asks to list/show: Show only caseId + title. Show the first 15, then ALWAYS end with: "There are more test cases available. Let me know if you'd like to see more!"

List test suites / show all suites:

exec: mcporter call testdino.health
exec: mcporter call testdino.list_manual_test_suites projectId=X

Show only the name field from each suite. Show the first 15 suites, then end with: "There are more suites available. Let me know if you'd like to see more!"

NEVER show: suite IDs, metadata, hierarchy, timestamps, descriptions (they are empty), or tags (they don't exist in the response). NEVER invent test case counts or test case names — list_manual_test_suites does NOT return test cases.

List spec files / show spec files / all spec files in project:

exec: mcporter call testdino.health
exec: mcporter call testdino.list_testruns projectId=X limit=1
exec: mcporter call testdino.get_run_details projectId=X counter=N

From the get_run_details response, extract testSuites[].fileName and list only the unique spec file names.


Missing parameters

  • No projectId → run health first to get it, unless user already provided it
  • No test name → ask: "Which test case name?"
  • No time range → ask: "Today (1d), 3 days (3d), 7 days, or monthly?"
  • Test run ID vs counter → testrun_id accepts the full string ID; counter accepts a sequential number (e.g. 42)

Time range mapping (apply strictly before every query)

User saysTIME value
today / right now / no time mentioned1d
yesterday / last day1d
last 2 days / 2 days3d
last 3 days / 3 days3d
last 5 days / 5 days3d
this week / last week / 7 days / last 7 days / weeklycompute 7-days-ago and today as YYYY-MM-DD,YYYY-MM-DD e.g. 2026-03-23,2026-03-30 — do NOT use the weekly keyword (it has an exclusive boundary that drops the oldest day)
last 2 weeks / 14 daysmonthly
this month / last month / 30 days / monthlymonthly
specific date e.g. "27 march 2026" / "27-03-2026" / "2026-03-27"2026-03-27,2026-03-28 (next day as end date)
date range e.g. "20 march to 25 march"2026-03-20,2026-03-25

Always normalize any date format the user gives to YYYY-MM-DD. For a specific single day, use YYYY-MM-DD,YYYY-MM-DD+1 (end date = next day). For "last N days" not in the table, use 3d if under 7 days, a computed date range (YYYY-MM-DD,YYYY-MM-DD) if exactly 7 days, monthly if over 7.


Response format

CRITICAL: NEVER invent or fabricate data. Only show fields that exist in the actual API response. If a field is missing or empty, do not guess or fill it in. If the response is too large to process fully, show what you can and say "there are more — ask if you'd like to see more." Never make up descriptions, tags, test counts, or test case names that are not in the response.

Keep responses short. Always use totalCount from the response for the count.

For list_testruns — use this exact format:

**[ProjectName]** — [N] runs

- Run #[counter] — [Mon DD] | [branch] | ✅ [passed] passed ❌ [failed] failed[ | 🔁 Retry of #[parentCounter]]
- Run #[counter] — ...

NEVER include: run IDs, start times, durations, environment, author, status, total tests, skipped/flaky/timedOut counts.

For list_testcase (failures) — use this exact format:

**[ProjectName]** — **[totalCount] failures**

**Run #[counter]**
1. [test title]
2. [test title]

_(showing N of totalCount — ask for more if needed)_

ONLY show run number and test title. NEVER show test case IDs, run IDs, durations, status, start times, spec file paths.

  • Merge safety: lead with yes/no then failure count
  • Debug: 2-3 sentences on dominant pattern, note if flaky
  • Show full details only if the user specifically asks

Comments

Loading comments...