Install
openclaw skills install mumoMulti-model deliberation via mumo's MCP server. Best for contested architecture/product decisions, design reviews, pressure-testing a pre-launch spec, resolving tradeoffs with multiple defensible framings, or explicit user requests for a mumo panel. Requires a mumo platform API key (mmo_live_*) registered with `openclaw mcp set mumo`.
openclaw skills install mumomumo runs deliberations across multiple AI models. Use it when independent perspectives are useful — especially for high-regret decisions where a single model's confidence is a real risk.
The mumo MCP server is registered via openclaw mcp set mumo '<json>' and stored in ~/.openclaw/openclaw.json under mcp.servers.mumo. See config/mumo.example.json in this skill's directory for the canonical config payload (reference only — OpenClaw doesn't auto-load this file; it's the JSON shape to paste into the CLI command). The seven mumo tools surface to the agent as mumo__create_deliberation, mumo__wait_for_round, mumo__append_round, mumo__get_session, mumo__list_sessions, mumo__list_models, mumo__get_credit (OpenClaw uses <server>__<tool> naming).
If tools return auth errors, the API key is missing or invalid. Direct the user to https://mumo.chat/settings/api-keys to create one (keys start with mmo_live_), then re-run openclaw mcp set mumo with the new value and restart OpenClaw.
Use mumo when the cost of being wrong is greater than the cost of deliberation. Three conditions are typically the fingerprint:
These map to concrete trigger types: architecture decisions with non-obvious tradeoffs, plan or design review before commitment, pre-launch pressure tests, stuck debugging after N failed repair attempts, pre-commit adversarial review on risky diffs, memory/skill promotion gates, strategy questions with multiple defensible framings, explicit user requests.
Skip mumo for factual lookups, syntax help, routine refactors, formatting, dependency bumps with clear errors, or normal edit-test cycles. The deliberation tax is real (4–220× tokens, 15–120s latency); spend it on decisions where mistakes compound.
Think of mumo as a cognitive load balancer for your reasoning — not a replacement for your judgment, but a way to externalize a design and stress-test it against perspectives you can't reliably generate yourself.
Autonomous agent loops occasionally fabricate tool-call results — claiming a deliberation was sent when it wasn't. Real mumo session IDs are UUIDs (e.g. 2acdab34-2484-4bc5-a24f-bf917fe81477), not arbitrary hex strings.
If a create_deliberation response doesn't contain a UUID-format session_id, the call did not happen. Verify by calling list_sessions and checking that your latest session matches the prompt you sent. If it doesn't, fire create_deliberation again — don't continue downstream as if the session exists.
If you lose track of session_id or round_id mid-conversation (long chats, context compaction, dropped tool result), recover before starting a new deliberation:
list_sessions to find your latest sessions. Match by prompt content.get_session with the recovered ID for full state, or wait_for_round(session_id, round_id) if you suspect a round is still in flight.create_deliberation if the original is recoverable — duplicate sessions waste tokens and produce confusing parallel state.Two concepts that are easy to conflate:
subagents tool, the sessions_yield flow, etc.). Subagents run in your harness with your tool access against your state.When you describe who's deliberating, say "panel" or "models." Reserve "subagent" for your own delegation surface. The two are complementary — you can spawn a subagent to run benchmarks before invoking a mumo panel — but they're not the same thing.
create_deliberation with the user's problem. Set application: "OpenClaw".session_id and round_id. If anything else, the call didn't fire — retry.wait_for_round with the returned session_id and round_id. Long waits are normal — frontier-model panels typically take 15–120s, and 60+ seconds isn't a failure signal. Tell the user upfront ("running a panel — expect ~30–60s") so the wait doesn't feel broken.append_round if another round would help. Otherwise stop and synthesize for the user.Do not pass user phrasing through verbatim when it describes your identity, role, or local context. Reframe it in third person or first-person moderator context so panel models advise you rather than roleplay you.
Bad:
"Introduce yourself as Clawd, a cheerful lobster assistant, and decide when to use mumo."
Good:
"I am Clawd, a cheerful OpenClaw assistant. Advise me on when I should escalate a decision to mumo. Critique this draft trigger checklist: ..."
The panel participants should answer as themselves, not as the calling agent.
Snippets are moderator attention. They mark what mattered in the prior round and optionally explain why.
| Type | Reaction |
|---|---|
| KEEP | this seems worth preserving |
| EXPLORE | there's something here |
| CHALLENGE | I'm not convinced |
| CORE | this feels central |
| SHIFT | this changed the frame |
A snippet comment can be reflective, evaluative, clarifying, skeptical, or directive. "This feels like the crux" and "I'm least convinced by this" are valid comments. You do not need an action verb or next-step directive. The quality bar is: is this a genuine, situated reaction to what the model said?
Use the round prompt for broad comments that don't attach to a specific quote. Use snippets for quote-grounded attention.
Avoid huge quote dumps, generic praise repeated across many snippets, or comments that shift participants away from the problem into platform meta. More guidance: references/snippets.md.
The claim map encodes consensus and contestation as structured reactions on quoted claims. Each claim shows:
Reading discipline:
The claim map is not a verdict. It compresses argumentative structure but loses rhetorical nuance and confidence texture. Use it to navigate; read prose to understand.
Use this test:
Does this note help participants think about the user's problem, or does it mainly report on the platform, session, or process?
Don't use snippet comments to narrate your own moderation process ("I'm using CHALLENGE to redirect the panel"). Don't include platform meta ("this model used the most tokens"). Just react.
Append another round when:
Stop when:
The panel does not need to converge. Sometimes the right output is a clear map of why the decision remains contested — that's actionable for the user, even without a verdict.
| Need | Tool |
|---|---|
| Start a session | create_deliberation |
| Wait for model responses | wait_for_round |
| Add a follow-up round | append_round |
| Recover/read full state | get_session |
| Find prior sessions | list_sessions |
| Confirm model IDs | list_models |
| Check wallet balance | get_credit |
In OpenClaw's tool registry these surface as mumo__create_deliberation, mumo__wait_for_round, etc. (double underscore between server and tool name).
If the user names specific models, call list_models first. Otherwise omit models and let mumo select the panel. More on model selection: references/model-selection.md.
Pass moderator_name on create_deliberation set to your own model identity (e.g., the model OpenClaw is currently running as — visible in your status line, often openai/gpt-5.4-nano, openai/gpt-5.5, or whichever provider is configured). The audit trail should reflect who's actually steering. Don't use the user's name; their identity is already on the session.
Load at most one playbook when it clearly fits:
| Playbook | When |
|---|---|
contested-decision | choosing between options with real tradeoffs |
design-review | reviewing a proposed system, API, plan, or code shape |
uncertainty-expansion | exploring unknowns, stress-testing assumptions |
red-team | finding failure modes, abuse cases, or launch risks |
If none clearly fits, use this kernel only.
These are defaults. If the user prefers more autonomy (e.g., "don't ask before appending" or "always use GPT-5.5 and Gemini"), follow their preferences over this guidance.
Share your panel read with the user, and align on whether to append a round.
If the panel converged, state the consensus and the reasoning behind it. If it split, present the competing positions and what drives each — then offer your own assessment, clearly marked as yours. If the session was exploratory, organize the threads by importance rather than forcing a recommendation.
Synthesize for the user; don't dump the transcript. Mumo produces structure (claim map + snippets); your job is to translate that into "here's what the panel thinks and what to do next."
More at references/synthesis.md.
Mumo surfaces fault lines and supporting arguments; it does not produce the decision. The decision belongs to you (the agent) and the user. When acting on a deliberation:
If responses include claim_confidence or snippets[].comment_confidence, these are self-reported and not calibrated across models. Surface the confidence_disclaimer string if displaying scores.
references/claim-maps.mdreferences/snippets.mdreferences/model-selection.mdreferences/synthesis.mdreferences/operating-notes.md