Install
openclaw skills install multi-model-critiqueUse multiple models in a 4-step cycle of drafting, cross-critique, revision, and synthesis to generate higher-quality answers for complex, high-stakes queries.
openclaw skills install multi-model-critiqueUse this skill only for complex tasks. Route multiple models through the same 4-step loop (Plan -> Execute -> Review -> Improve), then run cross-critique and synthesis to produce a higher-quality final answer than any single-model draft.
Enable this skill only when the request explicitly sets complex to true (or equivalent wording such as “this is complex/deep”).
If complex is false, skip this skill and respond with normal single-model behavior.
Collect or confirm these inputs before execution:
complex: boolean flag (must be true)question: user requestmodels: list of ACP agentId values (typically 3)constraints: output format, language, length, deadlines, forbidden assumptionsops: optional runtime controls (timeoutSec, maxRetries, maxRounds, budgetUsd)SKILL.md (this file): orchestration policy, trigger conditions, and execution sequence.references/prompt-templates.md: reusable prompts for draft, critique, revision, and final synthesis (includes scoring rubric usage).references/orchestration-template.md: practical OpenClaw orchestration flow using sessions_spawn, sessions_send, and sessions_history.references/output-schema.md: machine-parseable JSON output schema for final result and per-model scoring.scripts/build_round_prompts.py: utility to generate per-model prompt files for repeated runs.scripts/run_orchestration.py: local helper that builds a run plan JSON (model mapping, round prompts, runtime settings).Spawn one ACP session per model with the same task and constraints.
Per-model requirements:
Plan -> Execute -> Review -> ImproveDraft AnswerUse sessions_spawn with runtime:"acp" and explicit agentId.
Share peer Draft Answer outputs with each model and require structured critique:
Also require ranking of peer drafts with rationale.
Send critique feedback back to each original model and request revision:
Plan -> Execute -> Review -> ImproveChanges from CritiqueRevised AnswerIntegrate revised answers into one user-facing output:
Score each draft on a 1-5 scale:
accuracy: factual correctness and internal consistencycoverage: completeness against user request and constraintsevidence: quality of assumptions and supportactionability: usefulness for concrete decision/actionDefault weighted score:
0.40 * accuracy + 0.25 * coverage + 0.20 * evidence + 0.15 * actionability
Use this score to justify rankings and the final selected direction.
references/prompt-templates.md for canonical prompts.scripts/build_round_prompts.py when you need file-based prompt generation for repeated or batched runs.scripts/run_orchestration.py to generate a deterministic run-plan artifact for reproducible execution.references/orchestration-template.md for concrete OpenClaw tool-call flow.Final AnswerKey Improvements from CritiqueUncertaintiesNext Steps (optional)When machine consumption is needed, return JSON matching references/output-schema.md.
Do not expose private chain-of-thought. Provide concise reasoning summaries only.
timeoutSec: 180 per round per modelmaxRetries: 1 per failed model turnmaxRounds: fixed at 4 (draft, critique, revision, synthesis)budgetUsd: optional hard stop when cost-sensitive