Install
openclaw skills install llm-as-judgeCross-model verification for complex tasks. Spawn a judge subagent with a different model to review plans, code, architecture, or decisions before execution. Use when working on "architecture", "system design", "complex feature", "security review", "production deployment", financial/trading systems, or when stuck after 3+ attempts. NOT for simple edits, config changes, or routine tasks.
openclaw skills install llm-as-judgeCore principle: Same model = same blind spots. Different model = fresh perspective. Cross-model review catches ~85% of issues vs ~60% for self-reflection.
Use this pattern when:
Skip when:
Executor (Model A) → Output → Judge (Model B) → Verdict → Action
Verdicts: APPROVE | REVISE (with specific feedback) | REJECT (restart)
Use a different provider than the executor to avoid shared blind spots:
kimi or grok or gemini-proopusSee references/judge-prompts.md for full templates covering:
Simple task? → Self-review
Complex / high stakes? → LLM-as-Judge
Stuck after retries? → LLM-as-Judge (fresh perspective)
Financial/security? → LLM-as-Judge (mandatory)