ia-orchestrating-swarms

v3.0.4

Coordinate multi-agent swarms for parallel and pipeline workflows. Use when coordinating multiple agents, running parallel reviews, building pipeline workflo...

0· 265·0 current·0 all-time
byIlia Alshanetsky@iliaal

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for iliaal/compound-eng-orchestrating-swarms.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "ia-orchestrating-swarms" (iliaal/compound-eng-orchestrating-swarms) from ClawHub.
Skill page: https://clawhub.ai/iliaal/compound-eng-orchestrating-swarms
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Bare skill slug

openclaw skills install compound-eng-orchestrating-swarms

ClawHub CLI

Package manager switcher

npx clawhub@latest install compound-eng-orchestrating-swarms
Security Scan
Capability signals
CryptoCan make purchasesRequires OAuth tokenRequires sensitive credentials
These labels describe what authority the skill may exercise. They are separate from suspicious or malicious moderation verdicts.
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
The skill is an orchestration/workflow helper and all required behaviour (spawning teams/agents, task lifecycle, inboxes, use of tmux/iterm/in-process backends) is aligned with that purpose. It does not ask for unrelated credentials, remote URLs, or bizarre binaries.
Instruction Scope
SKILL.md and references explicitly instruct reading and writing agent runtime files (e.g., ~/.claude/teams/*, ~/.claude/tasks/*), running local commands (cat, grep, jq, tail, tmux, etc.), and installing helpers (it2) if needed. These actions are within the orchestrator's legitimate scope but do grant the skill access to local agent configuration and repo files — review and accept that local-file access is required before use.
Install Mechanism
This is an instruction-only skill with no install spec and no automatic downloads or extracted archives. That minimizes on-disk installation risk.
Credentials
The skill declares no required environment variables or credentials. It documents runtime environment variables (CLAUDE_CODE_*) that teammates receive; those are reasonable for agent orchestration and not secrets. No unrelated service keys or high-risk env vars are requested.
Persistence & Privilege
always:false (default). The skill does not request permanent platform presence or modifications to other skills' configurations. Autonomous invocation is allowed by default but is typical for skills of this class.
Assessment
This skill appears coherent for coordinating multi-agent swarms. Before installing, consider: (1) it includes many snippets that read and manipulate local agent runtime files under your home directory (~/.claude/*) and run shell commands (cat, grep, jq, tmux, tail). That is expected for an orchestrator but means it can access local config and task data — only install if you trust it with that. (2) It does not request external credentials or download code, reducing supply-chain risk. (3) If you plan to run agents that modify repository files, back up or use a disposable repo/worktree to avoid accidental changes; follow the skill's stated one-owner-per-file rule. (4) If you do not want visible panes or persistent subprocesses, force in-process backend or avoid tmux/iterm2 usage. (5) Review any prompts the skill will use to spawn subagents (they can execute tools) — avoid giving it sensitive tokens or secrets in task prompts. Overall: coherent and expected for an orchestrator, but verify you accept local file and shell access before using it.

Like a lobster shell, security has layers — review code before you run it.

latestvk970c27ngysnf5c3qtrt13p48x85nspn
265downloads
0stars
11versions
Updated 17h ago
v3.0.4
MIT-0

Swarm orchestration

Primitives

Agents, teams, teammates, leaders, tasks, inboxes, messages, backends — see primitives.md for definitions and the file-system layout.


Two Ways to Spawn Agents

AspectTask (subagent)Task + team_name + name (teammate)
LifespanUntil task completeUntil shutdown requested
CommunicationReturn valueInbox messages
Task accessNoneShared task list
Team membershipNoYes
CoordinationOne-offOngoing
Best forSearches, analysis, focused workParallel work, pipelines, collaboration

Subagent (short-lived, returns result):

Task({ subagent_type: "Explore", description: "Find auth files", prompt: "..." })

Teammate (persistent, communicates via inbox):

Teammate({ operation: "spawnTeam", team_name: "my-project" })
Task({ team_name: "my-project", name: "worker", subagent_type: "general-purpose",
       prompt: "...", run_in_background: true })

For detailed agent type descriptions, see agent-types.md.

Parallel Fan-Out (for independent work)

When dispatching multiple read-only or worktree-isolated agents whose work is independent, issue all Task calls in a SINGLE assistant message. Sequential dispatch across separate messages serializes what should run concurrently. Opus 4.7 does not parallelize by default -- state it explicitly.

// Correct: one message, multiple Task tool uses
Task({ subagent_type: "security-sentinel", ... })
Task({ subagent_type: "performance-oracle", ... })
Task({ subagent_type: "architecture-strategist", ... })

Sequential dispatch (each Task in its own message, waiting on the previous to return) is a serialization bug, not a coordination pattern. If agents truly depend on each other's output, that is a pipeline -- see Coordination Models below.


Quick Reference

For copy-paste spawn/message/task/shutdown snippets, load quick-reference.md.


Dispatch Discipline

Rules for when and how to dispatch agents. Getting these wrong wastes tokens and creates hard-to-debug failures.

When to dispatch a team vs. do it yourself:

Assess 5 signals: file count, module span, dependency chain, risk surface, parallelism potential. If 3+ fall in the "complex" column, dispatch a team. Below 3, do it yourself. When in doubt, prefer the simple path -- team overhead is only justified when parallelism provides a real speedup.

Task description template (for every dispatched task):

Every task prompt must include these fields to prevent integration failures:

  • Objective: what to accomplish (one sentence)
  • Owned Files: files this agent creates or modifies (exclusive -- no file assigned to multiple agents)
  • Interface Contracts: what to import from other agents' work, what to export for downstream agents
  • Acceptance Criteria: how the agent knows the task is correct
  • Out of Scope: what NOT to touch, even if it looks related

Cardinal rule: one owner per file. When files must be shared, designate a single owner; other agents send change requests, owner applies sequentially. If an upstream dependency isn't ready yet, write a stub/mock so downstream work can continue unblocked.

No parallel implementation agents (without worktrees):

Implementation agents share state via git by default, so parallel dispatch causes overwrites. Use isolation: "worktree" to give each agent its own copy. Without worktrees, dispatch implementation agents sequentially. Review, research, and analysis agents are always safe to parallelize (read-only).

Pre-dispatch file-intersection check -- operationalize the one-owner-per-file rule with a runnable safety gate before every parallel dispatch:

  1. Collect each unit's declared Owned Files / Test Paths / Modify Paths from its task spec.
  2. Build a {file → unit} map. If any file appears under more than one unit, the dispatch is unsafe. Quick check on Markdown task specs:
    grep -h "^Owned Files:" -A 20 tasks/*.md | grep -v "^Owned Files:" | grep -v "^--$" | sort | uniq -d
    
    Any output is an overlapping file path that needs resolution.
  3. On overlap: either downgrade to serial (log the overlap and the reason), or assign worktree isolation (isolation: "worktree" per agent), or rewrite unit boundaries so files become exclusive.
  4. Even with no declared overlap, include this constraint verbatim in every parallel-dispatch prompt: "Do not run git add, git commit, or the project's test suite while other parallel agents are active -- you'd race on the git index or thrash the test cache. Stage changes for the orchestrator to commit after integration."

The intersection check catches silent conflicts the controller misses at plan time; the dispatch-prompt constraint catches them when a unit's file list was incomplete.

Preset team compositions: Start from a named preset before designing a custom team. See team-compositions.md for the full table (Review / Debug / Feature / Fullstack / Migration / Security / Research), the cardinal subagent_type rule (read-only agents cannot implement), and custom-team guidelines. Use the smallest preset that covers all required dimensions — overlap between reviewers is a sizing signal to redefine focus areas, not add more agents.

Model selection by task complexity:

Task shapeModel
1-2 files, clear spec, mechanicalmodel: "haiku"
Multi-file integration, standard complexityDefault model
Architecture decisions, ambiguous scope, reviewmodel: "opus"

Handoff protocol -- structured agent-to-agent transfers:

When passing work between agents (leader→implementer, implementer→reviewer, reviewer→leader), include:

  1. Context: what was done, relevant files, constraints discovered
  2. Deliverable: specific output expected from the receiving agent
  3. Acceptance criteria: how the receiving agent knows the work is correct

The controller reads all tasks from the plan upfront and provides full task text directly to subagents. Never make subagents read plan files themselves -- they waste tokens navigating, may read different versions, and inherit unclear context. Paste the task content into the prompt. See handoff-templates.md for QA FAIL and Escalation Report formats.

Standardize implementer status signals:

Include the four statuses defined in ia-verification-before-completion (DONE, DONE_WITH_CONCERNS, BLOCKED, NEEDS_CONTEXT) in every teammate prompt so they know the reporting format. Expect teammates to report one. BLOCKED responses get further triage via the decision tree below.

BLOCKED triage decision tree -- when a teammate reports BLOCKED, classify the root cause before acting. Never retry the same prompt on the same model without changing a variable.

Root causeSignalResponse
Missing contextAgent asked for a file, spec, or decision it neededProvide the missing context, re-dispatch same agent
Reasoning ceilingAgent attempted, got stuck on a subtlety it cannot resolveEscalate model (haiku → sonnet → opus) and re-dispatch
Task too largeAgent made partial progress but hit token/complexity limitsSplit into smaller tasks with explicit interface contracts
Spec wrongAgent surfaces a contradiction in the plan or a missing requirementEscalate to the user -- do not re-dispatch

Never ignore an escalation. Never force the same agent to retry without changing at least one variable (context, model, or task scope).

Two-stage review gate on subagent outputs:

Verify spec compliance first: does the output match what was requested? Only then evaluate quality. A beautifully written solution to the wrong problem is still wrong. Structure review as two explicit passes -- pass 1 rejects on spec mismatch without reading further, pass 2 assesses correctness and quality on spec-compliant outputs.

QA retry loop:

Max 3 attempts per task. After each QA failure, pass structured feedback to the implementer using the QA FAIL template. After 3 failures, mark the task as blocked, continue the pipeline (don't halt everything), and let final integration catch remaining issues. Counter resets when advancing to the next task.


Integration Rules

Post-integration verification -- after all agents return: check overlapping file edits, review for conflicting approaches, run full test suite.

Spawned-session behavior -- when a skill runs inside an orchestrated pipeline (as a subagent, not user-invoked), suppress interactive prompts: do not use AskUserQuestion, auto-choose the conservative/safe default, skip upgrade checks and telemetry. Focus on completing the task and reporting results via prose output. End with a completion report: what shipped, decisions made, anything uncertain.


Context Carry-Forward

After each turn, five strategies exist for moving context forward: Continue, Rewind, /compact, Subagent, /clear+brief. Choose deliberately — the default "Continue" is rarely best, and Rewind is strictly better than "correcting in place" after a failed attempt. See context-carry-forward.md for the full decision table and rationale.

Coordination Models

Two approaches to multi-agent coordination exist. Choose based on the work pattern:

AspectStateless (copy-paste outputs)Stateful (file ownership + dependencies)
How agents share stateLeader copies full outputs between promptsAgents read/write shared task files, claim ownership
Best forShort pipelines, 2-3 agents, sequential handoffsParallel work, 4+ agents, complex dependency graphs
Failure modeContext grows linearly with agent countConcurrent modification conflicts
MitigationSummarize before passing (keep essentials, drop navigation)Use worktrees or exclusive file ownership per agent

For most work, start with stateless handoffs. Graduate to stateful coordination only when parallelism provides a real speedup and you have worktree isolation to prevent file conflicts.


Dispatch Anti-Patterns

Before designing any multi-agent workflow, check it against the four named failure modes in dispatch-anti-patterns.md: router persona, persona calls persona, sequential paraphraser, deep persona trees. Rule of thumb: if the proposed swarm has more coordinator roles than worker roles, collapse it.

Anti-Sycophancy and Resilience

When dispatching judge panels, running parallel reviewers, or iterating on subjective evaluations, load anti-sycophancy.md — cold-start isolation, fresh instances per round, label randomization, convergence detection.

When designing multi-agent workflows that must survive partial failure, load resilience-patterns.md — cascade prevention (timeouts, circuit breakers, bulkheads), failure classification (retry vs reassign vs escalate), mid-pipeline compensation for irreversible side effects, post-failure synthesis of partial results.

Verify

  • All tasks in terminal state (completed or blocked)
  • No orphaned teammates (git worktree list shows no stale entries)
  • Overlapping file edits reviewed and merged
  • Full test suite passes post-integration

References

DocumentWhen to loadWhat it covers
team-compositions.mdSizing a team or choosing a preset7 preset compositions, subagent_type cardinal rule, custom-team guidelines
agent-types.mdChoosing which agent to spawnBuilt-in and plugin agent types with examples
teammate-operations.mdUsing TeammateTool for persistent agentsAll 13 operations (spawnTeam, write, broadcast, requestShutdown, etc.)
task-system.mdManaging work items and dependenciesTaskCreate, TaskList, TaskGet, TaskUpdate, file structure
message-formats.mdSending structured messages between agentsAll JSON message examples (regular, shutdown, idle, plan approval)
orchestration-patterns.mdDesigning a multi-agent workflow6 patterns + 3 complete workflow examples
spawn-backends.mdTroubleshooting agent spawn issuesBackend comparison, auto-detection, in-process/tmux/iterm2
environment-config.mdConfiguring team environmentEnvironment variables and team config structure
handoff-templates.mdPassing work between agentsQA FAIL and Escalation Report formats
context-carry-forward.mdLong sessions with orchestrated subagentsContinue / Rewind / compact / Subagent / clear+brief decision table
anti-sycophancy.mdJudge panels, parallel reviewers, subjective evalsCold-start isolation, fresh instances per round, label randomization, convergence detection
resilience-patterns.mdDesigning workflows that survive partial failureCascade prevention, failure classification, mid-pipeline compensation, post-failure synthesis

Comments

Loading comments...