Install
openclaw skills install agent-architecture-guideBuild a more reliable OpenClaw agent with battle-tested architecture patterns. Covers WAL protocol, working buffer, memory anti-poisoning, layered memory com...
openclaw skills install agent-architecture-guidePractical patterns for building reliable OpenClaw agents.
Every pattern here solved a real problem in a production agent. They are strong defaults, not laws of nature.
For automated diagnostics based on these patterns, see the companion skill: agent-health-optimizer.
Source: Adapted from proactive-agent by halthelobster
Problem: User corrects you, you acknowledge, context resets, correction is lost.
Solution: Write to file BEFORE responding.
Trigger on inbound messages containing:
Protocol: STOP → WRITE (to memory file) → THEN respond.
Source: Adapted from proactive-agent by halthelobster
Problem: Context gets compressed. Recent conversation lost.
Solution: When context >60%, log every exchange to memory/working-buffer.md.
session_statusProblem: External content injects behavioral rules into persistent memory.
Rules:
(source: X, YYYY-MM-DD) to non-obvious factsSource: thoth-ix on Moltbook openclaw-explorers
Problem: Many agents fire bursty recurring cron at :00/:30 → API rate limit stampede.
Solution: Add stagger selectively to recurring jobs that do not need exact timing.
openclaw cron edit <id> --stagger 2m
Use stagger for: recurring polling, feed scans, periodic health checks, broad monitoring.
Avoid blind stagger for: exact-time reminders, scheduled restarts, market-open actions, or anything intentionally pinned to a precise wall-clock time.
Problem: Cron job has --announce and some other path forwards the same result → duplicate user messages.
Solution: pick one primary delivery path.
--announce--no-deliver and let the main agent forward onceThis is not about one universal default; it is about avoiding two send paths for the same event.
Insight from proactive-agent
| Type | Use When |
|---|---|
isolated agentTurn | Background work that must execute, or work that should survive main-session context drift |
main systemEvent | Interactive prompts needing conversation context or heartbeat context |
If the task must happen reliably and independently, prefer isolated.
Problem: Installing skills wholesale overrides your SOUL.md, AGENTS.md, onboarding.
Solution:
Example: From proactive-agent, take WAL + Working Buffer + Resourcefulness. Skip template-heavy onboarding if it conflicts with your existing workspace.
Problem: Many skills have 0 stars, are unmaintained, or overlap with better options.
Solution: Check stats before installing:
curl -s "https://clawhub.ai/api/v1/skills/SLUG" | python3 -c "
import sys,json
d=json.load(sys.stdin)['skill']
s=d.get('stats',{})
print(f'Stars:{s[\"stars\"]} Downloads:{s[\"downloads\"]} Installs:{s[\"installsCurrent\"]}')
"
Browse full catalog:
curl -s "https://clawhub.ai/api/v1/skills?sort=stars&limit=50"
curl -s "https://clawhub.ai/api/v1/skills?sort=trending&limit=30"
Community signals help, but do not replace judgment about fit.
Source: pinchy_mcpinchface on Moltbook (60% token reduction reported)
Problem: 5 separate cron jobs for periodic checks.
Solution: One heartbeat checking all 5. Token cost of 1 turn vs 5 isolated sessions.
Use cron for: exact timing, session isolation, different model Use heartbeat for: batched checks, needs conversation context, timing can drift
Source: proactive-agent by halthelobster
When something fails:
Problem: Agent wakes up fresh each session, doesn't know what skills/tools are installed. Tries which or npm list instead of checking workspace.
Solution: Maintain a categorized skill inventory in TOOLS.md.
Rules:
Suggested lookup priority:
skills/ directorymemory/ files for prior usagewhich, npm list, etc.) as a fallbackWhen you solve a problem, write down:
Add to AGENTS.md or MEMORY.md. Future sessions won't repeat the mistake.
Source: Inspired by TAMS project (18x compression, 97.8% recall) — adapted for OpenClaw's file-based memory.
Problem: MEMORY.md grows indefinitely. Old entries waste tokens every session load, but deleting them loses information.
Solution: Three-layer architecture with time-based compression and index pointers.
Layer 0: memory/YYYY-MM-DD.md ← Raw daily logs, never delete (source of truth)
Layer 1: MEMORY.md ← Active memory (recent 2 weeks: detailed)
Layer 2: memory/archive-YYYY-MM.md ← Monthly archive (highly compressed + index)
Monthly archive flow (run at start of each month):
memory/archive-YYYY-MM.mdCompression rules (general, scene-independent):
Decide compression level by information attributes, NOT by "what I think the user cares about":
| Dimension | Keep in full | Compress to one line | Index only |
|---|---|---|---|
| Reproducibility cost | Can't re-find (personal decisions, private conversation context) | Findable but effort-heavy (paper-specific data points) | Easily searchable (public product names, version numbers) |
| Information type | Actionable decisions / lessons / preferences | Specific numbers / names / dates (keep key identifiers) | Step-by-step procedures / process descriptions |
| Time decay | <2 weeks: keep as-is | 2 weeks – 2 months: refine + index | >2 months: into monthly archive |
Key principles:
Recall test method:
1. Pick 20 random facts from raw daily logs (cover all info types)
2. Try to answer each using ONLY MEMORY.md + archive files
3. Score: ✅ direct hit / ⚠️ partial (has index) / ❌ lost
4. If <80% direct hit: identify which compression rule was violated, fix, re-test
5. If any ❌ with no index pointer: compression was destructive — restore and re-compress
Tested results (real data, 40-question benchmark):
Complements Pattern #13. Compression handles proactive recall; vector search handles reactive retrieval.
Problem: Compressed memory achieves strong direct recall, but some queries still require pointer-tracing back to raw daily logs. Also, memory_search without an embedding provider only does keyword matching.
Solution: Configure OpenClaw's built-in vector search with a lightweight embedding provider. This indexes all memory layers and enables semantic retrieval across the whole history.
Setup (no self-hosted infra required):
# 1. Get a Gemini API key from https://aistudio.google.com/apikey
# 2. Configure OpenClaw
openclaw config set agents.defaults.memorySearch.provider gemini
openclaw config set agents.defaults.memorySearch.remote.apiKey "YOUR_GEMINI_API_KEY"
# 3. Restart gateway and force reindex
openclaw gateway restart
openclaw memory index --force
# 4. Verify
openclaw memory status --deep
Alternative providers:
OPENAI_API_KEY → auto-detectedVOYAGE_API_KEY → good for code-heavy memoryMISTRAL_API_KEY → lightweight alternativeollama → local optionHow it integrates with layered compression:
Query: "白萝卜英文怎么说"
Without vector search:
MEMORY.md → index pointer → manual read daily log
With vector search:
memory_search → hits daily log directly with full context
Also hits archive + MEMORY.md for cross-reference
All three layers get indexed:
MEMORY.md (L1)memory/archive-*.md (L2)memory/YYYY-MM-DD.md (L0)Result: Compression covers the frequently accessed 80-90%; vector search catches the long tail without manual pointer-tracing.
Problem: Short Chinese/Japanese/Korean queries (≤4 characters) consistently miss in vector search. Embedding models encode short CJK text poorly — cosine similarity falls below threshold even when the chunk exists.
Root cause (verified): The chunk is in the index, but similarity scores land at 0.22-0.25 vs a 0.3 minScore threshold. This is a fundamental embedding model limitation, not an indexing bug.
Solution: Expand short CJK queries before calling memory_search using pattern-based rewriting.
| Original pattern | Expand to | Example |
|---|---|---|
| "X了吗" / "X过吗" | Remove particles, search X itself | "装了吗" → "安装 配置 setup" |
| "怎么Y" | Y + method/flow/steps | "怎么部署" → "部署 流程 步骤" |
| "X叫什么" / "X英文" | X + English name | "豆腐英文" → "豆腐 tofu English name" |
| "为什么X" | X + reason | "为什么失败" → "失败 原因 error reason" |
| Pure CJK ≤3 chars | Add English synonym or context | "日志" → "日志 log file 记录" |
| "X停了吗" | X + stopped/paused/status | "服务停了吗" → "service 停止 status 状态" |
Execution: Not a tool modification — the agent expands the query string before calling memory_search. If expanded query still misses, retry with original (double attempt).
Measured impact: Queries like "怎么重启" went from miss (0 results) to direct hit (score 0.67) after combining with Pattern #16 (Ops Index).
Problem: Operational knowledge (restart flows, channel routing, tool configs) is scattered across daily logs, correction logs, and MEMORY.md. Hard to retrieve because the same fact exists in fragments across multiple files.
Solution: Create a single docs/ops-index.md that consolidates operational knowledge with search-friendly aliases.
Structure:
# Operational Index
## Gateway Restart Flow
<!-- aliases: restart, how to restart, restart steps -->
1. Update NOW.md
2. Send notification + set recovery cron
3. Restart → verify exit code
## Discord Channel Routing
<!-- aliases: which channel, message routing -->
| Content | Target | Channel ID |
|---------|--------|------------|
| Stocks | #stocks | 123... |
Key design decisions:
<!-- aliases: ... --> gets indexed by both FTS5 and vector searchMeasured impact: Ops/Config category went from ~60% to 83% recall rate.
Problem: User asks in Chinese, content is stored in English (or vice versa). Embedding models handle cross-language semantic matching poorly for short phrases.
Solution: When writing daily logs, always include both languages inline for any fact that bridges Chinese and English.
✅ 豆腐 (tofu) — firm tofu works best for stir-fry
✅ Docker 部署 (deployment) — port 8080, nginx reverse proxy
✅ 温度设置 (temperature setting) 定时调节 — schedule via app
❌ 豆腐 — 炒菜用老豆腐(missing English)
❌ Deployed Docker container(missing Chinese 部署)
Principle: User asks in Chinese → content might be in English. User searches English → content might be in Chinese. Bilingual anchors make both directions work.
Cost: Zero. It's a writing habit, not infrastructure.
Problem: Same entity has multiple names across languages and contexts (MU = Micron = 美光, 白萝卜 = daikon, 鹅鸭杀 = Goose Goose Duck). Search only finds one form.
Solution: Maintain memory/entities.json mapping canonical names to all known aliases.
{
"tools": {
"Docker": ["容器", "docker-compose", "container"],
"Nginx": ["反向代理", "reverse proxy", "web server"]
},
"food": {
"tofu": ["豆腐", "bean curd", "firm tofu"]
},
"concepts": {
"deployment": ["部署", "上线", "deploy", "release"]
}
}
Usage: When a search query contains a known alias, also search the canonical form (and vice versa). The registry itself doesn't need to be indexed — the agent reads it at query time.
Problem: After building a memory benchmark (N queries with known answers), it's tempting to add keywords to source files that directly match the failing queries. This inflates the score without improving the system.
Solution: Strict separation between eval set and optimization targets.
Rules:
Eval set is for observation, not optimization.
If you catch yourself copying a failing query's keywords into the source material — stop. That's overfitting. Find a structural fix instead.
Problem: Agent loads all memory files at session start, burning context tokens on information that's irrelevant to the current task.
Solution: Load only what the task needs. Use memory_search for precision retrieval instead of reading entire files.
| Scenario | Action |
|---|---|
| User asks "how did we do X last time" | memory_search → memory_get specific lines |
| User mentions a ticker/tool/project | memory_search(entity:XXX) |
| Need last 24h context | Read NOW.md highlights section |
| Heartbeat check | Only HEARTBEAT.md + state file |
| Sub-agent / cron task | Zero memory loading unless task explicitly needs it |
Core principle: If memory_search can pull it precisely, don't read the entire file. Every read consumes context — less waste = longer effective conversations.
Built from real production experience. Strong defaults, not dogma.
This work is licensed under CC BY-SA 4.0. You are free to share and adapt, with attribution and same-license requirement.