Install
openclaw skills install live-evoSelf-evolving memory system that learns from verifiable tasks. Use when completing tasks where you can verify the outcome (coding, predictions, analysis). Au...
openclaw skills install live-evoYou are using the Live-Evo memory system that learns from past mistakes through experience accumulation and adaptive evaluation.
IMPORTANT — Script location: All scripts are in the scripts/ subdirectory next to this SKILL.md file. When running scripts, use the absolute path to the scripts/ directory relative to where this file is located. For example, if this SKILL.md is at /path/to/live-evo/SKILL.md, the scripts are at /path/to/live-evo/scripts/.
Experience data is stored persistently at ~/.live-evo/experience_db.jsonl (independent of skill installation location).
Run the experience retrieval script to find relevant past experiences:
python <scripts-dir>/retrieve.py --query "YOUR_TASK_DESCRIPTION"
If experiences are found, they will be compiled into a task-specific guideline. Use this guideline to inform your approach.
You must judge whether contrastive verification (two attempts) is worthwhile based on:
| Factor | Do Contrastive Eval | Skip, Direct Apply |
|---|---|---|
| Cost of re-running | Low (e.g. run a test) | High (e.g. long build, API costs, heavy computation) |
| Verifiability | Clear ground truth exists (tests, known answer) | No easy way to verify programmatically |
| Task complexity | Simple enough to attempt twice | Too complex/large to reasonably duplicate |
| Guideline relevance | Retrieved guideline is highly relevant | Guideline is loosely related or no guideline found |
If contrastive eval IS worthwhile → Go to Step 2A If contrastive eval is NOT worthwhile → Go to Step 2B
Make two independent attempts:
Attempt A (Without Memory):
Attempt B (With Guideline):
Then verify and update weights:
python <scripts-dir>/update.py \
--task "TASK_DESCRIPTION" \
--result-a "RESULT_WITHOUT_MEMORY" \
--result-b "RESULT_WITH_GUIDELINE" \
--correct "CORRECT_ANSWER" \
--experience-ids "id1,id2,..."
When contrastive evaluation is not feasible:
python <scripts-dir>/add_experience.py \
--question "THE_TASK_QUESTION" \
--failure-reason "What went wrong (from feedback)" \
--improvement "Key lesson learned" \
--category "coding|analysis|prediction|debugging|other"
No contrastive comparison needed — just learn from what happened.
Whenever a task fails or feedback reveals a learnable lesson — regardless of which path you took — store the experience:
python <scripts-dir>/add_experience.py \
--question "THE_TASK_QUESTION" \
--failure-reason "What went wrong" \
--improvement "Key lesson learned" \
--category "coding|analysis|prediction|debugging|other"
If you used a retrieved guideline and can determine whether it helped:
python <scripts-dir>/update.py \
--task "TASK_DESCRIPTION" \
--result-a "WHAT_WOULD_HAVE_HAPPENED" \
--result-b "WHAT_ACTUALLY_HAPPENED" \
--correct "CORRECT_OUTCOME" \
--experience-ids "id1,id2,..."
If you cannot determine whether the guideline helped, skip weight updates — no update is better than a wrong update.
Use this system for:
Each experience contains:
question: The original task/questionfailure_reason: What went wrong in the original attemptimprovement: Key lesson or approach that would have helpedmissed_information: Information sources or considerations that were missedweight: Quality score (0.1-2.0) updated based on usefulnesscategory: Domain category for filteringView all experiences:
python <scripts-dir>/list_experiences.py
Search experiences:
python <scripts-dir>/retrieve.py --query "your search query" --top-k 5
Get statistics:
python <scripts-dir>/stats.py