Agent Cost Monitor

v1.1.1

Monitor token usage and estimated costs across all OpenClaw agents in real-time, with alerts and recommendations to stay within budgets and optimize spending.

0· 1.3k·8 current·8 all-time
byNealbuilds@neal-collab

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for neal-collab/agent-cost-monitor.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "Agent Cost Monitor" (neal-collab/agent-cost-monitor) from ClawHub.
Skill page: https://clawhub.ai/neal-collab/agent-cost-monitor
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Canonical install target

openclaw skills install neal-collab/agent-cost-monitor

ClawHub CLI

Package manager switcher

npx clawhub@latest install agent-cost-monitor
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
Name/description (cost monitoring across agents) match the instructions: it reads session_status/sessions_list, computes per-agent token/cost metrics, compares to budgets, and issues alerts/recommendations. No unrelated binaries, environment variables, or install steps are requested.
Instruction Scope
The SKILL.md stays within monitoring and recommendation tasks. It explicitly uses platform-internal tools (session_status, sessions_list) which is coherent. Note: the document contains example JSON snippets that change session/reset/cache/model parameters and recommends moving agents between model tiers — these are configuration-change suggestions. The skill does not include explicit commands that will automatically modify other agents, but an agent with write privileges could apply those snippets; review platform permissions before allowing autonomous writes.
Install Mechanism
Instruction-only skill with no install spec and no code files; nothing written to disk and no external packages downloaded.
Credentials
No environment variables, secrets, or external API keys are required. The SKILL.md mentions alert channels (e.g., Telegram DM) as an example but does not request corresponding tokens — if you configure such channels, you must supply credentials separately. No unexpected credentials are requested by the skill itself.
Persistence & Privilege
always is false (not forced into every agent run). The skill can be invoked autonomously by agents (default) which is normal for skills; combined with the configuration-change examples, consider restricting write permissions if you do not want automated configuration changes across agents.
Assessment
This skill appears coherent and low-risk as-is: it only describes reading agent session state and computing estimates and recommendations. Before enabling it, check two things: (1) what platform permissions an agent running this skill will have — if it can write agent configs, it could apply the example changes (cacheRetention, session resets, model swaps) automatically; restrict write permissions if you only want reports, and (2) if you configure external alert channels (Telegram, webhooks, etc.), provide tokens only to an agent you trust. Run the skill in an isolated/non-production agent first to confirm behavior and outputs.

Like a lobster shell, security has layers — review code before you run it.

latestvk970jv7qvs6h93appvwedmx791814bpf
1.3kdownloads
0stars
3versions
Updated 2mo ago
v1.1.1
MIT-0

Agent Cost Monitor — Know What Your Agents Cost

Track token usage, costs, and efficiency across all your OpenClaw agents in real-time. Get alerts before you blow your budget.

The Problem

Running multiple agents is powerful — but expensive if you're not watching:

  • Which agent is burning the most tokens?
  • Are heartbeats wasting money on expensive models?
  • Is caching actually saving you anything?
  • When will you hit your weekly rate limit?

What This Skill Does

When triggered (via cron or manually), the agent:

  1. Checks session_status for each agent
  2. Calculates per-agent and total costs
  3. Compares against budget thresholds
  4. Sends alerts if limits are approaching
  5. Suggests optimization moves

Usage

Ask your monitoring agent (or any agent with this skill):

"Give me a cost report for all agents"
"Which agent used the most tokens today?"
"Am I going to hit my rate limit this week?"

Automated Daily Report (Cron)

{
  "name": "Daily Cost Report",
  "schedule": { "kind": "cron", "expr": "0 20 * * *", "tz": "Europe/Berlin" },
  "payload": {
    "kind": "agentTurn",
    "message": "Run a cost report across all agents. Check session_status for each. Report: total tokens, cost per agent, top spender, budget warnings. Send summary to user."
  },
  "sessionTarget": "isolated",
  "delivery": { "mode": "announce" }
}

Cost Report Format

When generating a report, use this structure:

## 💰 Agent Cost Report — [Date]

### Per-Agent Breakdown
| Agent | Model | Tokens (24h) | Est. Cost | Status |
|-------|-------|-------------|-----------|--------|
| Central | Opus 4.6 | 125K | $1.87 | ⚠️ High |
| Techops | Opus 4.6 | 89K | $1.33 | ✅ Normal |
| Atlas | Sonnet 4.5 | 45K | $0.27 | ✅ Low |
| Closer | Haiku 4.5 | 23K | $0.02 | ✅ Minimal |
| Heartbeats | Ollama | 12K | $0.00 | ✅ Free |

### Summary
- **Total 24h:** 294K tokens (~$3.49)
- **Projected weekly:** ~$24.43
- **Budget:** $20/week → ⚠️ 122% projected

### Recommendations
1. Move Techops from Opus → Sonnet for routine tasks (-40% cost)
2. Increase heartbeat interval from 15m → 30m
3. Enable context pruning on Atlas (idle sessions burning cache)

Model Cost Reference

Use these rates for estimation (as of 2026):

Anthropic (Claude OAuth / API)

ModelInput/1MOutput/1MCache Read/1MCache Write/1M
Opus 4.6$5.00$25.00$0.50$6.25
Sonnet 4.5$3.00$15.00$0.30$3.75
Haiku 4.5$1.00$5.00$0.08$1.25

Free Options

ModelCostUse For
Ollama (local)$0Heartbeats, simple tasks
Gemini OAuth$0*Fallback (rate limited)

*Free tier with rate limits

Optimization Playbook

Quick Wins (Do These First)

  1. Heartbeats on Ollama
{ "heartbeat": { "model": "ollama/llama3.2:3b" } }

Saves: 100% of heartbeat costs (can be $5-10/week with Opus)

  1. Haiku Cache Retention Off
{ "anthropic/claude-haiku-4-5": { "params": { "cacheRetention": "none" } } }

Saves: Cache write costs on cheap model (not worth caching)

  1. Context Pruning
{ "contextPruning": { "mode": "cache-ttl", "ttl": "5m" } }

Saves: Stale context re-reads on every turn

  1. Opus/Sonnet Cache Retention Long
{ "anthropic/claude-opus-4-6": { "params": { "cacheRetention": "long" } } }

Saves: Re-sending system prompt every turn (biggest single saving)

Model Tiering (Biggest Impact)

Task TypeUse ThisNot ThisSaving
Coordination, complex reasoningOpusJustified
Finance, data analysisSonnetOpus-40%
Sales drafts, marketing copyHaikuSonnet-67%
Heartbeats, health checksOllamaAny paid-100%
Tweet draftsHaiku or GrokOpus-80%

Session Management

  • Daily reset: Sessions auto-clear at a set hour (reduces token accumulation)
{ "session": { "reset": { "mode": "daily", "atHour": 4, "idleMinutes": 45 } } }
  • Memory flush: Save important context before compaction
{ "compaction": { "memoryFlush": { "enabled": true } } }

Alert Thresholds

Configure in your monitoring agent's memory:

## Budget Alerts
- Daily budget: $5.00 (warn at 80% = $4.00)
- Weekly budget: $20.00 (warn at 70% = $14.00)
- Per-agent daily max: $2.00
- Alert channel: Telegram DM

Integration with DevOps Agent

If you have a DevOps/monitoring agent (e.g. your DevOps agent), add to its AGENTS.md:

## Cost Monitoring
- Run daily cost report at 20:00
- Alert if any agent exceeds $2/day
- Weekly summary every Monday 09:00
- Track trends: is usage going up or down?

FAQ

Q: Does this skill make API calls? A: No. It uses OpenClaw's built-in session_status tool. No external APIs, no additional costs.

Q: How accurate are cost estimates? A: Based on published model pricing. Actual costs may vary with caching hits. Estimates are conservative (slightly high).

Q: Can I track costs per conversation? A: Not directly. Costs are tracked per session. Use sessions_list to see per-session token counts.

Q: Works with non-Anthropic models? A: Yes. Token counts work for all providers. Cost estimation requires known pricing (add custom rates in the cost reference section).

Changelog

v1.1.0

  • Generalized all agent names in examples
  • No specific setup references

v1.0.0

  • Initial release

Comments

Loading comments...