Julia's OpenClaw Token Optimizer

AdvisoryAudited by Static analysis on May 7, 2026.

Overview

No suspicious patterns detected.

Findings (0)

Artifact-based informational review of SKILL.md, metadata, install specs, static scan signals, and capability signals. ClawScan does not execute the skill or run runtime probes.

What this means

Your agent's default model could be changed persistently, which may affect future answer quality, cost, provider routing, or rate-limit behavior.

Why it was flagged

The skill instructs the agent to patch model-selection configuration, including defaults, but the artifacts do not require explicit user approval, scope the patch to a session/profile, or provide a rollback plan.

Skill content
3. gateway config.schema.lookup 'agents.defaults.modelSelection'. ... 6. gateway config.patch {modelSelection: {primary: 'best-cheap-model'}}.
Recommendation

Use this only with an explicit approve-before-patch workflow: require the agent to show the proposed config diff, expected cost/quality tradeoff, target scope, and rollback command before making changes.

What this means

Benchmark prompts may be processed by additional models and may consume tokens; private or sensitive prompts should not be used as test data unless intentionally approved.

Why it was flagged

Benchmarking through a subagent/model is aligned with the optimizer purpose, but it does involve sending prompts through another agent/model boundary.

Skill content
4. Benchmark: spawn subagent with test prompts on cheap models.
Recommendation

Use synthetic benchmark prompts, disclose which models/providers will be tested, and get user approval before sending any sensitive content.