Install
openclaw skills install clawd-throttleRoutes LLM requests to the cheapest capable model across 8 providers (Anthropic, Google, OpenAI, DeepSeek, xAI, Moonshot, Mistral, Ollama) and 25+ models. Scores prompts on 8 dimensions in under 1ms, supports three routing modes (eco, standard, gigachad), and logs all decisions for cost tracking.
openclaw skills install clawd-throttleRoute every LLM request to the cheapest model that can handle it. Stop paying Opus prices for "hello" and "summarize this."
Supports 8 providers and 25+ models: Anthropic (Claude), Google (Gemini), OpenAI (GPT / o-series), xAI (Grok), DeepSeek, Moonshot (Kimi), Mistral, and Ollama (local).
| Mode | Simple | Standard | Complex |
|---|---|---|---|
| eco | Grok 4.1 Fast | Gemini Flash | Haiku |
| standard | Grok 4.1 Fast | Haiku | Sonnet |
| gigachad | Haiku | Sonnet | Opus 4.6 |
Each cell shows the first-choice model. The router tries a preference list and falls through to the next available provider if the first is not configured.
| Command | What It Does |
|---|---|
route_request | Send a prompt and get a response from the cheapest capable model |
classify_prompt | Analyze prompt complexity without making an LLM call |
get_routing_stats | View cost savings and model distribution stats |
get_config | View current configuration (keys redacted) |
set_mode | Change routing mode at runtime |
get_recent_routing_log | Inspect recent routing decisions |
/opus, /sonnet, /haiku, /flash, or /grok-fast to force a specific modelnpm run setup