OnDeckLLM
v1.4.3Localhost dashboard for managing LLM providers, model routing, and batting-order fallback chains. Auto-discovers providers from OpenClaw config or works stan...
Like a lobster shell, security has layers — review code before you run it.
License
SKILL.md
OnDeckLLM — AI Model Lineup Manager
Prerequisites
npm install -g ondeckllm
Verify: ondeckllm --help or check install with npm list -g ondeckllm.
What It Does
OnDeckLLM is a localhost web dashboard that:
- Auto-discovers LLM providers from OpenClaw config (
~/.openclaw/openclaw.json) - Manages a batting-order priority list for model routing (primary + fallbacks)
- Tests provider health and latency
- Syncs model lineup back to OpenClaw config with one click
- Tracks session costs (JSONL usage log + Chart.js)
- Supports Anthropic, OpenAI, Google AI, Groq, xAI/Grok, Ollama (local + remote), Mistral, DeepSeek, Together, OpenRouter
Starting the Dashboard
# Default port 3900
ondeckllm
# Custom port
PORT=3901 ondeckllm
The dashboard runs at http://localhost:3900 (or custom port).
As a Background Service
Use the helper script to check status or start OnDeckLLM:
node scripts/status.js
Output: JSON with running (bool), port, url, and pid if active.
Agent Workflow
Check if OnDeckLLM is running
node scripts/status.js
Open the dashboard for the user
Direct them to http://localhost:3900 (or the configured port/URL).
Provider management
OnDeckLLM reads provider config from ~/.openclaw/openclaw.json automatically.
Changes made in the dashboard sync back to OpenClaw config.
No separate API or CLI commands needed — it's a web UI tool.
Configuration
OnDeckLLM stores its data in ~/.ondeckllm/:
config.json— provider settings, port, Ollama URLusage.jsonl— cost tracking logprofiles/— saved batting-order profiles
Remote Ollama
To connect to a remote Ollama instance, configure in the dashboard UI:
Settings → Ollama → Remote URL (e.g., http://192.168.55.80:11434)
Links
- 🌐 ondeckllm.com
- 📦 npm
- 🐛 GitHub Issues
Files
2 totalComments
Loading comments…
