{"skill":{"slug":"llmrouter","displayName":"Llmrouter","summary":"Intelligent LLM proxy that routes requests to appropriate models based on complexity. Save money by using cheaper models for simple tasks. Tested with Anthropic, OpenAI, Gemini, Kimi/Moonshot, and Ollama.","tags":{"latest":"0.1.1"},"stats":{"comments":2,"downloads":2567,"installsAllTime":8,"installsCurrent":8,"stars":6,"versions":2},"createdAt":1770040968713,"updatedAt":1777524962645},"latestVersion":{"version":"0.1.1","createdAt":1770055010755,"changelog":"llmrouter v0.1.1\n\n- Expanded provider support: now tested with Anthropic, OpenAI, Google Gemini, Kimi/Moonshot, and Ollama.\n- Added provider-agnostic classification: classifier can run locally on Ollama or remotely on Anthropic, OpenAI, Google, or Kimi.\n- Updated configuration instructions and defaults for broader provider compatibility.\n- Improved OpenClaw integration documentation and setup.\n- Minor dependency and environment requirements changes (Ollama now optional; Python 3.10+ and venv use encouraged).\n- No functional code changes—README/metadata/documentation only.","license":null},"metadata":{"os":["darwin","linux"],"systems":null},"owner":{"handle":"alexrudloff","userId":"publishers:alexrudloff","displayName":"alexrudloff","image":"https://avatars.githubusercontent.com/u/1984999?v=4"},"moderation":null}