{"skill":{"slug":"prompt-token-counter","displayName":"Prompt Token Counter","summary":"Count tokens and estimate costs for 300+ LLM models. Primary use: audit OpenClaw workspace token consumption (memory, persona, skills).","tags":{"latest":"1.0.11"},"stats":{"comments":0,"downloads":351,"installsAllTime":2,"installsCurrent":2,"stars":0,"versions":12},"createdAt":1773486819803,"updatedAt":1777526025553},"latestVersion":{"version":"1.0.11","createdAt":1773638267540,"changelog":"**Changelog for version 1.0.11:**\n\n- Updated documentation and CLI usage for consistency: now supports batch mode by default, allowing multiple file paths directly as arguments (no segmentation).\n- Clarified example commands to demonstrate batch mode and simplified file handling syntax.\n- README and usage examples emphasize file-path-first usage and streamlined batch counting.\n- No core logic/code changes; this release is focused on improved documentation and clearer workflow presentation.","license":"MIT-0"},"metadata":null,"owner":{"handle":"zhaobudaoyuema","userId":"publishers:zhaobudaoyuema","displayName":"Zhaobudaoyuema","image":"https://avatars.githubusercontent.com/u/66905526?v=4"},"moderation":null}