{"skill":{"slug":"llama-params-optimizer","displayName":"llama.cpp Parameter Optimizer","summary":"Complete methodology for local LLM performance optimization. Core principle: maximize context while fully covering GPU memory — find the sweet spot where GPU...","tags":{"latest":"3.0.0","llama.cpp":"2.0.0","local-llm":"2.0.0","optimization":"2.0.0","performance":"2.0.0"},"stats":{"comments":0,"downloads":73,"installsAllTime":0,"installsCurrent":0,"stars":0,"versions":9},"createdAt":1777178782014,"updatedAt":1777206714393},"latestVersion":{"version":"3.0.0","createdAt":1777206168654,"changelog":"Fixed 0.0.0.0 references in all example files to use 127.0.0.1","license":"MIT-0"},"metadata":{"os":null,"systems":null},"owner":{"handle":"hoperealize","userId":"s17ad6sxtk65d4y20pyavaxmys83mb6w","displayName":"Hoperealize","image":"https://avatars.githubusercontent.com/u/21057450?v=4"},"moderation":null}