llmfit

v0.2.2

Detect local hardware (RAM, CPU, GPU/VRAM) and recommend the best-fit local LLM models with optimal quantization, speed estimates, and fit scoring.

1· 960·7 current·7 all-time
byAlex Jones@alexsjones

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for alexsjones/llmfit.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "llmfit" (alexsjones/llmfit) from ClawHub.
Skill page: https://clawhub.ai/alexsjones/llmfit
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Required binaries: llmfit
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Canonical install target

openclaw skills install alexsjones/llmfit

ClawHub CLI

Package manager switcher

npx clawhub@latest install llmfit
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Suspicious
medium confidence
Purpose & Capability
Name and description match the runtime instructions: the SKILL.md tells the agent to run the llmfit CLI to detect hardware and produce model recommendations. Required binaries (llmfit) are consistent with that purpose and no unrelated credentials or files are requested.
Instruction Scope
Instructions are narrowly scoped to running llmfit commands (system, recommend) and mapping outputs to local providers (Ollama, vLLM, LM Studio). They do not instruct reading arbitrary system files or exfiltrating secrets. The skill suggests editing openclaw.json to configure models, which is coherent with its goal.
!
Install Mechanism
Install metadata is inconsistent and potentially risky: the SKILL.md / registry lists a Homebrew formula 'AlexsJones/llmfit' (a third‑party tap) and a second install entry that is labeled as 'cargo install llmfit' but is marked kind: 'node' (and registry also lists 'node'). This mismatch (node vs cargo label) is sloppy and prevents clear vetting of the install source. Homebrew taps from unknown owners should be reviewed before use because they install binaries from third parties.
Credentials
The skill requests no environment variables or credentials. That is proportionate for a local hardware-detection and recommendation tool.
Persistence & Privilege
always is false and the skill does not request or auto-modify other skills' configs. It only recommends edits to openclaw.json (user-driven). No privileged or persistent presence is requested by the skill metadata or instructions.
What to consider before installing
This skill appears to do what it says (run the llmfit CLI and recommend models), but the install metadata is inconsistent and the Homebrew tap is a third-party source. Before installing or running anything: 1) ask the maintainer for the upstream source or GitHub repo and a homepage; 2) inspect the Homebrew formula or cargo package code on a trusted repo; 3) prefer installing from an official, traceable release (GitHub releases or crates.io) rather than an anonymous tap; 4) if you must run the binary first, run `llmfit --version` and `llmfit --json system` in a sandbox/container to inspect output; 5) do not provide credentials or elevated privileges to this tool. If you want help vetting the brew formula or package repository, provide the install URL and I can point out risky patterns.

Like a lobster shell, security has layers — review code before you run it.

Runtime requirements

🧠 Clawdis
Binsllmfit

Install

Install llmfit (brew tap AlexsJones/llmfit && brew install llmfit)
Bins: llmfit
brew install AlexsJones/llmfit
Install llmfit (cargo install llmfit)
Bins: llmfit
latestvk979mvm0m95ka4j8fgh5h5z9zs819h3r
960downloads
1stars
2versions
Updated 2mo ago
v0.2.2
MIT-0

llmfit-advisor

Hardware-aware local LLM advisor. Detects your system specs (RAM, CPU, GPU/VRAM) and recommends models that actually fit, with optimal quantization and speed estimates.

When to use (trigger phrases)

Use this skill immediately when the user asks any of:

  • "what local models can I run?"
  • "which LLMs fit my hardware?"
  • "recommend a local model"
  • "what's the best model for my GPU?"
  • "can I run Llama 70B locally?"
  • "configure local models"
  • "set up Ollama models"
  • "what models fit my VRAM?"
  • "help me pick a local model for coding"

Also use this skill when:

  • The user wants to configure models.providers.ollama or models.providers.lmstudio
  • The user mentions running models locally and you need to know what fits
  • A model recommendation is needed and the user has local inference capability (Ollama, vLLM, LM Studio)

Quick start

Detect hardware

llmfit --json system

Returns JSON with CPU, RAM, GPU name, VRAM, multi-GPU info, and whether memory is unified (Apple Silicon).

Get top recommendations

llmfit recommend --json --limit 5

Returns the top 5 models ranked by a composite score (quality, speed, fit, context) with optimal quantization for the detected hardware.

Filter by use case

llmfit recommend --json --use-case coding --limit 3
llmfit recommend --json --use-case reasoning --limit 3
llmfit recommend --json --use-case chat --limit 3

Valid use cases: general, coding, reasoning, chat, multimodal, embedding.

Filter by minimum fit level

llmfit recommend --json --min-fit good --limit 10

Valid fit levels (best to worst): perfect, good, marginal.

Understanding the output

System JSON

{
  "system": {
    "cpu_name": "Apple M2 Max",
    "cpu_cores": 12,
    "total_ram_gb": 32.0,
    "available_ram_gb": 24.5,
    "has_gpu": true,
    "gpu_name": "Apple M2 Max",
    "gpu_vram_gb": 32.0,
    "gpu_count": 1,
    "backend": "Metal",
    "unified_memory": true
  }
}

Recommendation JSON

Each model in the models array includes:

FieldMeaning
nameHuggingFace model ID (e.g. meta-llama/Llama-3.1-8B-Instruct)
providerModel provider (Meta, Alibaba, Google, etc.)
params_bParameter count in billions
scoreComposite score 0–100 (higher is better)
score_componentsBreakdown: quality, speed, fit, context (each 0–100)
fit_levelPerfect, Good, Marginal, or TooTight
run_modeGPU, CPU+GPU Offload, or CPU Only
best_quantOptimal quantization for the hardware (e.g. Q5_K_M, Q4_K_M)
estimated_tpsEstimated tokens per second
memory_required_gbVRAM/RAM needed at this quantization
memory_available_gbAvailable VRAM/RAM detected
utilization_pctHow much of available memory the model uses
use_caseWhat the model is designed for
context_lengthMaximum context window

Fit levels explained

  • Perfect: Model fits comfortably with room to spare. Ideal choice.
  • Good: Model fits but uses most available memory. Will work well.
  • Marginal: Model barely fits. May work but expect slower performance or reduced context.
  • TooTight: Model does not fit. Do not recommend.

Run modes explained

  • GPU: Full GPU inference. Fastest. Model weights loaded entirely into VRAM.
  • CPU+GPU Offload: Some layers on GPU, rest in system RAM. Slower than pure GPU.
  • CPU Only: All inference on CPU using system RAM. Slowest but works without GPU.

Configuring OpenClaw with results

After getting recommendations, configure the user's local model provider.

For Ollama

Map the HuggingFace model name to its Ollama tag. Common mappings:

llmfit nameOllama tag
meta-llama/Llama-3.1-8B-Instructllama3.1:8b
meta-llama/Llama-3.3-70B-Instructllama3.3:70b
Qwen/Qwen2.5-Coder-7B-Instructqwen2.5-coder:7b
Qwen/Qwen2.5-72B-Instructqwen2.5:72b
deepseek-ai/DeepSeek-Coder-V2-Lite-Instructdeepseek-coder-v2:16b
deepseek-ai/DeepSeek-R1-Distill-Qwen-32Bdeepseek-r1:32b
google/gemma-2-9b-itgemma2:9b
mistralai/Mistral-7B-Instruct-v0.3mistral:7b
microsoft/Phi-3-mini-4k-instructphi3:mini
microsoft/Phi-4-mini-instructphi4-mini

Then update openclaw.json:

{
  "models": {
    "providers": {
      "ollama": {
        "models": ["ollama/<ollama-tag>"]
      }
    }
  }
}

And optionally set as default:

{
  "agents": {
    "defaults": {
      "model": {
        "primary": "ollama/<ollama-tag>"
      }
    }
  }
}

For vLLM / LM Studio

Use the HuggingFace model name directly as the model identifier with the appropriate provider prefix (vllm/ or lmstudio/).

Workflow example

When a user asks "what local models can I run?":

  1. Run llmfit --json system to show hardware summary
  2. Run llmfit recommend --json --limit 5 to get top picks
  3. Present the recommendations with scores and fit levels
  4. If the user wants to configure one, map it to the appropriate Ollama/vLLM/LM Studio tag
  5. Offer to update openclaw.json with the chosen model

When a user asks for a specific use case like "recommend a coding model":

  1. Run llmfit recommend --json --use-case coding --limit 3
  2. Present the coding-specific recommendations
  3. Offer to pull via Ollama and configure

Notes

  • llmfit detects NVIDIA GPUs (via nvidia-smi), AMD GPUs (via rocm-smi), and Apple Silicon (unified memory).
  • Multi-GPU setups aggregate VRAM across cards automatically.
  • The best_quant field tells you the optimal quantization — higher quant (Q6_K, Q8_0) means better quality if VRAM allows.
  • Speed estimates (estimated_tps) are approximate and vary by hardware and quantization.
  • Models with fit_level: "TooTight" should never be recommended to users.

Comments

Loading comments...