Lemonade Server Manager

v1.0.7

Manage Lemonade Servers natively. Use when checking system info, health status, listing available models, pulling or loading new models, completing LLM chats...

0· 184·0 current·0 all-time
byJames Martinez@james-martinez

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for james-martinez/lemonade-server-manager.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "Lemonade Server Manager" (james-martinez/lemonade-server-manager) from ClawHub.
Skill page: https://clawhub.ai/james-martinez/lemonade-server-manager
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Bare skill slug

openclaw skills install lemonade-server-manager

ClawHub CLI

Package manager switcher

npx clawhub@latest install lemonade-server-manager
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
Name/description match the declared operations (system info, health, list/pull/load/unload models, chat, image generation). The only runtime requirement (curl) and the optional LEMONADE_API_KEY are appropriate for an HTTP-based server-management skill.
Instruction Scope
SKILL.md confines behavior to making HTTP requests to a base server_url (default http://localhost:8000) and explicitly warns prompts/model requests will be transmitted to that URL. It does not instruct reading arbitrary host files or other env vars. Note: if a remote server_url is supplied, user prompts and model payloads will be sent to that remote endpoint — expected for this skill but important for user privacy.
Install Mechanism
Instruction-only skill with no install spec or code files — lowest install risk. README suggests manual clone or ClawHub install, but nothing is automatically downloaded by the skill itself.
Credentials
Only an optional LEMONADE_API_KEY is declared and used by the documented curl examples. No unrelated credentials, config paths, or high-privilege env vars are requested.
Persistence & Privilege
always:false and user-invocable:true (defaults). The skill can be invoked autonomously by the agent (platform default). Combined with the ability to send user prompts to arbitrary server_url values, this is a privacy consideration but not an incoherence or disproportionate privilege in itself.
Assessment
This skill appears coherent for managing Lemonade servers. Before installing: (1) Be aware that any content you send through the skill will be forwarded to whatever server_url you (or the agent) supplies — avoid sending sensitive data to untrusted remote endpoints. (2) Prefer using local servers (http://localhost:8000) or HTTPS endpoints you control; verify TLS and host identity for remote servers. (3) The API key is optional — only set LEMONADE_API_KEY if you trust the target server. (4) The registry metadata has no homepage while SKILL.md lists a GitHub repo; if provenance matters, manually inspect the repository/owner to confirm trustworthiness before installing or exporting credentials. (5) Because the skill can be invoked by the agent autonomously (platform default), review agent autonomy settings if you want to restrict automatic calls to remote servers.

Like a lobster shell, security has layers — review code before you run it.

Runtime requirements

🍋 Clawdis
OSLinux · macOS · Windows
Any bincurl
latestvk97ejk6p4tffb48k95y0z6k17183dqfp
184downloads
0stars
7versions
Updated 1mo ago
v1.0.7
MIT-0
Linux, macOS, Windows

Lemonade Server Management

Interact with and manage local or remote Lemonade AI Server hardware directly via standard native network requests (curl).

When to Use

  • Checking local GPU/NPU health and currently loaded VRAM resources
  • Listing available, downloaded text/image models on a Lemonade cluster
  • Pulling, loading, or unloading multimodal models
  • Generating text from LLMs (chat completions)
  • Generating stable-diffusion image responses

Setup Instructions

  1. Every endpoint requires a base server_url. If one is not specified by the user, assume http://localhost:8000.
  2. Use the LEMONADE_API_KEY environment variable to authenticate.
  3. This skill stays focused on making HTTP requests to a base server_url (default http://localhost:8000) and using an optional LEMONADE_API_KEY. It does not instruct reading arbitrary host files or other env vars.
  4. Operational note: The skill will transmit prompts and model requests to whichever server_url is used. If a remote URL is supplied, the agent will send user content and model commands there (expected for this skill, but worth the user's attention).

API Operations

System Info

Get hardware capabilities and device enumeration limits.

Returns: JSON object with hardware capabilities.

# Example Request
curl -X GET "http://localhost:8000/api/v1/system-info" \
  -H "Authorization: Bearer ${LEMONADE_API_KEY}"

Health Check

Verify status and monitor currently loaded models to prevent VRAM overflow.

Returns: JSON object containing server status and currently loaded models.

# Example Request
curl -X GET "http://localhost:8000/api/v1/health" \
  -H "Authorization: Bearer ${LEMONADE_API_KEY}"

List Models

Get an array of downloaded models available to load into memory.

Returns: JSON list of all downloaded and available models.

# Example Request
curl -X GET "http://localhost:8000/api/v1/models" \
  -H "Authorization: Bearer ${LEMONADE_API_KEY}"

Pull Model

Download and install a new model string to the target machine.

Returns: JSON stream or object confirming download status.

# Example Request
curl -X POST "http://localhost:8000/api/v1/pull" \
  -H "Authorization: Bearer ${LEMONADE_API_KEY}" \
  -H "Content-Type: application/json" \
  -d '{"model": "llama3"}'

Load Model

Load a model into VRAM/NPU to prepare for prompt responses.

Returns: JSON object confirming model loaded into memory.

# Example Request
curl -X POST "http://localhost:8000/api/v1/load" \
  -H "Authorization: Bearer ${LEMONADE_API_KEY}" \
  -H "Content-Type: application/json" \
  -d '{"model": "llama3"}'

Unload Model

Force unload a model to free up memory before loading a larger variant.

Returns: JSON object confirming model unloaded.

# Example Request
curl -X POST "http://localhost:8000/api/v1/unload" \
  -H "Authorization: Bearer ${LEMONADE_API_KEY}" \
  -H "Content-Type: application/json" \
  -d '{"model": "llama3"}'

Chat Completion

Send a standard chat request to the LLM backend.

Returns: JSON object containing standard chat completion response.

# Example Request
curl -X POST "http://localhost:8000/api/v1/chat/completions" \
  -H "Authorization: Bearer ${LEMONADE_API_KEY}" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "llama3",
    "messages": [
      {"role": "user", "content": "Hello, world!"}
    ]
  }'

Generate Image

Submit a stable-diffusion prompt for image generation.

Returns: JSON object containing base64 generated image data.

# Example Request
curl -X POST "http://localhost:8000/api/v1/images/generations" \
  -H "Authorization: Bearer ${LEMONADE_API_KEY}" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "sdxl",
    "prompt": "A majestic lion standing on a cliff, digital art"
  }'

Tips

  • The flm, ryzenai-llm, and whispercpp backends are mutually exclusive on an NPU. Always use the /api/v1/unload endpoint on conflicting LLMs before attempting to load a Whisper model on an NPU.
  • Check /api/v1/health first to assess available hardware VRAM before pulling or loading a multi-gigabyte LLM to avoid out-of-memory errors on the host.
  • Provide clear feedback to the user if an API request fails due to resource constraints and suggest unloading inactive background models.

Comments

Loading comments...