Nirvana Plugin

v1.0.0

Local-first privacy-first inference. Your OpenClaw agent thinks locally and asks the cloud intelligently. Saves 85%+ tokens, protects privacy, agent learns f...

0· 18·0 current·0 all-time
byShiva&G@shivaclaw

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for shivaclaw/project-nirvana-plugin.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "Nirvana Plugin" (shivaclaw/project-nirvana-plugin) from ClawHub.
Skill page: https://clawhub.ai/shivaclaw/project-nirvana-plugin
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Canonical install target

openclaw skills install shivaclaw/project-nirvana-plugin

ClawHub CLI

Package manager switcher

npx clawhub@latest install project-nirvana-plugin
Security Scan
Capability signals
CryptoRequires walletCan make purchasesRequires sensitive credentials
These labels describe what authority the skill may exercise. They are separate from suspicious or malicious moderation verdicts.
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
medium confidence
Purpose & Capability
Name/description (local-first privacy-preserving inference) align with what the package implements: local Ollama integration, a router, context stripper, audit logging, and response integrator. Declared permissions to read identity files and write memory/cache are coherent with implementing context-stripping and caching. There are no unrelated environment variables or mysterious third-party credentials requested.
Instruction Scope
SKILL.md and INSTALL.md instruct the agent/operator to run Docker (ollama/ollama), install the plugin via ClawHub/OpenClaw, and to check local audit logs. These instructions reference sensitive local files (memory/nirvana-audit.log, memory/*, SOUL.md, USER.md, MEMORY.md) — which is expected for a privacy enforcer but should be reviewed. The delivery checklist also asks for uploading an archive to Google Drive (external endpoint) as part of publication; this is a publishing workflow artifact, not a runtime requirement, but it is an external upload step worth reviewing for data exposure.
Install Mechanism
There is no automated install spec embedded with the skill bundle; installation is manual via ClawHub/OpenClaw and Docker. The runtime will pull official Docker images (ollama/ollama) and auto-pull models (qwen2.5:7b) — these are standard but large downloads. No obscure or shortened URLs are used. Verify the Docker image source and model provenance before pulling (official Ollama image expected).
Credentials
The skill requests no environment variables or external API keys by default (cloud fallback is optional). The plugin manifest and code explicitly require read access to identity files (SOUL.md, USER.md, MEMORY.md) and write access to memory/* for audit/caching. That access is proportionate to implementing context stripping/auditing, but those are highly sensitive files — granting read access to them is a significant privilege and should only be given to trusted code after code review.
Persistence & Privilege
The skill does not set always:true and uses normal plugin hooks (on-query, on-response). It writes to its own memory/cache paths and audit log. There is no evidence it attempts to persistently modify unrelated plugin configs or request platform-wide always-on privileges. Autonomous invocation is allowed (platform default) but not combined with an unusual persistent privilege.
Scan Findings in Context
[system-prompt-override] expected: A prompt-injection pattern was detected in SKILL.md. For a context-stripper/privacy plugin it can be expected because the plugin must inspect and rewrite agent prompts before export; nevertheless, modifying or overriding system prompts is a sensitive capability — review the exact text and code paths that rewrite or generate prompts to ensure they do not leak private data or grant broader control.
Assessment
What to check before installing: - Review the code that reads SOUL.md, USER.md, MEMORY.md (src/context-stripper.js and privacy-auditor) to confirm stripping logic is correct and thorough. These files are very sensitive — the plugin needs to read them to strip data, but that requires trust. - Verify Docker image and model provenance before pulling (docker run ollama/ollama will pull from Docker Hub). If you need higher assurance, pull images from a verified registry or inspect image contents first. - The plugin caches cloud responses and writes audit logs to memory/*. Confirm audit logs are stored only locally at the configured path and are rotated/protected as appropriate. - The DELIVERY checklist includes uploading an archive to Google Drive — that is a publication step, not a runtime requirement. Do not upload any private data to external drives as part of setup. - If you enable cloud fallback, only add API keys you control and check that the context-stripping configuration (contextStripDepth, identityFilesNeverExport) is enforced before any cloud calls. Test with synthetic sensitive data to validate stripping. - Because the plugin will handle system prompts and can rewrite queries, do a small controlled test session and inspect outgoing HTTP payloads (or run in an air-gapped environment) to confirm no private fields are transmitted. Overall: the package looks coherent for its stated purpose, but it operates on very sensitive data and downloads large models — perform a code review and controlled testing before trusting it with real private files.

Like a lobster shell, security has layers — review code before you run it.

latestvk971ajd30wzcvgdqzyq2395zs185e4wd
18downloads
0stars
1versions
Updated 4h ago
v1.0.0
MIT-0

Project Nirvana: Local-First, Privacy-First Inference

A new way of thinking about LLM access. Your agent thinks locally, asks the cloud intelligently, and learns from the response. The cloud never sees your private data.


The Problem

Today's approach leaks your privacy and wastes 85% of your API budget.

Every time you ask your OpenClaw agent a question:

  1. Your agent builds a "system prompt" containing:

    • Excerpts from its SOUL.md and MEMORY.md
    • Your personal information from its USER.md
    • Your entire chat history (context window)
  2. All of this gets sent to cloud APIs (OpenAI, Anthropic, Google)

  3. You pay for thousands of extra tokens

  4. The cloud provider trains its next model on your private data

This is the current default. It's inefficient and it's a privacy disaster.


The Solution: Nirvana

Local-first inference that protects privacy and slashes costs.

Nirvana flips the paradigm:

  1. Your agent thinks locally using Ollama (free, private, on your hardware)
  2. For complex questions, it asks the cloud — but only sends its own carefully-crafted queries
  3. Your private data never leaves your system
  4. The cloud's responses are cached locally — your agent learns from them

The Paradigm Shift

AspectToday (Default)Nirvana
Where thinking happensCloud onlyLocal first, cloud when needed
What gets sent to cloudYour full context + system promptsAgent's sanitized query only
Who learns from your dataCloud providerYou (local agent)
Token cost per interaction2,000–5,000 tokens50–300 tokens
Savings85%+ token reduction
PrivacyLeakedProtected

What Nirvana Does

Local Inference

  • Bundled Ollama with qwen2.5:7b (free, 3.5GB model)
  • Handles 80%+ of queries locally (no API calls)
  • ~200 tokens/second on CPU; 3–5x faster with GPU
  • Works offline, no internet required

Privacy Enforcement

  • Context stripper — Removes SOUL.md, USER.md, MEMORY.md before cloud queries
  • Prompt sanitizer — Agent rewrites its own questions for the cloud (never sends yours)
  • Audit trail — Every decision logged; transparent boundary crossing
  • Zero telemetry — No data sent to third parties

Intelligent Routing

  • Complexity analyzer — Decides: local vs cloud?
  • Semantic understanding — "Can qwen2.5:7b handle this, or do I need Claude?"
  • Seamless fallback — Cloud APIs used transparently when needed
  • User override@local or @cloud hints respected

Learning & Caching

  • Response integrator — Cloud responses cached locally
  • Agent learns — Reuse cached answers for similar future questions
  • No repeated payments — Cloud only answers novel questions

How It Works

User asks your agent a question
    ↓
┌─────────────────────────────────────────┐
│ Nirvana Router                          │
│ "Can qwen2.5:7b answer this locally?"  │
└─────────────────────────────────────────┘
    ↙                                   ↘

[LOCAL PATH]                        [CLOUD PATH]
80%+ of queries                     20%- of queries
Ollama (qwen2.5:7b)                 OpenAI/Anthropic/Google
Free                                Pay for answer
Private                             Cloud sees sanitized query only
~1s latency                         ~3s latency
Result cached locally               Result cached locally
    ↓                                   ↓
    └─────────────┬─────────────────────┘
                  ↓
    Agent answers your question using:
    - Local inference (primary)
    - Cloud intelligence (if needed)
    - Cached knowledge (if available)
    
    YOUR PRIVATE DATA NEVER LEFT YOUR SYSTEM

Installation

Prerequisites

  • OpenClaw 2026.3.24+
  • Docker (for Ollama) — or pre-existing local LLM at any endpoint

Two Paths

Path A: Use Bundled Ollama + qwen2.5:7b (Out-of-box)

# Install the plugin
clawhub install shivaclaw/nirvana

# Start Ollama container (pulls auto on first run)
docker run -d -p 11434:11434 ollama/ollama

# Verify
openclaw nirvana status

Path B: Use Existing Local LLM (Any Provider)

# Install the skill (context stripping only)
clawhub install shivaclaw/nirvana-local

# Configure endpoint
openclaw nirvana configure --local-endpoint http://your-llm:5000

# Verify
openclaw nirvana status

Cost Impact

Token Savings

ScenarioTodayWith NirvanaSavings
10 questions/day20,000 tokens/day3,000 tokens/day85%
100 questions/day200,000 tokens/day30,000 tokens/day85%
Monthly cost (OpenAI GPT-4)$500–$1,000$75–$15085%

Local inference is free. Only pay for the 15%–20% of queries that truly need frontier models.


Privacy Guarantee

What Never Leaves Your System

  • ✅ SOUL.md (agent identity)
  • ✅ USER.md (your personal info)
  • ✅ MEMORY.md (agent memories)
  • ✅ Chat history (your actual questions)
  • ✅ Code snippets, documents, secrets

What Optionally Goes to Cloud

  • ✅ Agent's own sanitized query (no personal data)
  • ✅ Task-specific context (never your full context)
  • ✅ Result gets cached locally for future reuse

Privacy Audit Trail

# View what was sent to cloud this session
openclaw nirvana audit-log

# Output:
# 2026-04-24 14:23:45 — CLOUD API CALL
# Original query: [REDACTED]
# Sanitized query sent: "Explain quantum entanglement"
# Response cached: Yes
# User data in request: None

Platform Support

PlatformStatusNotes
Linux (Ubuntu/Debian)✅ FullOllama container + native binaries
macOS (Intel/ARM)✅ FullOllama via Docker or native
Windows (WSL2)✅ FullOllama in WSL2 container
VPS (Hostinger, DigitalOcean, AWS)✅ FullDocker Compose ready
Docker container✅ FullOrchestrated via docker-compose
Air-gapped (offline)✅ FullLocal-only mode (no cloud fallback)

Configuration

Basic Setup

{
  "nirvana": {
    "mode": "local-first",
    "local_model": {
      "provider": "ollama",
      "endpoint": "http://ollama:11434",
      "model": "qwen2.5:7b",
      "timeout_ms": 180000
    },
    "routing": {
      "local_threshold": 0.75,
      "max_local_tokens": 8000,
      "cloud_fallback": true
    },
    "privacy": {
      "strip_soul": true,
      "strip_user": true,
      "strip_memory": true,
      "audit_logging": true
    }
  }
}

Custom Local LLM (Non-Ollama)

{
  "nirvana": {
    "local_model": {
      "provider": "custom",
      "endpoint": "http://your-llm-server:5000",
      "api_format": "openai-compatible",
      "model": "your-model-name",
      "timeout_ms": 120000
    }
  }
}

Use Cases

✅ Perfect For

  • Personal AI agents (maximize budget, minimize cost)
  • Private/sensitive workloads (code, healthcare, legal, finance)
  • Latency-critical tasks (local response < 2s)
  • Air-gapped environments (fully offline)
  • Cost-conscious organizations (85% savings)
  • Privacy-first deployments (zero external data exposure)

⚠️ When to Use Cloud

  • Advanced reasoning (Claude Opus for complex problems)
  • Specialized tasks (image generation, audio synthesis)
  • Extreme scale (millions of tokens/day)

Philosophy

Your agent should train itself. The cloud should not train on you.

Today's default paradigm:

  • Cloud provider gains knowledge from every interaction with you
  • You pay for the privilege of training their next model
  • Your private data becomes their training data

Nirvana's paradigm:

  • Your agent gains knowledge from selective cloud interactions
  • You pay only for what you actually need
  • Your private data never leaves your system
  • Cloud providers contribute intelligence; you keep the learning

What's Included

ComponentPurpose
router.tsDecides local vs cloud routing
context-stripper.tsRemoves private data before cloud API calls
privacy-auditor.tsLogs all boundary crossings
response-integrator.tsCaches cloud responses locally
ollama-manager.tsHandles Ollama lifecycle + model management
metrics-collector.tsTracks performance + cost + privacy
config.schema.jsonConfiguration validation

Performance

Benchmarks (qwen2.5:7b on 4-core CPU)

  • Latency (P50): 800ms–1.2s per response
  • Throughput: 180–220 tokens/second
  • Memory: 4.6GB RAM running
  • Accuracy: 85–92% accuracy vs Claude 3.5 on typical tasks
  • GPU acceleration: 3–5x faster with CUDA/Metal

Optimization

  • Use GPU (CUDA/Metal) for production
  • Upgrade to qwen3.5:9b for complex reasoning
  • Enable response caching for repeated patterns
  • Monitor metrics dashboard for bottlenecks

Support & Community


License

MIT-0 — Free to use, modify, and redistribute. No attribution required.


Your agent deserves privacy. Nirvana makes it real.

Comments

Loading comments...