Skill flagged — suspicious patterns detected

ClawHub Security flagged this skill as suspicious. Review the scan results before using.

CertainLogic Hallucination Guard

v1.0.1

Install, configure, and use CertainLogic Verifier (hallucination‑guard) – deterministic AI verification middleware that catches hallucinations before they re...

0· 61·0 current·0 all-time

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for certainlogicai/certainlogic-hallucination-guard.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "CertainLogic Hallucination Guard" (certainlogicai/certainlogic-hallucination-guard) from ClawHub.
Skill page: https://clawhub.ai/certainlogicai/certainlogic-hallucination-guard
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Bare skill slug

openclaw skills install certainlogic-hallucination-guard

ClawHub CLI

Package manager switcher

npx clawhub@latest install certainlogic-hallucination-guard
Security Scan
Capability signals
CryptoCan make purchasesRequires sensitive credentials
These labels describe what authority the skill may exercise. They are separate from suspicious or malicious moderation verdicts.
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Suspicious
high confidence
!
Purpose & Capability
Name/description, API docs, and integration guides consistently describe a self‑hosted verifier middleware, which matches the bundled docs and assets. However the SKILL.md simultaneously claims 'self‑hosted & air‑gapped' while recommending external services (OpenRouter) for cache warm‑up and referencing an external docker image (certainlogic/hallucination-guard:latest). That contradicts the air‑gapped claim and is not justified in the metadata.
!
Instruction Scope
Runtime instructions are concrete (git clone, pip install, start uvicorn, endpoints for validate/cache/facts). They do not instruct arbitrary file scanning or secret harvesting, but they do tell users to set/use an OPENROUTER_API_KEY for cache warm‑ups and suggest using an external image or OpenRouter model calls — which introduces network calls and potential data flow off your host despite the 'air‑gapped' marketing. The docs also reference HALLUCINATION_GUARD_URL and other env vars that are not declared in registry metadata.
!
Install Mechanism
There is no formal install spec in the registry, but the included scripts/install.sh clones a GitHub repo and runs 'pip install -r requirements.txt'. The script header advertises a curl|bash one‑liner (raw.githubusercontent URL) which encourages executing remote code. Using GitHub is common, but pip installing requirements from a remote repo without inspecting requirements.txt or the Dockerfile can pull arbitrary third‑party packages — a moderate risk. The repository build and Docker recommendations will also execute code not included in the skill bundle (Dockerfile, requirements, application code).
!
Credentials
Registry metadata declares no required env vars/credentials, but SKILL.md and references require/mention several environment variables (OPENROUTER_API_KEY for warming/run fallback, PRODUCT_MODE, CACHE_DIR, LOG_LEVEL, HALLUCINATION_GUARD_URL). In particular, OPENROUTER_API_KEY is effectively a credential for an external service yet is not declared. This mismatch is a red flag: the runtime requires an external API key for optional features, and that contradicts 'no required env vars' in the metadata.
Persistence & Privilege
The skill does not request 'always: true' or any privileged persistent presence. It doesn't declare required config paths in the registry. The included docker-compose mounts local facts_db.json and audit_log.jsonl (expected for this product) but there is no evidence the skill modifies other skills or system‑wide settings. Autonomous invocation (disable-model-invocation: false) is default — not a standalone concern here.
What to consider before installing
What to check before installing: - Inspect the remote repository (https://github.com/CertainLogicAI/hallucination-guard) before running the installer or a curl|bash one‑liner. Review the Dockerfile and requirements.txt for unexpected packages or startup scripts. - Confirm whether you will use OpenRouter or any external model provider: SKILL.md mentions OPENROUTER_API_KEY and warming.run (external network calls). If you need true air‑gapped operation, plan to disable or audit those features. - The registry metadata lists no required env vars, but the docs require OPENROUTER_API_KEY and a HALLUCINATION_GUARD_URL in integrations — treat this as an undocumented credential dependency and document/lock where keys are stored. - Prefer pulling the repo and reviewing files locally rather than piping a remote install script to bash. If you do run the installer, run it in an isolated test environment (container or VM) first and monitor outbound network traffic. - Verify file permissions for facts_db.json and audit_log.jsonl (they may contain sensitive/PII or audit trails). Consider using the provided PostgreSQL option for persistent logs if you need harder isolation. - If you rely on compliance claims (HIPAA/GDPR/FedRAMP), validate the implementation and cryptographic log design independently — marketing claims are not proof of compliance. Given the inconsistencies (air‑gapped vs external services, undocumented env vars) and the moderate install risks, treat this skill as suspicious until you or your security team review the upstream repo and dependency list.

Like a lobster shell, security has layers — review code before you run it.

latestvk97fvmp2fspsvar01m8me97m5s85dz9r
61downloads
0stars
1versions
Updated 4d ago
v1.0.1
MIT-0

Hallucination Guard – CertainLogic Verifier

Overview

CertainLogic Verifier is an open‑source, self‑hosted middleware layer that sits between your LLM calls and your application. It validates every AI response against a verified facts database, flags hallucinations, caches verified answers (bypassing the LLM), and provides cryptographic audit logs.

Key capabilities:

  • 99%+ hallucination block rate – rule‑based checks + TF‑IDF memory search against your facts_db
  • 85‑98% token savings – semantic cache hits skip the LLM entirely
  • Self‑hosted & air‑gapped – nothing leaves your infrastructure; ready for HIPAA/GDPR/SOC2/FedRAMP
  • MIT licensed – no proprietary lock‑in; inspect every validation rule
  • Deterministic grounding – same query → same verified answer, every time
  • Cryptographic audit logs – SHA‑256 chained JSONL for compliance

Quick Start (2‑Minute Install)

# Clone the repository
git clone https://github.com/CertainLogicAI/hallucination-guard
cd hallucination-guard

# Set up Python environment
python3 -m venv venv
source venv/bin/activate  # On Windows: venv\Scripts\activate
pip install -r requirements.txt

# Start the service
uvicorn main:app --host 0.0.0.0 --port 8000

Verify it's working:

curl -X POST http://localhost:8000/validate \
  -d '{"query": "What is the price of GPT‑5?", "response": "\$200/month"}'

Installation Options

1. Docker (Recommended for Production)

docker build -t hallucination-guard .
docker run -p 8000:8000 hallucination-guard

2. Kubernetes/Helm

See deploy/helm/ in the repository for production‑ready Helm charts.

3. Systemd Service

A sample systemd unit file is included at deploy/systemd/hallucination-guard.service.

Configuration

Facts Database

The verifier checks responses against facts_db.json. Populate it with your domain‑specific verified facts.

Example entry:

{
  "fact": "Python was created in 1991 by Guido van Rossum",
  "category": "programming",
  "source": "official Python history",
  "verified_at": "2026‑04‑20"
}

Adding facts:

  • Manually edit facts_db.json
  • Use the /facts/add endpoint (POST with JSON)
  • Bulk‑load from documents via the /warming/extract endpoint

Environment Variables

Set these in .env or as environment variables:

PRODUCT_MODE=coder           # coder|agent (determines rate limits)
OPENROUTER_API_KEY=your_key  # Required for cache‑miss fallback
LOG_LEVEL=INFO              # DEBUG|INFO|WARNING|ERROR
CACHE_DIR=./cache           # Persistent cache storage

Usage

Validating a Single Response

import requests

response = requests.post(
    "http://localhost:8000/validate",
    json={
        "query": "What year was Python created?",
        "response": "Python was created in 1991."
    }
)
print(response.json())

Integrating with AI Agent Pipelines

Place the verifier between your LLM call and your application logic:

def get_ai_response(query):
    # 1. Check cache first
    cache_check = requests.post("http://localhost:8000/cache/check", 
                                json={"query": query})
    if cache_check.json().get("cached"):
        return cache_check.json()["response"]
    
    # 2. Call LLM
    llm_response = call_llm(query)
    
    # 3. Validate
    validation = requests.post("http://localhost:8000/validate",
                               json={"query": query, "response": llm_response})
    
    if validation.json().get("valid"):
        return llm_response
    else:
        # Handle hallucination
        raise ValueError(f"Hallucination detected: {validation.json()}")

Cache Management

  • View cache stats: GET /cache/stats
  • Clear cache: POST /cache/clear
  • Warm cache: POST /warming/run (requires OpenRouter API key)

Advanced Features

Deterministic Memory Search

The verifier uses TF‑IDF similarity to match queries against known facts, even with paraphrasing.

Uncertainty Detection

Responses containing "I think", "might be", "not sure" are penalized and flagged for review.

Numeric‑Unit Matching

Checks that numeric values match known facts with correct units (e.g., "5 km" vs "5 miles").

Audit Logs

All validations are logged to audit_log.jsonl with SHA‑256 chaining for tamper evidence.

Resources

scripts/

  • install.sh – One‑line installer for Linux/macOS
  • docker-compose.yml – Multi‑service setup with PostgreSQL for audit logs

references/

  • api-reference.md – Complete API documentation
  • facts-schema.md – Facts database schema and validation rules
  • integration-guide.md – Step‑by‑step integration with popular AI frameworks

assets/

  • sample-facts.json – Example facts database with 50+ verified entries
  • docker-compose.prod.yml – Production‑ready Docker Compose configuration

Support & Community

License

MIT – see LICENSE in the repository.

Comments

Loading comments...