N8N QDrant workflow expert

v1.0.0

Expertise in designing, building, and troubleshooting production-grade n8n workflows for Qdrant ingestion, retrieval, hybrid search, and RAG pipelines.

0· 111·0 current·0 all-time

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for encryptshawn/n8n-qdrant.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "N8N QDrant workflow expert" (encryptshawn/n8n-qdrant) from ClawHub.
Skill page: https://clawhub.ai/encryptshawn/n8n-qdrant
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Bare skill slug

openclaw skills install n8n-qdrant

ClawHub CLI

Package manager switcher

npx clawhub@latest install n8n-qdrant
Security Scan
VirusTotalVirusTotal
Pending
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
The name/description (n8n workflows for Qdrant ingestion, retrieval, hybrid search, RAG) match the provided SKILL.md and example workflows. References to Qdrant, LangChain vector store, OpenAI/Gemini embeddings, Slack, and optional sparse encoders are expected for this use case.
Instruction Scope
SKILL.md and docs are comprehensive and stay within the domain of designing n8n workflows. They include instructions to call external APIs (OpenAI, Qdrant, Google Gemini), to deploy or call a sparse-encoder service, and to use code nodes inside n8n (JS). No instructions attempt to read unrelated system files or secretly exfiltrate data, but the workflows explicitly send user content to third‑party services — review that carefully before enabling in production.
Install Mechanism
There is no install spec and no code shipped that would be downloaded or executed by the platform. This is instruction-only documentation and example JSON for n8n — lowest install risk.
Credentials
The registry metadata lists no required env vars/credentials, but the workflows and docs clearly expect credentials and environment variables for OpenAI, Qdrant, Slack, Google/Gemini, and possibly a sparse-encoder service. Those credentials are appropriate for the skill's purpose, but the metadata omission means users must manually supply and configure them in n8n; verify all credential placeholders (e.g., CONFIGURE_ME_*, $env.OPENAI_API_KEY) before use.
Persistence & Privilege
always:false and there is no code that modifies other skills or system-wide settings. The skill does not request persistent platform privileges.
Assessment
This skill is documentation and example workflows for n8n→Qdrant RAG pipelines and appears internally consistent. Before importing or running any of the example workflows: 1) Replace all CONFIGURE_ME_* placeholders and verify the intended credential method (n8n credentials vs environment variables). 2) Confirm where sensitive data will be sent — examples call OpenAI, Qdrant, Gemini, Slack, and show an optional sparse-encoder HTTP endpoint; avoid sending confidential data to third-party endpoints you don't control. 3) Don’t paste production secrets into workflow JSON — configure them via n8n credentials or secure env storage. 4) If you run the sparse encoder as a service, host it in your environment or add authentication to avoid leaking queries. 5) Test in a sandbox account and audit logs/rate limits/costs before enabling scheduled ingestion. The only metadata inconsistency: the registry declares no required env vars while the examples depend on external service credentials — that’s expected for instruction-only skills but worth noting.

Like a lobster shell, security has layers — review code before you run it.

latestvk974ytd75adja0k29g2m2122dn844v0c
111downloads
0stars
1versions
Updated 3w ago
v1.0.0
MIT-0

n8n + Qdrant: Ingestion & RAG Pipeline Skill

Overview

This skill enables AI agents to design, build, and troubleshoot production-grade Qdrant ingestion and retrieval pipelines in n8n. It covers the full lifecycle: source data extraction → chunking → metadata enrichment → vector embedding → Qdrant upsert → retrieval (dense, sparse, hybrid) → RAG response generation.

Always read the supporting docs in /docs/ before building workflows:

  • docs/NODE-REFERENCE.md — Every Qdrant node, mode, and parameter explained
  • docs/INGESTION-PIPELINE.md — Step-by-step ingestion architecture
  • docs/RAG-RETRIEVAL.md — Dense, sparse, and hybrid retrieval patterns
  • docs/CHUNKING-METADATA.md — Chunking strategies and metadata schema design
  • docs/examples/ — Annotated workflow JSON examples

Two Node Systems to Know

n8n has two separate Qdrant integration systems — knowing which to use is critical:

1. Official Qdrant Node (n8n-nodes-qdrant)

  • Package: n8n-nodes-qdrant (community node, install via n8n Settings → Community Nodes)
  • Purpose: Direct Qdrant API operations — collection management, point upsert/delete/scroll, search queries
  • Node name in editor: Qdrant
  • Use for: Building custom ingestion pipelines, running Query Points (dense/sparse/hybrid search), collection setup, point management
  • GitHub: https://github.com/qdrant/n8n-nodes-qdrant

2. LangChain Vector Store Node (built-in)

  • Package: Built into n8n's AI/LangChain nodes
  • Purpose: LangChain-compatible vector store integration — connects with Document Loaders, Text Splitters, Embeddings, and AI Agents
  • Node name in editor: Qdrant Vector Store (@n8n/n8n-nodes-langchain.vectorStoreQdrant)
  • Use for: LangChain-style RAG pipelines, AI Agent tool integration, retrieve-as-tool mode
  • Modes: insert (ingest documents), retrieve (similarity search), retrieve-as-tool (AI agent tool)

Rule of thumb: Use LangChain Vector Store for LangChain-native agent/RAG flows. Use the Official Qdrant Node for direct API control, hybrid search, payload operations, and production ingestion pipelines.


Quick Decision Matrix

GoalUse This NodeMode/Operation
Ingest documents via LangChain chainLangChain Vector Storeinsert
AI Agent retrieves from Qdrant as toolLangChain Vector Storeretrieve-as-tool
Run hybrid (dense+sparse) searchOfficial Qdrant NodeSearch → Query Points
Create/manage collectionsOfficial Qdrant NodeCollection → Create Collection
Upsert raw points with custom payloadsOfficial Qdrant NodePoint → Upsert Points
Delete points by filter (e.g. file_id)Official Qdrant NodePoint → Delete Points
Scroll all points for audit/exportOfficial Qdrant NodePoint → Scroll Points
Batch ingest large datasetsOfficial Qdrant NodePoint → Batch Update Points

Canonical Ingestion Pipeline Architecture

[Trigger]
    │
    ▼
[Source Node] ──────────────────────────────────────────────
(Slack, Fireflies, Google Drive, HTTP, DB, etc.)             │
    │                                                         │
    ▼                                                         │
[Split in Batches]  ←── Loop for large datasets              │
    │                                                         │
    ▼                                                         │
[Extract/Normalize]                                          │
(Set node: build content string + raw metadata)              │
    │                                                         │
    ▼                                                         │
[AI: Extract Metadata]                                       │
(Information Extractor or LLM Chain)                         │
Produces: themes, keywords, entities, summary, tags          │
    │                                                         │
    ▼                                                         │
[Text Splitter]                                              │
(Token Splitter or Recursive Character Splitter)             │
chunkSize: 512–2000 tokens, overlap: 10–15%                  │
    │                                                         │
    ▼                                                         │
[Embeddings Node]                                            │
(OpenAI text-embedding-3-large or similar)                   │
    │                                                         │
    ▼                                                         │
[Qdrant Vector Store — insert mode]  OR                      │
[Official Qdrant Node — Upsert Points]                       │
    │                                                         │
    ▼                                                         │
[Wait Node]  ←── Rate limiting / backpressure                │
    │                                                         │
    └─────────────────── back to Split in Batches ───────────┘

See docs/INGESTION-PIPELINE.md for full node-by-node configuration.


Canonical RAG Retrieval Architecture

[Chat Trigger / Webhook]
    │
    ▼
[AI Agent Node]
    │
    ├── [LLM: Gemini / GPT-4o / Claude]
    ├── [Memory: Window Buffer Memory]
    └── [Tool: Qdrant Vector Store — retrieve-as-tool]
              │
              └── [Embeddings Node]

For hybrid search (dense + sparse), use the Official Qdrant Node → Query Points with a prefetch array combining dense and sparse queries + RRF fusion. See docs/RAG-RETRIEVAL.md.


Credentials Setup

Official Qdrant Node

  • Credential type: qdrantApi
  • Fields: URL (e.g. https://your-cluster.cloud.qdrant.io) + API Key

LangChain Vector Store Node

  • Credential type: qdrantApi (same credential, shared)

Qdrant Cloud Setup

  1. Open https://cloud.qdrant.io → select cluster
  2. Copy Endpoint → use as URL
  3. Go to API Keys tab → copy key

Local (Docker / AI Starter Kit)

  • URL: http://qdrant:6333/
  • Set QDRANT_API_KEY=your_key in docker-compose environment

Naming Conventions

Use consistent naming across workflows:

ElementConventionExample
Collection name{org}-{source}-{content-type}acme-slack-messages
Metadata key for source IDsource_id"source_id": "C01234-1709123456"
Metadata key for document IDdoc_id"doc_id": "file_abc123"
Metadata key for chunk indexchunk_index"chunk_index": 3
Metadata key for timestampcreated_atISO 8601 string
Metadata key for source typesource_type"slack", "fireflies", "gdrive"
Metadata key for channel/foldersource_context"#engineering"

Critical Rules

  1. Always set file_id or doc_id in metadata — enables targeted deletion without full collection wipe
  2. Always use onError: continueRegularOutput on the Qdrant Vector Store node — prevents single-item failures from crashing the whole batch
  3. Always use retryOnFail: true on the Qdrant node for production ingestion
  4. Chunk before embedding — never embed full documents; always split first
  5. Never store raw text in collection names or keys — normalize to lowercase slug format
  6. Use Split in Batches with a Wait node for large datasets — prevents API rate limit errors and memory exhaustion
  7. Run metadata extraction BEFORE the text splitter — extract from the full document, then attach metadata to each chunk
  8. For delete operations, always add human-in-the-loop confirmation (Telegram sendAndWait, Slack approval, etc.)

Comments

Loading comments...