{"skill":{"slug":"cm-hallucination-detector","displayName":"Hallucination Detector","summary":"Detect and flag hallucinations in LLM outputs by cross-referencing claims against source documents, code, and verifiable data. Essential for RAG pipelines an...","tags":{"latest":"1.0.0"},"stats":{"comments":0,"downloads":36,"installsAllTime":0,"installsCurrent":0,"stars":0,"versions":1},"createdAt":1777509696594,"updatedAt":1777510609234},"latestVersion":{"version":"1.0.0","createdAt":1777509696594,"changelog":"Initial release of hallucination-detector.\n\n- Detects and flags hallucinations in LLM outputs by checking claims against source code, documentation, and data.\n- Classifies issues as fabrication, contradiction, exaggeration, outdated, or unverifiable.\n- Provides confidence scores for each claim and detailed output reports.\n- Supports integration into CI pipelines, RAG workflows, and AI content review processes.\n- Includes detection of common hallucination patterns and actionable recommendations.","license":"MIT-0"},"metadata":null,"owner":{"handle":"charlie-morrison","userId":"s17cttbdxry5kkyafjw983mq8s83p4y3","displayName":"charlie-morrison","image":"https://avatars.githubusercontent.com/u/271589886?v=4"},"moderation":null}