{"skill":{"slug":"csam-shield","displayName":"Csam Shield","summary":"Detects, blocks, and reports Child Sexual Abuse Material using AI-driven image, video, and behavior analysis with automatic NCMEC reporting and evidence pres...","tags":{"latest":"1.0.0"},"stats":{"comments":0,"downloads":529,"installsAllTime":0,"installsCurrent":0,"stars":0,"versions":1},"createdAt":1771687948844,"updatedAt":1777525303408},"latestVersion":{"version":"1.0.0","createdAt":1771687948844,"changelog":"Initial public release of CSAM Shield — a critical safety system for detecting and preventing CSAM.\n\n- Provides advanced computer vision, hash matching, age estimation, anatomical and behavioral analysis to identify child sexual abuse material.\n- Automates NCMEC CyberTipline reporting, evidence preservation, law enforcement notifications, and user account actions.\n- Offers configurable thresholds, extensive API methods, and strict security controls (encryption, audit logging, access restriction).\n- Delivers real-time content blocking, automatic user suspension, and detection of grooming or distribution networks.\n- Strong emphasis on legal compliance, privacy, staff mental health, and evidence management.","license":null},"metadata":null,"owner":{"handle":"raghulpasupathi","userId":"s17ea3bzwjbttkwd39h6mqkr6n83g14c","displayName":"raghulpasupathi","image":"https://avatars.githubusercontent.com/u/15859053?v=4"},"moderation":{"isSuspicious":true,"isMalwareBlocked":false,"verdict":"suspicious","reasonCodes":["suspicious.llm_suspicious","suspicious.vt_suspicious"],"summary":"Detected: suspicious.llm_suspicious, suspicious.vt_suspicious","engineVersion":"v2.4.5","updatedAt":1777525303408}}