{"skill":{"slug":"lieutenant","displayName":"Lieutenant - AI Agent Security","summary":"AI agent security and trust verification. Scan messages, agent cards, and A2A communications for prompt injection, jailbreaks, and malicious patterns. Use when protecting agents from attacks, verifying external agents, or scanning untrusted content.","tags":{"latest":"1.0.0"},"stats":{"comments":0,"downloads":1282,"installsAllTime":1,"installsCurrent":0,"stars":0,"versions":1},"createdAt":1770418801744,"updatedAt":1777525026603},"latestVersion":{"version":"1.0.0","createdAt":1770418801744,"changelog":"- Initial release of Lieutenant, an AI agent security and trust verification tool.\n- Scans messages, agent cards, and A2A communications for prompt injection, jailbreaks, and malicious patterns.\n- Detects 65+ threat patterns across 10 categories, including prompt injection, jailbreak, data exfiltration, and more.\n- Supports semantic analysis for paraphrased threat detection (requires OpenAI API key).\n- Integrates with TrustAgents API to enhance detection with reputation and crowdsourced threat intelligence.\n- Provides command-line tools, Python API, and A2A SDK middleware for flexible use and integration.","license":null},"metadata":null,"owner":{"handle":"jd-delatorre","userId":"publishers:jd-delatorre","displayName":"jd-delatorre","image":"https://avatars.githubusercontent.com/u/1874259?v=4"},"moderation":{"isSuspicious":true,"isMalwareBlocked":false,"verdict":"suspicious","reasonCodes":["suspicious.llm_suspicious","suspicious.prompt_injection_instructions"],"summary":"Detected: suspicious.llm_suspicious, suspicious.prompt_injection_instructions","engineVersion":"v2.4.5","updatedAt":1777525026603}}