Podcast Generation with Microsoft Foundry

v0.1.0

Generate AI-powered podcast-style audio narratives using Azure OpenAI's GPT Realtime Mini model via WebSocket. Use when building text-to-speech features, audio narrative generation, podcast creation from content, or integrating with Azure OpenAI Realtime API for real audio output. Covers full-stack implementation from React frontend to Python FastAPI backend with WebSocket streaming.

3· 2.4k·7 current·7 all-time
MIT-0
Download zip
LicenseMIT-0 · Free to use, modify, and redistribute. No attribution required.
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Suspicious
high confidence
!
Purpose & Capability
The skill's name/description (podcast audio via Azure OpenAI Realtime) matches the implementation. However, the SKILL.md and referenced code expect AZURE_OPENAI_AUDIO_API_KEY, AZURE_OPENAI_AUDIO_ENDPOINT, and AZURE_OPENAI_AUDIO_DEPLOYMENT (and a settings object) while the registry metadata lists no required environment variables or config paths. That discrepancy is significant: consumers cannot see from metadata that sensitive credentials are needed.
!
Instruction Scope
The runtime instructions focus on WebSocket to Azure, streaming audio chunks, and PCM→WAV conversion — all within the declared purpose. But the references and example service code also describe fetching content from a database (tags/bookmarks), building prompts from user data, saving audio to a DB, and exposing streaming endpoints. Those broader actions (DB reads/writes and content aggregation) are not surfaced in the skill metadata and expand the scope beyond mere TTS conversion.
Install Mechanism
No install specification is provided (instruction-only + a small utility script). Nothing is downloaded or written to disk by an install step, which lowers supply-chain risk.
!
Credentials
The SKILL.md explicitly requires AZURE_OPENAI_AUDIO_API_KEY, AZURE_OPENAI_AUDIO_ENDPOINT, and AZURE_OPENAI_AUDIO_DEPLOYMENT, but the package metadata declared no required environment variables or config paths. Additionally, the code examples assume access to application settings and a database (db operations and settings.*), which would require additional credentials (DB connection string, possibly cloud storage). The absence of these declarations is disproportionate and opaque.
Persistence & Privilege
The skill does not request always:true and contains no install hooks or instructions to modify other skills or system-wide config. It does describe saving audio artifacts to a database in examples, which is normal for this application but should be explicitly declared to users.
What to consider before installing
This skill's code and documentation require an Azure OpenAI realtime API key, endpoint, and deployment name and also include example code that reads from and writes to a database — but the registry entry does not declare those requirements. Before installing: (1) ask the publisher to update metadata to list required env vars (AZURE_OPENAI_AUDIO_API_KEY, AZURE_OPENAI_AUDIO_ENDPOINT, AZURE_OPENAI_AUDIO_DEPLOYMENT) and any DB/config needs; (2) verify where secrets will be stored and supplied (do not paste production Azure keys into unknown skills); (3) review / audit any database integration code and ensure least-privilege credentials (read-only where possible) and proper access controls; (4) run the skill in an isolated/test environment first and rotate keys after verification; and (5) if you cannot obtain clear metadata or a trusted publisher, treat the skill as untrusted and do not install it with production credentials.

Like a lobster shell, security has layers — review code before you run it.

latestvk97czrjzakybr4rp5wtgaxpkp9809sbb

License

MIT-0
Free to use, modify, and redistribute. No attribution required.

Comments