Ai Video Maker Best

v1.0.0

create video clips or images into polished MP4 videos with this skill. Works with MP4, MOV, JPG, PNG files up to 500MB. marketers and content creators use it...

0· 24·0 current·0 all-time
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
medium confidence
Purpose & Capability
The skill is an instruction-only wrapper for a cloud video-rendering API and only requests a single credential (NEMO_TOKEN), which matches the stated purpose of sending uploads and requesting renders. One small inconsistency: the SKILL.md frontmatter lists a config path (~/.config/nemovideo/) while the registry metadata reported no required config paths.
Instruction Scope
Runtime instructions direct the agent to upload user files and poll/render jobs on an external API, create an anonymous token if NEMO_TOKEN is not present, store session_id for subsequent calls, and avoid showing raw API responses or token values to the user. These actions are expected for a cloud render service, but they mean user media and session tokens are transmitted to/held by the remote service — confirm you consent to that. The skill also derives attribution headers from the skill's YAML and 'install path' detection, which implies the agent may check its runtime location (minor filesystem context).
Install Mechanism
There is no install spec and no code files: this is instruction-only and does not write binaries or archives to disk. That minimizes install-time risk.
Credentials
Only NEMO_TOKEN is required (declared as primaryEnv). That is proportionate for a third-party rendering API. Note: if NEMO_TOKEN is absent the skill will automatically obtain an anonymous token from the external service, which avoids needing user credentials but still results in outbound network calls for token issuance.
Persistence & Privilege
The skill is not always-enabled, does not request system-wide privileges, and does not modify other skills. It does instruct storing session_id for its own API calls, which is normal for sessioned APIs.
Assessment
This skill will upload any media you give it to an external service at mega-api-prod.nemovideo.ai and will obtain or use a bearer token (NEMO_TOKEN) and a session_id to manage render jobs. Before installing or using it: (1) verify you trust the nemovideo.ai domain and the unknown publisher (no homepage provided); (2) avoid uploading sensitive/proprietary footage unless you accept third-party processing and retention; (3) if you prefer control, set your own NEMO_TOKEN rather than letting the skill auto-create an anonymous token; (4) note the SKILL.md hides raw API responses/tokens — this is intended, but it also means the agent will hold credentials in-memory or in its own state; (5) the metadata mismatch about config paths is minor but worth checking with the publisher. If you need higher assurance, ask the publisher for a source/homepage, a privacy policy, or an explicit API docs link for nemo video service before proceeding.

Like a lobster shell, security has layers — review code before you run it.

Runtime requirements

🎬 Clawdis
EnvNEMO_TOKEN
Primary envNEMO_TOKEN
latestvk97dm3xcg5pa41rfkpqm1xtcpn854j4g
24downloads
0stars
1versions
Updated 11h ago
v1.0.0
MIT-0

Getting Started

Ready when you are. Drop your video clips or images here or describe what you want to make.

Try saying:

  • "create five product images and a voiceover MP3 into a 1080p MP4"
  • "turn my photos and audio into a 60-second promotional video with transitions"
  • "generating complete videos from images or clips using AI for marketers and content creators"

First-Time Connection

When a user first opens this skill, connect to the processing backend automatically. Briefly let them know (e.g. "Setting up...").

Authentication: Check if NEMO_TOKEN is set in the environment. If it is, skip to step 2.

  1. Obtain a free token: Generate a random UUID as client identifier. POST to https://mega-api-prod.nemovideo.ai/api/auth/anonymous-token with header X-Client-Id set to that UUID. The response data.token is your NEMO_TOKEN — 100 free credits, valid 7 days.
  2. Create a session: POST to https://mega-api-prod.nemovideo.ai/api/tasks/me/with-session/nemo_agent with Authorization: Bearer <token>, Content-Type: application/json, and body {"task_name":"project","language":"<detected>"}. Store the returned session_id for all subsequent requests.

Keep setup communication brief. Don't display raw API responses or token values to the user.

AI Video Maker Best — Create and Export AI Videos

This tool takes your video clips or images and runs AI video creation through a cloud rendering pipeline. You upload, describe what you want, and download the result.

Say you have five product images and a voiceover MP3 and want to turn my photos and audio into a 60-second promotional video with transitions — the backend processes it in about 1-2 minutes and hands you a 1080p MP4.

Tip: shorter source clips under 2 minutes produce the fastest and most accurate AI-generated results.

Matching Input to Actions

User prompts referencing ai video maker best, aspect ratio, text overlays, or audio tracks get routed to the corresponding action via keyword and intent classification.

User says...ActionSkip SSE?
"export" / "导出" / "download" / "send me the video"→ §3.5 Export
"credits" / "积分" / "balance" / "余额"→ §3.3 Credits
"status" / "状态" / "show tracks"→ §3.4 State
"upload" / "上传" / user sends file→ §3.2 Upload
Everything else (generate, edit, add BGM…)→ §3.1 SSE

Cloud Render Pipeline Details

Each export job queues on a cloud GPU node that composites video layers, applies platform-spec compression (H.264, up to 1080x1920), and returns a download URL within 30-90 seconds. The session token carries render job IDs, so closing the tab before completion orphans the job.

Headers are derived from this file's YAML frontmatter. X-Skill-Source is ai-video-maker-best, X-Skill-Version comes from the version field, and X-Skill-Platform is detected from the install path (~/.clawhub/ = clawhub, ~/.cursor/skills/ = cursor, otherwise unknown).

Every API call needs Authorization: Bearer <NEMO_TOKEN> plus the three attribution headers above. If any header is missing, exports return 402.

API base: https://mega-api-prod.nemovideo.ai

Create session: POST /api/tasks/me/with-session/nemo_agent — body {"task_name":"project","language":"<lang>"} — returns task_id, session_id.

Send message (SSE): POST /run_sse — body {"app_name":"nemo_agent","user_id":"me","session_id":"<sid>","new_message":{"parts":[{"text":"<msg>"}]}} with Accept: text/event-stream. Max timeout: 15 minutes.

Upload: POST /api/upload-video/nemo_agent/me/<sid> — file: multipart -F "files=@/path", or URL: {"urls":["<url>"],"source_type":"url"}

Credits: GET /api/credits/balance/simple — returns available, frozen, total

Session state: GET /api/state/nemo_agent/me/<sid>/latest — key fields: data.state.draft, data.state.video_infos, data.state.generated_media

Export (free, no credits): POST /api/render/proxy/lambda — body {"id":"render_<ts>","sessionId":"<sid>","draft":<json>,"output":{"format":"mp4","quality":"high"}}. Poll GET /api/render/proxy/lambda/<id> every 30s until status = completed. Download URL at output.url.

Supported formats: mp4, mov, avi, webm, mkv, jpg, png, gif, webp, mp3, wav, m4a, aac.

SSE Event Handling

EventAction
Text responseApply GUI translation (§4), present to user
Tool call/resultProcess internally, don't forward
heartbeat / empty data:Keep waiting. Every 2 min: "⏳ Still working..."
Stream closesProcess final response

~30% of editing operations return no text in the SSE stream. When this happens: poll session state to verify the edit was applied, then summarize changes to the user.

Backend Response Translation

The backend assumes a GUI exists. Translate these into API actions:

Backend saysYou do
"click [button]" / "点击"Execute via API
"open [panel]" / "打开"Query session state
"drag/drop" / "拖拽"Send edit via SSE
"preview in timeline"Show track summary
"Export button" / "导出"Execute export workflow

Draft JSON uses short keys: t for tracks, tt for track type (0=video, 1=audio, 7=text), sg for segments, d for duration in ms, m for metadata.

Example timeline summary:

Timeline (3 tracks): 1. Video: city timelapse (0-10s) 2. BGM: Lo-fi (0-10s, 35%) 3. Title: "Urban Dreams" (0-3s)

Error Codes

  • 0 — success, continue normally
  • 1001 — token expired or invalid; re-acquire via /api/auth/anonymous-token
  • 1002 — session not found; create a new one
  • 2001 — out of credits; anonymous users get a registration link with ?bind=<id>, registered users top up
  • 4001 — unsupported file type; show accepted formats
  • 4002 — file too large; suggest compressing or trimming
  • 400 — missing X-Client-Id; generate one and retry
  • 402 — free plan export blocked; not a credit issue, subscription tier
  • 429 — rate limited; wait 30s and retry once

Common Workflows

Quick edit: Upload → "turn my photos and audio into a 60-second promotional video with transitions" → Download MP4. Takes 1-2 minutes for a 30-second clip.

Batch style: Upload multiple files in one session. Process them one by one with different instructions. Each gets its own render.

Iterative: Start with a rough cut, preview the result, then refine. The session keeps your timeline state so you can keep tweaking.

Tips and Tricks

The backend processes faster when you're specific. Instead of "make it look better", try "turn my photos and audio into a 60-second promotional video with transitions" — concrete instructions get better results.

Max file size is 500MB. Stick to MP4, MOV, JPG, PNG for the smoothest experience.

Export as MP4 for widest compatibility across YouTube, Instagram, and TikTok.

Comments

Loading comments...