Install
openclaw skills install video-learn-onlineSkip the learning curve of professional editing software. Describe what you want — break this lecture into short topic chapters with subtitles and a summary...
openclaw skills install video-learn-onlineGot educational video clips to work with? Send it over and tell me what you need — I'll take care of the AI learning video creation.
Try saying:
Before handling any user request, establish a connection to the backend API. Show a brief status like "Connecting...".
If NEMO_TOKEN is in the environment, use it directly and create a session. Otherwise, acquire a free starter token:
https://mega-api-prod.nemovideo.ai/api/auth/anonymous-token with the X-Client-Id headertoken with 100 free credits valid for 7 days — use it as NEMO_TOKENThen create a session by POSTing to https://mega-api-prod.nemovideo.ai/api/tasks/me/with-session/nemo_agent with Bearer authorization and body {"task_name":"project","language":"en"}. The session_id in the response is needed for all following requests.
Tell the user you're ready. Keep the technical details out of the chat.
This tool takes your educational video clips and runs AI learning video creation through a cloud rendering pipeline. You upload, describe what you want, and download the result.
Say you have a 10-minute lecture recording and want to break this lecture into short topic chapters with subtitles and a summary overlay — the backend processes it in about 1-2 minutes and hands you a 1080p MP4.
Tip: splitting a long lecture into 3-5 minute segments improves viewer retention and processes faster.
User prompts referencing video learn online, aspect ratio, text overlays, or audio tracks get routed to the corresponding action via keyword and intent classification.
| User says... | Action | Skip SSE? |
|---|---|---|
| "export" / "导出" / "download" / "send me the video" | → §3.5 Export | ✅ |
| "credits" / "积分" / "balance" / "余额" | → §3.3 Credits | ✅ |
| "status" / "状态" / "show tracks" | → §3.4 State | ✅ |
| "upload" / "上传" / user sends file | → §3.2 Upload | ✅ |
| Everything else (generate, edit, add BGM…) | → §3.1 SSE | ❌ |
Each export job queues on a cloud GPU node that composites video layers, applies platform-spec compression (H.264, up to 1080x1920), and returns a download URL within 30-90 seconds. The session token carries render job IDs, so closing the tab before completion orphans the job.
Include Authorization: Bearer <NEMO_TOKEN> and all attribution headers on every request — omitting them triggers a 402 on export.
Three attribution headers are required on every request and must match this file's frontmatter:
| Header | Value |
|---|---|
X-Skill-Source | video-learn-online |
X-Skill-Version | frontmatter version |
X-Skill-Platform | auto-detect: clawhub / cursor / unknown from install path |
API base: https://mega-api-prod.nemovideo.ai
Create session: POST /api/tasks/me/with-session/nemo_agent — body {"task_name":"project","language":"<lang>"} — returns task_id, session_id.
Send message (SSE): POST /run_sse — body {"app_name":"nemo_agent","user_id":"me","session_id":"<sid>","new_message":{"parts":[{"text":"<msg>"}]}} with Accept: text/event-stream. Max timeout: 15 minutes.
Upload: POST /api/upload-video/nemo_agent/me/<sid> — file: multipart -F "files=@/path", or URL: {"urls":["<url>"],"source_type":"url"}
Credits: GET /api/credits/balance/simple — returns available, frozen, total
Session state: GET /api/state/nemo_agent/me/<sid>/latest — key fields: data.state.draft, data.state.video_infos, data.state.generated_media
Export (free, no credits): POST /api/render/proxy/lambda — body {"id":"render_<ts>","sessionId":"<sid>","draft":<json>,"output":{"format":"mp4","quality":"high"}}. Poll GET /api/render/proxy/lambda/<id> every 30s until status = completed. Download URL at output.url.
Supported formats: mp4, mov, avi, webm, mkv, jpg, png, gif, webp, mp3, wav, m4a, aac.
0 — success, continue normally1001 — token expired or invalid; re-acquire via /api/auth/anonymous-token1002 — session not found; create a new one2001 — out of credits; anonymous users get a registration link with ?bind=<id>, registered users top up4001 — unsupported file type; show accepted formats4002 — file too large; suggest compressing or trimming400 — missing X-Client-Id; generate one and retry402 — free plan export blocked; not a credit issue, subscription tier429 — rate limited; wait 30s and retry onceThe backend responds as if there's a visual interface. Map its instructions to API calls:
| Event | Action |
|---|---|
| Text response | Apply GUI translation (§4), present to user |
| Tool call/result | Process internally, don't forward |
heartbeat / empty data: | Keep waiting. Every 2 min: "⏳ Still working..." |
| Stream closes | Process final response |
~30% of editing operations return no text in the SSE stream. When this happens: poll session state to verify the edit was applied, then summarize changes to the user.
Draft field mapping: t=tracks, tt=track type (0=video, 1=audio, 7=text), sg=segments, d=duration(ms), m=metadata.
Timeline (3 tracks): 1. Video: city timelapse (0-10s) 2. BGM: Lo-fi (0-10s, 35%) 3. Title: "Urban Dreams" (0-3s)
The backend processes faster when you're specific. Instead of "make it look better", try "break this lecture into short topic chapters with subtitles and a summary overlay" — concrete instructions get better results.
Max file size is 500MB. Stick to MP4, MOV, WebM, AVI for the smoothest experience.
Export as MP4 for widest compatibility with learning platforms like Moodle and Teachable.
Quick edit: Upload → "break this lecture into short topic chapters with subtitles and a summary overlay" → Download MP4. Takes 1-2 minutes for a 30-second clip.
Batch style: Upload multiple files in one session. Process them one by one with different instructions. Each gets its own render.
Iterative: Start with a rough cut, preview the result, then refine. The session keeps your timeline state so you can keep tweaking.