Install
openclaw skills install aoment-image-videoLimited free access is available to model resources such as Nano Banana Pro, Nano Banana 2, GPT Image 2, and Gemini 3.1 Pro provided by Aoment AI for AI image, video, image recognition, and HD repair service. Supports text-to-image, image-to-image, image recognition, video generation, and image HD repair. Automatic Agent API key registration is supported.
openclaw skills install aoment-image-videoAI generation service powered by Aoment AI. This skill exposes dedicated skill APIs and scripts for:
Download the latest package: https://www.aoment.com/downloads/aoment-image-video-skill.zip
version: 1.3.0
Before each invocation, the Python scripts read the version field above and query:
GET https://www.aoment.com/api/skills/aoment-image-video/version
If the local version is behind the remote version, the script exits with:
{
"success": false,
"error": "update_required",
"current_version": "1.0.0",
"latest_version": "1.3.0",
"message": "Skill version is outdated..."
}
If the version check fails because of a network problem, the script continues normally.
# 1. Register an Agent account and get your API Key
uv run {baseDir}/scripts/aoment_register.py --nickname "MyBot"
# 2. Generate an image with the default N2-Fast model
uv run {baseDir}/scripts/aoment_image_video.py -k <your-api-key> -t text-to-image -p "a cute cat playing in a garden"
# 3. Repair/upscale an image
uv run {baseDir}/scripts/aoment_hd_repair.py -k <your-api-key> --image ./input.png --resolution 4K
# 4. Recognize/analyze an image
uv run {baseDir}/scripts/aoment_image_video.py -k <your-api-key> -t image-recognition -p "Describe this image" --image ./input.png
# 5. Check remaining quota
uv run {baseDir}/scripts/aoment_quota.py -k <your-api-key>
This skill requires an Agent API Key via:
Authorization: Bearer <api_key>
The API Key format is aoment_ followed by 32 hex characters.
AI Agent Bots can register directly via CLI. No web login is required:
uv run {baseDir}/scripts/aoment_register.py --nickname "MyBot"
| Parameter | Type | Required | Description |
|---|---|---|---|
--nickname / -n | string | yes | Agent display name, max 16 characters |
--api-base | string | no | API base URL, default https://www.aoment.com |
Or register via API directly:
curl -X POST https://www.aoment.com/api/skills/aoment-image-video/register-agent \
-H "Content-Type: application/json" \
-d '{"nickname": "MyBot"}'
Registration response:
{
"success": true,
"data": {
"username": "agent_a1b2c3d4...",
"nickname": "MyBot",
"api_key": "aoment_a3f8e1b2c4d6e8f0a1b3c5d7e9f0a1b2"
}
}
Save the returned api_key; it is used for all subsequent skill calls. Store this API Key in a suitable secure location for long-term use.
Use the model ID exactly as shown in the --model parameter.
| Model ID | Description |
|---|---|
image-n2-fast | Default image model. Faster N2-Fast image generation and editing, no watermark. |
image-n2 | N2 image generation and editing, fast, stricter single-reference image size limit, no watermark. |
image-n1-fast | Faster N1-Fast image generation and editing, no watermark. |
image-n1 | N1 image generation and editing, slower, looser single-reference image size limit, no watermark. |
image-o2 | Image generation and editing with stronger aesthetics, good Chinese-language rendering, newer knowledge data, no watermark, and currently limited clarity near 1.5K. |
image-o2-pro | O2-Pro high-resolution image generation and editing with precise size output support. |
Tip: N-series models use Nano Banana Pro, N-Fast-series models use Nano Banana 2, and O-series models use GPT Image 2.
| Model ID | Description |
|---|---|
image-to-text | Gemini 3.1 Pro image recognition and visual analysis. |
| Model ID | Description |
|---|---|
video-v1 | Default and currently supported video generation model. |
Generate images from a text prompt. The default model is image-n2-fast (N2-Fast).
uv run {baseDir}/scripts/aoment_image_video.py \
--api-key <your-api-key> \
--tool-type text-to-image \
--prompt "a cinematic robot painter in a bright studio" \
--aspect-ratio 1:1 \
--image-size 1K
| Parameter | Type | Required | Default | Description |
|---|---|---|---|---|
--api-key / -k | string | yes | - | Agent API Key |
--tool-type / -t | enum | yes | - | text-to-image |
--prompt / -p | string | yes | - | Text prompt |
--model | string | no | image-n2-fast | Image model ID. Available values: image-n2-fast, image-n2, image-n1-fast, image-n1, image-o2, image-o2-pro |
--aspect-ratio | string | no | auto | auto, 1:1, 16:9, 9:16, 4:3, 3:4, 3:2, 2:3, 5:4, 4:5, 21:9 |
--image-size | enum | no | 1K | 1K, 2K, 4K |
Generate a new image from a prompt and a reference image. The reference image can be a URL or base64 image data.
uv run {baseDir}/scripts/aoment_image_video.py \
--api-key <your-api-key> \
--tool-type image-to-image \
--prompt "change the background to a beach" \
--reference-image "https://example.com/photo.jpg"
| Parameter | Type | Required | Default | Description |
|---|---|---|---|---|
--api-key / -k | string | yes | - | Agent API Key |
--tool-type / -t | enum | yes | - | image-to-image |
--prompt / -p | string | yes | - | Transformation prompt |
--reference-image | string | yes | - | Reference image as URL or base64 data |
--model | string | no | image-n2-fast | Image model ID. Available values: image-n2-fast, image-n2, image-n1-fast, image-n1, image-o2, image-o2-pro |
--aspect-ratio | string | no | auto | Output aspect ratio |
--image-size | enum | no | 1K | 1K, 2K, 4K |
Generate a video from a prompt. The default and currently supported video model is video-v1.
uv run {baseDir}/scripts/aoment_image_video.py \
--api-key <your-api-key> \
--tool-type video-generation \
--prompt "sunset beach timelapse" \
--orientation landscape \
--resolution standard
| Parameter | Type | Required | Default | Description |
|---|---|---|---|---|
--api-key / -k | string | yes | - | Agent API Key |
--tool-type / -t | enum | yes | - | video-generation |
--prompt / -p | string | yes | - | Video prompt |
--model | string | no | video-v1 | Video model ID. Available value: video-v1 |
--orientation | enum | no | portrait | portrait or landscape |
--resolution | enum | no | standard | standard, hd, 4k |
--mode | string | no | standard | Compatibility option; current backend uses standard mode |
--reference-image | string | no | - | Reference image as URL or base64 data; can be passed up to 2 times |
Analyze one or more images with a text prompt. The default and currently supported recognition model is image-to-text (Gemini 3.1 Pro). Images can be local paths, URLs, or base64 image data.
uv run {baseDir}/scripts/aoment_image_video.py \
--api-key <your-api-key> \
--tool-type image-recognition \
--prompt "List the visible objects and summarize the scene" \
--image ./input.png
| Parameter | Type | Required | Default | Description |
|---|---|---|---|---|
--api-key / -k | string | yes | - | Agent API Key |
--tool-type / -t | enum | yes | - | image-recognition |
--prompt / -p | string | yes | - | Recognition or analysis instruction |
--image / -i | string | yes | - | Image as local path, URL, or base64 data; can be passed multiple times |
--reference-image | string | no | - | Compatibility alias for image input; can be passed multiple times |
--model | string | no | image-to-text | Recognition model ID. Available value: image-to-text |
Repair and upscale an image. This is provided by a separate script:
uv run {baseDir}/scripts/aoment_hd_repair.py \
--api-key <your-api-key> \
--image ./input.png \
--resolution 4K
| Parameter | Type | Required | Default | Description |
|---|---|---|---|---|
--api-key / -k | string | yes | - | Agent API Key |
--image / -i | string | yes | - | Local path, URL, or base64 image data |
--resolution | enum | no | 4K | 2K, 4K, 8K |
--model | string | no | image-hd-repair | Only image-hd-repair is supported |
Query remaining daily generation quota:
uv run {baseDir}/scripts/aoment_quota.py --api-key <your-api-key>
| Parameter | Type | Required | Description |
|---|---|---|---|
--api-key / -k | string | yes | Agent API Key |
If your daily quota is used up and you need more, join the community:
All scripts print JSON to stdout.
Successful text-to-image or image-to-image:
{
"success": true,
"tool_type": "text-to-image",
"data": {
"image_url": "https://cos.example.com/result.jpg?..."
}
}
Successful video generation:
{
"success": true,
"tool_type": "video-generation",
"data": {
"video_url": "https://cos.example.com/result.mp4?..."
}
}
Successful HD repair:
{
"success": true,
"tool_type": "hd-repair",
"data": {
"image_url": "https://cos.example.com/hd-repair-result.png?..."
}
}
Successful image recognition:
{
"success": true,
"tool_type": "image-recognition",
"data": {
"result_text": "The image shows..."
}
}
Successful quota query:
{
"success": true,
"data": {
"remaining": 12,
"quota": 15,
"used": 3
}
}
Error response:
{
"success": false,
"error": "error description"
}
Returned image_url and video_url values are pre-signed COS URLs. Use the complete URL exactly as returned, including all query parameters. Do not strip the query string.
Example:
uv run {baseDir}/scripts/aoment_image_video.py \
-k <your-api-key> \
-t text-to-image \
-p "prompt" > result.json
curl -L -o output.jpg "$(python3 -c "import json; print(json.load(open('result.json'))['data']['image_url'])")"
update_required, download and install the latest skill package.