Vision Tool

v1.1.3

Image recognition using Ollama + qwen3.5:4b with think=False for reliable content extraction.

0· 75·0 current·0 all-time
byRuilizhen Hu@huruilizhen
MIT-0
Download zip
LicenseMIT-0 · Free to use, modify, and redistribute. No attribution required.
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
Name/description match the implementation: the code reads an image, Base64-encodes it, and posts to an Ollama /api/chat endpoint using model qwen3.5:4b. Required binaries (ollama, python3) are appropriate and no unrelated credentials or tools are requested.
Instruction Scope
Runtime instructions only run local Python code and call the Ollama API at http://127.0.0.1:11434/api/chat; they read the provided image file and send its Base64 payload. This is coherent with image-analysis purpose, but be aware that if the Ollama service URL is changed from the default, image data could be sent to a remote host — the code itself does not exfiltrate to external endpoints by default.
Install Mechanism
No install spec that downloads external artifacts; included code is pure Python using the requests library. There are no archive downloads or external installers declared in the skill metadata.
Credentials
The skill declares no required environment variables or credentials. It uses sensible defaults (local Ollama URL). No secret or cloud credentials are requested, which is proportionate for a local-model vision tool.
Persistence & Privilege
always:false and user-invocable:true (defaults) — the skill does not request forced persistent inclusion or elevated platform privileges. It does not modify other skills or system-wide configs.
Assessment
This skill appears to do exactly what it claims: it reads a local image file, Base64-encodes it, and POSTs it to an Ollama /api/chat endpoint on localhost. Before installing or running it, ensure you: 1) run a trusted Ollama instance locally (ollama serve) and have pulled qwen3.5:4b, 2) confirm the Ollama service is not proxying/forwarding requests to an untrusted remote endpoint (if you change the default URL the skill will send images to wherever that URL points), and 3) review and run the included tests in a safe environment. Because the skill does not request secrets or remote installs and the code is readable, there are no incoherent or disproportionate requests — but always verify you trust the Ollama server you will use (local vs remote).

Like a lobster shell, security has layers — review code before you run it.

image-recognitionvk976tkj86h8stff4smj0vwc3ax84q4jclatestvk97e5029kqsh7a8enzamkwftq984qymxollamavk976tkj86h8stff4smj0vwc3ax84q4jcvisionvk976tkj86h8stff4smj0vwc3ax84q4jc

License

MIT-0
Free to use, modify, and redistribute. No attribution required.

Runtime requirements

👁️ Clawdis
Binsollama, python3

Comments