Install
openclaw skills install guoshun-industrial-vision-advisor国顺工业视觉顾问技能。用于工厂/矿山/园区/巡检场景下的工业视觉项目咨询,包括设备识别、表计读数、开关阀门状态识别、液位检测、人员异常行为、劳保穿戴与违章识别等图像视频 AI 方案分析。适用于用户需要判断现场是否适合做视觉 AI、该用 YOLO/RT-DETR、开放词汇检测、SAM、VLM/OCR、关键点、姿态动作识别、跟踪规则,或需要输出 PoC/实施/验收方案时。
openclaw skills install guoshun-industrial-vision-advisor当用户提出工厂、矿山、园区巡检、设备点检、人员安全监管等视觉识别需求时,使用本技能把问题拆解成可执行的技术路线。
核心原则:先定义业务决策和视觉任务,再选择模型。不要一上来就默认“训练 YOLO”或“直接上 VLM”,必须先明确可见性、数据条件、风险边界和验收标准。
Prefer concrete evidence over abstract descriptions. Ask for:
Read references/intake-template.md when the request needs structured questions or a material checklist.
Use this quick map, then read references/task-taxonomy.md for details.
| User asks for | Usually decompose into |
|---|---|
| Find people, vehicles, gauges, switches, valves, devices | Detection plus optional tracking |
| Read pointer/analog gauges | Detection -> keypoints/segmentation -> OCR/config -> geometry |
| Determine switch/valve state | Detection -> keypoints/classification -> device binding rules |
| Detect liquid level | Detection -> segmentation/keypoints -> OCR/config -> measurement |
| PPE/violation recognition | Person/object detection -> tracking -> region/relationship/time rules |
| Abnormal movement/action | Person detection -> tracking -> pose/action model -> time-window rules |
| Smoke, leakage, crack, dirt, spill, boundary | Segmentation/anomaly detection, sometimes thermal/3D/special lighting |
| Unknown or changing target names | Open-vocabulary detection for discovery/auto-labeling, then dedicated model if production use |
| Explain scene, read labels, produce report | VLM/OCR as low-frequency assistant or reviewer |
Use current official docs before finalizing model/API choices because model versions and deployment support change. Read references/toolchain.md for the maintained toolchain summary and source links.
Default production posture:
Read references/guardrails.md for the full red lines. Always enforce these:
Every answer should include, scaled to the request:
Use references/output-template.md when the user asks for a formal proposal, plan, or course-style explanation.
For most production projects:
Site samples and definitions
-> task decomposition
-> camera/lighting feasibility check
-> auto-labeling with open-vocabulary/SAM where useful
-> manual label correction and hard-negative collection
-> train dedicated detector/segmenter/keypoint/action model
-> add tracking, geometry, OCR, and rules
-> VLM only for review/reporting/low-confidence cases
-> offline test on separated data
-> shadow-mode field trial
-> monitored production with sample feedback and retraining
For a new scenario with weak data, output a staged route rather than a final architecture.