Ml Ops

v1.0.0

Deep MLOps workflow—reproducible training, experiment tracking, packaging, deployment, monitoring (drift, performance), governance, and rollback for ML. Use...

0· 114·0 current·0 all-time
Security Scan
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
Name, description, and content consistently describe an MLOps workflow; no unexpected binaries, env vars, or config paths are requested.
Instruction Scope
SKILL.md contains advisory steps and checklists for MLOps stages only — it does not instruct the agent to read files, access environment variables, call external endpoints, or run commands.
Install Mechanism
No install spec or code files are present; this is instruction-only so nothing will be written to disk or downloaded during install.
Credentials
The skill declares no required environment variables, credentials, or config paths; requested access is proportional (none) to its advisory purpose.
Persistence & Privilege
always is false and model invocation is allowed (default); the skill is user-invocable and does not request persistent installation or elevated privileges.
Assessment
This skill is a pure guidance document about MLOps and appears internally coherent. It does not request secrets or install code, so installing it has low technical risk. However, if you or an agent later use this guidance to wire up real systems (artifact registries, monitoring, feature stores, cloud deploys), those integrations will require credentials and privileged access — evaluate each connector (CI/CD, cloud accounts, feature stores, monitoring hooks) for least privilege, audit logging, and secret handling before granting them. If a future version adds tooling or install steps, re-evaluate for downloads, unexpected URLs, or broad env var requirements.

Like a lobster shell, security has layers — review code before you run it.

latestvk978yyj7jjvhyb0ptacg2wg7pn83j1zr
114downloads
0stars
1versions
Updated 3w ago
v1.0.0
MIT-0

MLOps (Deep Workflow)

MLOps connects research velocity to production reliability: version data, code, and artifacts together; monitor behavior after deploy.

When to Offer This Workflow

Trigger conditions:

  • First production model; batch or online serving
  • Drift, bias, or latency SLO misses
  • Compliance needs for lineage and explainability

Initial offer:

Use six stages: (1) problem & risk class, (2) data & reproducibility, (3) training & evaluation, (4) packaging & deployment, (5) monitoring & feedback, (6) governance & rollback). Confirm batch vs real-time and regulatory tier.


Stage 1: Problem & Risk Class

Goal: Align ML to decision risk (credit, health vs recommendation).

Exit condition: Offline and online success metrics defined.


Stage 2: Data & Reproducibility

Goal: Snapshot training data; deterministic pipelines; PII handling.

Practices

  • Feature stores optional but valuable for consistency
  • Secrets not in notebooks; orchestrated jobs

Exit condition: Run id reproduces artifact hash within agreed bounds.


Stage 3: Training & Evaluation

Goal: Train/val/test without leakage; time-series splits careful.

Practices

  • Model card with limits and metrics
  • Fairness slices where policy requires

Stage 4: Packaging & Deployment

Goal: Immutable artifacts; canary or shadow before full cutover.

Practices

  • Model + preprocessing code version pinned together

Exit condition: Rollback to previous artifact id documented.


Stage 5: Monitoring & Feedback

Goal: Data drift, concept drift, latency; business KPIs tied to model decisions.

Practices

  • Human review queue for low-confidence predictions when needed

Stage 6: Governance & Rollback

Goal: Approvals for retrain/deploy; audit trail; A/B for big changes.


Final Review Checklist

  • Offline metrics aligned with business risk
  • Data and code reproducibility
  • Packaged artifacts with versioning and rollback
  • Online monitoring and drift strategy
  • Governance and approval path

Tips for Effective Guidance

  • Training-serving skew is a top bug—feature parity tests help.
  • Offline accuracy ≠ online business outcome.
  • Fairness needs explicit slices—not one headline number.

Handling Deviations

  • LLM-heavy products: lean on eval harnesses and prompt versioning (see llm-evaluation).
  • Tiny teams: start with artifact registry + dashboards before a full feature store.

Comments

Loading comments...