home / skills / anton-abyzov / specweave / ml-engineer
/plugins/specweave-ml/skills/ml-engineer
This skill helps build and optimize ML pipelines by enforcing best practices, experiment tracking, cross-validation, and explainability throughout stages.
npx playbooks add skill anton-abyzov/specweave --skill ml-engineerReview the files below or copy the command above to add this skill to your agents.
---
name: ml-engineer
description: ML system builder enforcing best practices - baseline comparison, cross-validation, experiment tracking, explainability (SHAP/LIME). Use for ML pipelines, model training, production ML.
model: opus
context: fork
---
# ML Engineer Agent
## ⚠️ Chunking Rule
Large ML pipelines = 1000+ lines. Generate ONE stage per response: Data/EDA → Features → Training → Evaluation → Deployment.
This skill builds ML systems that enforce engineering best practices for robust, production-ready models. It guides pipeline construction, baseline comparison, cross-validation, experiment tracking, and model explainability using SHAP/LIME. The implementation targets TypeScript environments and integrates with CI/CD and developer tooling.
The agent inspects project artifacts and generates one pipeline stage per response: Data/EDA → Features → Training → Evaluation → Deployment, keeping large pipelines manageable. For each stage it emits concrete specs, tests, and TypeScript scaffold code, plus configuration for experiment tracking and model explainability hooks. It validates choices against baseline models and cross-validation protocols, and adds CI/CD and documentation snippets suited to production ML.
How does the one-stage-per-response rule affect workflow?
It forces small, reviewable increments: request a specific stage and receive focused specs, code, and tests for that stage only.
Which explainability tools are supported?
The workflow includes SHAP and LIME integrations for feature-level explanations and exports explainability artifacts alongside experiment logs.