home / skills / vadimcomanescu / codex-skills / senior-computer-vision
This skill helps design, evaluate, and deploy computer vision systems by guiding task definition, dataset strategy, and performance-oriented workflows.
npx playbooks add skill vadimcomanescu/codex-skills --skill senior-computer-visionReview the files below or copy the command above to add this skill to your agents.
---
name: senior-computer-vision
description: "Computer vision engineering workflow for dataset design, model selection (detection/segmentation/classification), evaluation, inference optimization, and deployment. Use when planning or reviewing CV systems, auditing datasets, defining metrics/splits, or diagnosing model/inference issues."
---
# Senior Computer Vision
Build CV systems that generalize, are measurable, and are deployable.
## Quick Start
1) Specify the task precisely: classification vs detection vs segmentation; latency and target hardware.
2) Dataset first: define label taxonomy, edge cases, split strategy, and evaluation metrics.
3) Train with discipline: baselines, ablations, and error analysis (not just “more epochs”).
4) Deploy with realism: preprocessing parity, batching, quantization/trt where needed, monitoring.
## Optional tool: dataset inventory (no ML deps)
For a directory like `data/train/<class>/...` or any image folder:
```bash
python ~/.codex/skills/senior-computer-vision/scripts/dataset_inventory.py data/ --out /tmp/dataset_report.json
```
## References
- Metrics and splits: `references/metrics-and-splits.md`
This skill provides a practical, end-to-end engineering workflow for building robust computer vision systems. It covers dataset design, model selection across classification/detection/segmentation, evaluation strategies, and deployment optimizations. The guidance is focused on measurable outcomes, generalization, and realistic production constraints.
The skill inspects and formalizes the problem (task type, latency, target hardware) and enforces a dataset-first approach: defining taxonomies, edge cases, splits, and metrics. It prescribes disciplined training practices (baselines, ablations, error analysis) and checks inference parity during deployment using techniques like batching, quantization, and TensorRT where appropriate. An optional dataset inventory script creates a lightweight report of folder-structured image datasets to speed audits without ML dependencies.
Does this skill include training code or pretrained models?
It provides workflow guidance and utilities for dataset inspection; it does not bundle heavyweight training code or pretrained models.
Can I use the dataset inventory on non-folder datasets (COCO, TFRecord)?
The included script targets folder-structured image datasets. For COCO/TFRecord, use a small custom adapter or export a folder view for inspection.