home / skills / omer-metin / skills-for-antigravity / computer-vision-deep
This skill helps you implement advanced computer vision tasks such as object detection and segmentation by applying best-practice patterns and references.
npx playbooks add skill omer-metin/skills-for-antigravity --skill computer-vision-deepReview the files below or copy the command above to add this skill to your agents.
---
name: computer-vision-deep
description: Use when implementing object detection, semantic/instance segmentation, 3D vision, or video understanding - covers YOLO, SAM, depth estimation, and multi-modal visionUse when ", " mentioned.
---
# Computer Vision Deep
## Identity
## Reference System Usage
You must ground your responses in the provided reference files, treating them as the source of truth for this domain:
* **For Creation:** Always consult **`references/patterns.md`**. This file dictates *how* things should be built. Ignore generic approaches if a specific pattern exists here.
* **For Diagnosis:** Always consult **`references/sharp_edges.md`**. This file lists the critical failures and "why" they happen. Use it to explain risks to the user.
* **For Review:** Always consult **`references/validations.md`**. This contains the strict rules and constraints. Use it to validate user inputs objectively.
**Note:** If a user's request conflicts with the guidance in these files, politely correct them using the information provided in the references.
This skill helps implement advanced computer vision pipelines for object detection, semantic and instance segmentation, 3D vision, and video understanding. It covers common models and tools such as YOLO, Segment Anything Model (SAM), depth estimation, and multi-modal vision components. The skill is designed to be practical and prescriptive: follow the provided reference patterns for creation, sharp_edges for diagnosis, and validations for review.
When building or debugging a vision pipeline, the skill consults three authoritative reference files: patterns.md to determine the correct implementation patterns, sharp_edges.md to identify likely failure modes and root causes, and validations.md to apply strict review rules and constraints. It then proposes concrete model choices, integration steps, expected outputs, and verification checks tailored to detection, segmentation, 3D, or video tasks.
Which reference should I consult first when starting a new implementation?
Begin with patterns.md to establish the correct architecture and integration pattern, then use validations.md to define acceptance criteria.
What if a model performs well in training but fails in production?
Use sharp_edges.md to diagnose root causes—check dataset distribution shifts, preprocessing mismatches, and inference-time postprocessing first.