home / skills / sickn33 / antigravity-awesome-skills / data-quality-frameworks
This skill helps automate data quality validation with Great Expectations, dbt tests, and data contracts across pipelines, CI/CD, and monitoring.
npx playbooks add skill sickn33/antigravity-awesome-skills --skill data-quality-frameworksReview the files below or copy the command above to add this skill to your agents.
---
name: data-quality-frameworks
description: Implement data quality validation with Great Expectations, dbt tests, and data contracts. Use when building data quality pipelines, implementing validation rules, or establishing data contracts.
---
# Data Quality Frameworks
Production patterns for implementing data quality with Great Expectations, dbt tests, and data contracts to ensure reliable data pipelines.
## Use this skill when
- Implementing data quality checks in pipelines
- Setting up Great Expectations validation
- Building comprehensive dbt test suites
- Establishing data contracts between teams
- Monitoring data quality metrics
- Automating data validation in CI/CD
## Do not use this skill when
- The data sources are undefined or unavailable
- You cannot modify validation rules or schemas
- The task is unrelated to data quality or contracts
## Instructions
- Identify critical datasets and quality dimensions.
- Define expectations/tests and contract rules.
- Automate validation in CI/CD and schedule checks.
- Set alerting, ownership, and remediation steps.
- If detailed patterns are required, open `resources/implementation-playbook.md`.
## Safety
- Avoid blocking critical pipelines without a fallback plan.
- Handle sensitive data securely in validation outputs.
## Resources
- `resources/implementation-playbook.md` for detailed frameworks, templates, and examples.
This skill implements production-ready data quality validation using Great Expectations, dbt tests, and formal data contracts. It provides patterns to identify critical datasets, define expectations and tests, and automate validation within CI/CD and scheduled pipelines. The goal is reliable data pipelines with clear ownership, alerting, and remediation steps.
The skill inspects schema, completeness, uniqueness, value distributions, and business rules across datasets and surfaces failures through validation suites. It codifies checks as Great Expectations expectations, dbt test cases, and contract rules that can be enforced at ingestion, transformation, and delivery points. Validation runs are automated in CI/CD pipelines and schedulers, and failures trigger alerts and defined remediation workflows. It also supports metrics collection for monitoring data quality trends.
Can I run these checks on streaming data?
Yes. Use lightweight expectations on micro-batches or windowed jobs and stream checkpoints into your validation system; avoid heavy aggregations in low-latency paths.
How do I avoid noisy alerts from transient failures?
Tune thresholds, require consecutive failures before alerting, add guardrails like minimum sample sizes, and implement backoff/retry policies before firing pager alerts.