home / skills / gtmagents / gtm-agents / nurture-testing
This skill guides planning, executing, and logging nurture experiments and regression tests to ensure quality, compliance, and measurable outcomes.
npx playbooks add skill gtmagents/gtm-agents --skill nurture-testingReview the files below or copy the command above to add this skill to your agents.
---
name: nurture-testing
description: Use when planning, executing, and logging nurture experiments and regression
tests.
---
# Nurture Testing & QA Skill
## When to Use
- Running experiments (subject lines, timing, offers, personalization) in nurture flows.
- Performing regression tests before go-live or updates.
- Documenting QA evidence for compliance/audits.
## Framework
1. **Test Design** – hypothesis, variable, control, variant, KPI, sample size, duration.
2. **QA Checklist** – links, tokens, segmentation, tracking, device/browser coverage, accessibility.
3. **Evidence Logging** – screenshots, seed inbox captures, MAP logs, CRM task verification.
4. **Significance & Rollout** – evaluation method (frequentist/Bayesian), rollout criteria, holdback rules.
5. **Regression Cadence** – schedule for periodic audits, triggered when assets/tokens change.
## Templates
- Experiment brief + tracker.
- QA matrix (scenario, expected result, status, owner).
- Evidence archive folder structure.
## Tips
- Always include control groups or holdouts for long-running nurtures.
- Automate reminders to re-run QA after major MAP/CRM updates.
- Pair with `personalization-logic` and `marketing-ops-partner` to catch edge cases early.
---
This skill helps teams plan, execute, and document nurture experiments and regression tests for marketing automation and CRM-driven flows. It standardizes test design, QA checks, evidence capture, and rollout rules so experiments run reliably and audits have clear proof. Use it to reduce risk during updates and to validate personalization and segmentation at scale.
The skill guides you through a five-step framework: test design (hypothesis, variables, KPIs, sample sizing), a QA checklist (links, tokens, segmentation, tracking, accessibility), evidence logging (screenshots, seed inbox captures, MAP/CRM logs), significance evaluation (frequentist or Bayesian) and controlled rollout rules. It also defines a regression cadence for recurring audits and triggers when assets or tokens change. Templates for briefs, QA matrices, and evidence archives make execution repeatable and auditable.
What evidence should I collect for audits?
Collect screenshots, seeded inbox captures, MAP send logs, tracking pixel confirmations, and CRM task or lead records. Store them in a structured evidence archive.
How do I decide between frequentist and Bayesian evaluation?
Choose frequentist methods for traditional hypothesis tests and fixed-horizon experiments; use Bayesian approaches for adaptive experiments or when you want continuous probability updates during rollout.