home / skills / anton-abyzov / specweave / data-scientist

data-scientist skill

/plugins/specweave-ml/skills/data-scientist

This skill helps you design and analyze experiments, forecast trends, and optimize business outcomes through robust statistical modeling and data analytics.

This is most likely a fork of the sw-data-scientist skill from openclaw
npx playbooks add skill anton-abyzov/specweave --skill data-scientist

Review the files below or copy the command above to add this skill to your agents.

Files (1)
SKILL.md
427 B
---
name: data-scientist
description: Statistical modeling - A/B testing, causal inference, customer analytics (CLV, churn), time series forecasting. Use for business analytics or experiment design.
model: opus
context: fork
---

## ⚠️ Chunking Rule

Large analyses (EDA + modeling + visualization) = 800+ lines. Generate ONE phase per response: EDA → Feature Engineering → Modeling → Evaluation → Recommendations.

Overview

This skill provides statistical modeling and business analytics expertise for A/B testing, causal inference, customer lifetime value (CLV) and churn modeling, and time series forecasting. It is designed to produce reproducible analyses and production-ready code in TypeScript, integrating with AI-assisted developer workflows. The skill emphasizes clear project phases and one-phase-per-response delivery for large analyses.

How this skill works

The skill inspects datasets, experiment designs, and business questions to recommend appropriate statistical methods and produce TypeScript implementations, tests, and documentation. For experiments it runs power calculations, randomization checks, and treatment effect estimation; for customer analytics it implements CLV, churn risk models, and cohort analyses; for time series it fits forecasts with diagnostics. Outputs include model specifications, evaluation metrics, code snippets, and recommended next steps.

When to use it

  • Designing or analyzing A/B experiments and estimating treatment effects
  • Estimating causal effects from observational data with appropriate controls
  • Calculating CLV or building churn prediction pipelines for customer analytics
  • Building time series forecasts for demand, revenue, or KPIs
  • Converting analytic results into production-ready TypeScript code and tests

Best practices

  • Follow the chunking rule: for large analyses generate one phase per response (EDA, feature engineering, modeling, evaluation, recommendations)
  • Start with clear business questions and pre-registered hypotheses for experiments
  • Validate assumptions: balance checks, overlap, stationarity, and model diagnostics
  • Include reproducible tests and documentation with every model or pipeline change
  • Prefer interpretable baselines before complex models; quantify uplift and uncertainty

Example use cases

  • Run power analysis and analyze an A/B test to estimate lift on conversion rates with confidence intervals
  • Perform causal inference using propensity score methods to estimate incremental revenue from a marketing campaign
  • Build a CLV model and deploy a TypeScript scoring function with unit tests for real-time segmentation
  • Train a churn prediction model, produce feature importances, and recommend retention actions
  • Forecast monthly demand using ARIMA/ETS or modern state-space models and produce forecast intervals

FAQ

Can this skill produce production-ready TypeScript code?

Yes. It generates TypeScript implementations, tests, and documentation suitable for integration into CI/CD pipelines.

How does the chunking rule affect workflow?

For analyses expected to be 800+ lines, responses are limited to a single phase (e.g., EDA) to keep outputs focused, reviewable, and iterative.