home / skills / bobmatnyc / claude-mpm-skills / reporting-pipelines
This skill generates and exports timestamped CSV/JSON/markdown reports from GitFlow analytics pipelines, with summaries and post-processing.
npx playbooks add skill bobmatnyc/claude-mpm-skills --skill reporting-pipelinesReview the files below or copy the command above to add this skill to your agents.
---
name: reporting-pipelines
description: Reporting pipelines for CSV/JSON/Markdown exports with timestamped outputs, summaries, and post-processing.
version: 1.0.0
category: universal
author: Claude MPM Team
license: MIT
progressive_disclosure:
entry_point:
summary: "Generate CSV/JSON/markdown reports with timestamped filenames and summary outputs."
when_to_use: "Building reporting flows, exporting analytics results, or standardizing CSV/JSON/markdown outputs across projects."
quick_start: "1. Run the CLI that produces base data 2. Export CSV/JSON/markdown with timestamps 3. Save to reports/"
tags:
- reporting
- csv
- json
- markdown
- analytics
---
# Reporting Pipelines
## Overview
Your reporting pattern is consistent across repos: run a CLI or script that emits structured data, then export CSV/JSON/markdown reports with timestamped filenames into `reports/` or `tests/results/`.
## GitFlow Analytics Pattern
```bash
# Basic run
gitflow-analytics -c config.yaml --weeks 8 --output ./reports
# Explicit analyze + CSV
gitflow-analytics analyze -c config.yaml --weeks 12 --output ./reports --generate-csv
```
Outputs include CSV + markdown narrative reports with date suffixes.
## EDGAR CSV Export Pattern
`edgar/scripts/create_csv_reports.py` reads a JSON results file and emits:
- `executive_compensation_<timestamp>.csv`
- `top_25_executives_<timestamp>.csv`
- `company_summary_<timestamp>.csv`
This script uses pandas for sorting and percentile calculations.
## Standard Pipeline Steps
1. **Collect base data** (CLI or JSON artifacts)
2. **Normalize** into rows/records
3. **Export** CSV/JSON/markdown with timestamp suffixes
4. **Summarize** key metrics in stdout
5. **Store** outputs in `reports/` or `tests/results/`
## Naming Conventions
- Use `YYYYMMDD` or `YYYYMMDD_HHMMSS` suffixes
- Keep one output directory per repo (`reports/` or `tests/results/`)
- Prefer explicit prefixes (e.g., `narrative_report_`, `comprehensive_export_`)
## Troubleshooting
- **Missing output**: ensure output directory exists and is writable.
- **Large CSVs**: filter or aggregate before export; keep summary CSVs for quick review.
## Related Skills
- `universal/data/sec-edgar-pipeline`
- `toolchains/universal/infrastructure/github-actions`
This skill provides a set of reporting pipelines that export structured project outputs as CSV, JSON, and Markdown files with timestamped filenames, summaries, and optional post-processing. It standardizes a simple flow: collect, normalize, export, summarize, and store outputs in a single reports directory. The goal is repeatable, timestamped exports suitable for auditing and downstream processing.
Pipelines ingest CLI or JSON artifacts, normalize data into rows/records (often using pandas), and write exports to a designated output folder with YYYYMMDD or YYYYMMDD_HHMMSS suffixes. Each run emits CSV/JSON/Markdown files plus a short stdout summary and supports optional post-processing steps like sorting, percentile calculations, and aggregation. Filenames use explicit prefixes (e.g., narrative_report_, executive_compensation_) to make outputs discoverable.
What timestamp formats should I use?
Prefer YYYYMMDD or YYYYMMDD_HHMMSS so files sort chronologically and remain filesystem-friendly.
Where should I store outputs in my repo?
Use a single, dedicated folder such as reports/ or tests/results/ to keep exports organized and discoverable.