home / skills / jeremylongshore / claude-code-plugins-plus-skills / detecting-performance-regressions

This skill automatically detects performance regressions in CI/CD by comparing current metrics to baselines and reporting notable degradations.

npx playbooks add skill jeremylongshore/claude-code-plugins-plus-skills --skill detecting-performance-regressions

Review the files below or copy the command above to add this skill to your agents.

Files (7)
SKILL.md
4.1 KB
---
name: detecting-performance-regressions
description: Automatically detect performance regressions in CI/CD pipelines by comparing metrics against baselines. Use when validating builds or analyzing performance trends. Trigger with phrases like "detect performance regression", "compare performance metrics", or "analyze performance degradation".
version: 1.0.0
allowed-tools: "Read, Write, Edit, Grep, Glob, Bash(ci:*), Bash(metrics:*), Bash(testing:*)"
license: MIT
author: Jeremy Longshore <[email protected]>
---
# Performance Regression Detector

This skill provides automated assistance for performance regression detector tasks.

## Overview

This skill automates the detection of performance regressions within a CI/CD pipeline. It utilizes various methods, including baseline comparison, statistical analysis, and threshold violation checks, to identify performance degradation. The skill provides insights into potential performance bottlenecks and helps maintain application performance.

## How It Works

1. **Analyze Performance Data**: The plugin gathers performance metrics from the CI/CD environment.
2. **Detect Regressions**: It employs methods like baseline comparison, statistical analysis, and threshold checks to detect regressions.
3. **Report Findings**: The plugin generates a report summarizing the detected performance regressions and their potential impact.

## When to Use This Skill

This skill activates when you need to:
- Identify performance regressions in a CI/CD pipeline.
- Analyze performance metrics for potential degradation.
- Compare current performance against historical baselines.

## Examples

### Example 1: Identifying a Response Time Regression

User request: "Detect performance regressions in the latest build. Specifically, check for increases in response time."

The skill will:
1. Analyze response time metrics from the latest build.
2. Compare the response times against a historical baseline.
3. Report any statistically significant increases in response time that exceed a defined threshold.

### Example 2: Detecting Throughput Degradation

User request: "Analyze throughput for performance regressions after the recent code merge."

The skill will:
1. Gather throughput data (requests per second) from the post-merge CI/CD run.
2. Compare the throughput to pre-merge values, looking for statistically significant drops.
3. Generate a report highlighting any throughput degradation, indicating a potential performance regression.

## Best Practices

- **Define Baselines**: Establish clear and representative performance baselines for accurate comparison.
- **Set Thresholds**: Configure appropriate thresholds for identifying significant performance regressions.
- **Monitor Key Metrics**: Focus on monitoring critical performance metrics relevant to the application's behavior.

## Integration

This skill can be integrated with other CI/CD tools to automatically trigger regression detection upon new builds or code merges. It can also be combined with reporting plugins to generate detailed performance reports.

## Prerequisites

- Historical performance baselines in {baseDir}/performance/baselines/
- Access to CI/CD performance metrics
- Statistical analysis tools
- Defined regression thresholds

## Instructions

1. Collect performance metrics from current build
2. Load historical baseline data
3. Apply statistical analysis to detect significant changes
4. Check for threshold violations
5. Identify specific regressed metrics
6. Generate regression report with root cause analysis

## Output

- Performance regression detection report
- Statistical comparison with baselines
- List of regressed metrics with severity
- Visualization of performance trends
- Recommendations for investigation

## Error Handling

If regression detection fails:
- Verify baseline data availability
- Check metrics collection configuration
- Validate statistical analysis parameters
- Ensure threshold definitions are valid
- Review CI/CD integration setup

## Resources

- Statistical process control for performance testing
- CI/CD performance testing best practices
- Regression detection algorithms
- Performance monitoring strategies

Overview

This skill automates detection of performance regressions in CI/CD pipelines by comparing current metrics to historical baselines and threshold rules. It flags statistically significant degradations, summarizes severity, and produces actionable reports to speed investigation. Use it to validate builds and monitor performance trends automatically.

How this skill works

The skill ingests performance metrics from the CI/CD run and loads representative historical baselines. It applies statistical comparisons, significance tests, and configurable threshold checks to identify deviations that indicate regressions. Results are compiled into a report listing regressed metrics, severity, trend visualizations, and investigation recommendations.

When to use it

  • Validate new builds before deployment to catch performance regressions early
  • Analyze performance impact after code merges or configuration changes
  • Monitor long-term trends and detect gradual degradation
  • Trigger automated alerts in CI pipelines when key metrics cross thresholds
  • Prioritize performance investigations by severity and statistical significance

Best practices

  • Define representative baselines covering typical workloads and release phases
  • Set realistic, metric-specific thresholds and allow for configurable sensitivity
  • Focus on a small set of key metrics (response time, throughput, error rate, resource utilization)
  • Run statistical tests over sufficient sample sizes to avoid false positives
  • Integrate detection into CI pipelines and link reports to issue trackers for traceability

Example use cases

  • Detect increased API response times in the latest build by comparing latency distributions to baseline
  • Identify throughput drops after a merge by analyzing requests-per-second against historical runs
  • Spot rising error rates or CPU usage trends that may indicate emerging bottlenecks
  • Automatically fail a CI job when critical metrics breach configured regression thresholds
  • Generate a regression report with visual trend lines and ranked recommendations for triage

FAQ

What inputs are required to run detection?

You need current CI performance metrics, historical baseline files, and configured thresholds or sensitivity settings.

How does it avoid false positives from noisy data?

It uses statistical significance tests and configurable thresholds, and recommends using sufficient sample sizes and representative baselines to reduce noise-driven alerts.