home / skills / athola / claude-night-market / math-review

This skill verifies mathematical code for correctness and numerical stability, aligning implementations with formal standards to ensure reliable, auditable

npx playbooks add skill athola/claude-night-market --skill math-review

Review the files below or copy the command above to add this skill to your agents.

Files (5)
SKILL.md
5.2 KB
---
name: math-review
description: 'Use this skill for mathematical code verification. Use when reviewing
  math-heavy code, verifying algorithm correctness, checking numerical stability,
  aligning with mathematical standards. Do not use when general algorithm review -
  use architecture-review. DO NOT use when: performance optimization - use parseltongue:python-performance.'
category: specialized
tags:
- math
- algorithms
- numerical
- stability
- verification
- scientific
tools:
- derivation-checker
- stability-analyzer
- reference-finder
usage_patterns:
- algorithm-review
- numerical-analysis
- derivation-verification
- stability-assessment
complexity: advanced
estimated_tokens: 200
progressive_loading: true
dependencies:
- pensive:shared
- imbue:evidence-logging
---
## Table of Contents

- [Quick Start](#quick-start)
- [When to Use](#when-to-use)
- [Required TodoWrite Items](#required-todowrite-items)
- [Core Workflow](#core-workflow)
- [1. Context Sync](#1-context-sync)
- [2. Requirements Mapping](#2-requirements-mapping)
- [3. Derivation Verification](#3-derivation-verification)
- [4. Stability Assessment](#4-stability-assessment)
- [5. Evidence Logging](#5-evidence-logging)
- [Progressive Loading](#progressive-loading)
- [Essential Checklist](#essential-checklist)
- [Output Format](#output-format)
- [Summary](#summary)
- [Context](#context)
- [Requirements Analysis](#requirements-analysis)
- [Derivation Review](#derivation-review)
- [Stability Analysis](#stability-analysis)
- [Issues](#issues)
- [Recommendation](#recommendation)
- [Exit Criteria](#exit-criteria)


# Mathematical Algorithm Review

Intensive analysis ensuring numerical stability and alignment with standards.

## Quick Start

```bash
/math-review
```
**Verification:** Run the command with `--help` flag to verify availability.

## When To Use

- Changes to mathematical models or algorithms
- Statistical routines or probabilistic logic
- Numerical integration or optimization
- Scientific computing code
- ML/AI model implementations
- Safety-critical calculations

## When NOT To Use

- General algorithm review -
  use architecture-review
- Performance optimization - use parseltongue:python-performance
- General algorithm review -
  use architecture-review
- Performance optimization - use parseltongue:python-performance

## Required TodoWrite Items

1. `math-review:context-synced`
2. `math-review:requirements-mapped`
3. `math-review:derivations-verified`
4. `math-review:stability-assessed`
5. `math-review:evidence-logged`

## Core Workflow

### 1. Context Sync
```bash
pwd && git status -sb && git diff --stat origin/main..HEAD
```
**Verification:** Run `git status` to confirm working tree state.
Enumerate math-heavy files (source, tests, docs, notebooks). Classify risk: safety-critical, financial, ML fairness.

### 2. Requirements Mapping
Translate requirements → mathematical invariants. Document pre/post conditions, conservation laws, bounds. **Load**: `modules/requirements-mapping.md`

### 3. Derivation Verification
Re-derive formulas using CAS. Challenge approximations. Cite authoritative standards (NASA-STD-7009, ASME VVUQ). **Load**: `modules/derivation-verification.md`

### 4. Stability Assessment
Evaluate conditioning, precision, scaling, randomness. Compare complexity. Quantify uncertainty. **Load**: `modules/numerical-stability.md`

### 5. Evidence Logging
```bash
pytest tests/math/ --benchmark
jupyter nbconvert --execute derivation.ipynb
```
**Verification:** Run `pytest -v tests/math/` to verify.
Log deviations, recommend: Approve / Approve with actions / Block. **Load**: `modules/testing-strategies.md`

## Progressive Loading

**Default (200 tokens)**: Core workflow, checklists
**+Requirements** (+300 tokens): Invariants, pre/post conditions, coverage analysis
**+Derivation** (+350 tokens): CAS verification, standards, citations
**+Stability** (+400 tokens): Numerical properties, precision, complexity
**+Testing** (+350 tokens): Edge cases, benchmarks, reproducibility

**Total with all modules**: ~1600 tokens

## Essential Checklist

**Correctness**: Formulas match spec | Edge cases handled | Units consistent | Domain enforced
**Stability**: Condition number OK | Precision sufficient | No cancellation | Overflow prevented
**Verification**: Derivations documented | References cited | Tests cover invariants | Benchmarks reproducible
**Documentation**: Assumptions stated | Limitations documented | Error bounds specified | References linked

## Output Format

```markdown
## Summary
[Brief findings]

## Context
Files | Risk classification | Standards

## Requirements Analysis
| Invariant | Verified | Evidence |

## Derivation Review
[Status and conflicts]

## Stability Analysis
Condition number | Precision | Risks

## Issues
[M1] [Title]: Location | Issue | Fix

## Recommendation
Approve / Approve with actions / Block
```
**Verification:** Run the command with `--help` flag to verify availability.

## Exit Criteria

- Context synced, requirements mapped, derivations verified, stability assessed, evidence logged with citations
## Troubleshooting

### Common Issues

**Command not found**
Ensure all dependencies are installed and in PATH

**Permission errors**
Check file permissions and run with appropriate privileges

**Unexpected behavior**
Enable verbose logging with `--verbose` flag

Overview

This skill verifies math-heavy code for correctness and numerical stability. It focuses on aligning implementations with mathematical specifications, documenting derivations, and producing evidence-backed recommendations for approval or remediation.

How this skill works

The workflow syncs repository context, maps requirements to mathematical invariants, and re-derives key formulas using CAS or notebook execution. It then assesses numerical stability (conditioning, precision, scaling), runs targeted tests and benchmarks, and logs evidence with citations and recommended actions.

When to use it

  • When changes touch mathematical models, statistical routines, or scientific computing code
  • When implementing numerical integration, optimization, or linear algebra algorithms
  • When verifying ML/AI model math, loss functions, or probabilistic logic
  • When safety-critical or financial calculations require formal verification
  • When you need documented derivations, error bounds, and reproducible tests

Best practices

  • Start by syncing context: list math files, tests, notebooks, and classify risk
  • Map requirements to explicit invariants, pre/post conditions, and conserved quantities
  • Re-derive formulas with a CAS or executed notebook and cite authoritative standards
  • Quantify conditioning, rounding risks, and uncertainty; include unit and scale checks
  • Log evidence with tests, notebooks, and benchmarks; recommend Approve / Approve with actions / Block

Example use cases

  • Reviewing a new numerical solver for PDEs to confirm convergence criteria and stability
  • Verifying implemented formulas for a financial risk model against regulatory standards
  • Checking ML loss function derivations and numerical stability of gradient computations
  • Auditing a signal processing library to ensure units, scaling, and overflow are handled
  • Validating Monte Carlo or probabilistic code for reproducibility and variance estimation

FAQ

Can this skill be used for general architecture or performance reviews?

No. Use an architecture-focused review for design-level concerns and a dedicated performance skill for optimization tasks.

What outputs should I expect?

A concise report with Summary, Context, Requirements Analysis, Derivation Review, Stability Analysis, Issues, and a clear Recommendation with evidence links.