home / skills / 0xdarkmatter / claude-mods / testgen

testgen skill

/skills/testgen

This skill generates comprehensive tests by automatically detecting frameworks and project standards, routing to expert agents, and creating well-structured

npx playbooks add skill 0xdarkmatter/claude-mods --skill testgen

Review the files below or copy the command above to add this skill to your agents.

Files (6)
SKILL.md
8.1 KB
---
name: testgen
description: "Generate tests with expert routing, framework detection, and auto-TaskCreate. Triggers on: generate tests, write tests, testgen, create test file, add test coverage."
allowed-tools: "Read Write Edit Bash Glob Grep Task TaskCreate"
---

# TestGen Skill - AI Test Generation

Generate comprehensive tests with automatic framework detection, expert agent routing, and project convention matching.

## Architecture

```
testgen <target> [--type] [--focus] [--depth]
    │
    ├─→ Step 1: Analyze Target
    │     ├─ File exists? → Read and parse
    │     ├─ Function specified? → Extract signature
    │     ├─ Directory? → List source files
    │     └─ Find existing tests (avoid duplicates)
    │
    ├─→ Step 2: Detect Framework (parallel)
    │     ├─ package.json → jest/vitest/mocha/cypress/playwright
    │     ├─ pyproject.toml → pytest/unittest
    │     ├─ go.mod → go test
    │     ├─ Cargo.toml → cargo test
    │     ├─ composer.json → phpunit/pest
    │     └─ Check existing test patterns
    │
    ├─→ Step 3: Load Project Standards
    │     ├─ AGENTS.md, CLAUDE.md conventions
    │     ├─ Existing test file structure
    │     └─ Naming conventions (*.test.ts vs *.spec.ts)
    │
    ├─→ Step 4: Route to Expert Agent
    │     ├─ .ts → typescript-expert
    │     ├─ .tsx/.jsx → react-expert
    │     ├─ .vue → vue-expert
    │     ├─ .py → python-expert
    │     ├─ .go → go-expert
    │     ├─ .rs → rust-expert
    │     ├─ .php → laravel-expert
    │     ├─ E2E/Cypress → cypress-expert
    │     ├─ Playwright → typescript-expert
    │     ├─ --visual → Chrome DevTools MCP
    │     └─ Multi-file → parallel expert dispatch
    │
    ├─→ Step 5: Generate Tests
    │     ├─ Create test file in correct location
    │     ├─ Follow detected conventions
    │     └─ Include: happy path, edge cases, error handling
    │
    └─→ Step 6: Integration
          ├─ Auto-create task (TaskCreate) for verification
          └─ Suggest: run tests, /review, /save
```

## Execution Steps

### Step 1: Analyze Target

```bash
# Check if target exists
test -f "$TARGET" && echo "FILE" || test -d "$TARGET" && echo "DIRECTORY"

# For function-specific: extract signature
command -v ast-grep >/dev/null 2>&1 && ast-grep -p "function $FUNCTION_NAME" "$FILE"

# Fallback to ripgrep
rg "(?:function|const|def|public|private)\s+$FUNCTION_NAME" "$FILE" -A 10
```

**Check for existing tests:**
```bash
fd -e test.ts -e spec.ts -e test.js -e spec.js | rg "$BASENAME"
fd "test_*.py" | rg "$BASENAME"
```

### Step 2: Detect Framework

**JavaScript/TypeScript:**
```bash
cat package.json 2>/dev/null | jq -r '.devDependencies | keys[]' | grep -E 'jest|vitest|mocha|cypress|playwright|@testing-library'
```

**Python:**
```bash
grep -E "pytest|unittest|nose" pyproject.toml setup.py requirements*.txt 2>/dev/null
```

**Go:**
```bash
test -f go.mod && echo "go test available"
```

**Rust:**
```bash
test -f Cargo.toml && echo "cargo test available"
```

**PHP:**
```bash
cat composer.json 2>/dev/null | jq -r '.["require-dev"] | keys[]' | grep -E 'phpunit|pest|codeception'
```

### Step 3: Load Project Standards

```bash
# Claude Code conventions
cat AGENTS.md 2>/dev/null | head -50
cat CLAUDE.md 2>/dev/null | head -50

# Test config files
cat jest.config.* vitest.config.* pytest.ini pyproject.toml 2>/dev/null | head -30
```

**Test location conventions:**
```
# JavaScript
src/utils/helper.ts → src/utils/__tests__/helper.test.ts  # __tests__ folder
                    → src/utils/helper.test.ts            # co-located
                    → tests/utils/helper.test.ts          # separate tests/

# Python
app/utils/helper.py → tests/test_helper.py               # tests/ folder
                    → tests/utils/test_helper.py         # mirror structure

# Go
pkg/auth/token.go → pkg/auth/token_test.go               # co-located (required)

# Rust
src/auth.rs → src/auth.rs (mod tests { ... })            # inline tests
            → tests/auth_test.rs                          # integration tests
```

### Step 4: Route to Expert Agent

| File Pattern | Primary Expert | Secondary |
|--------------|----------------|-----------|
| `*.ts` | typescript-expert | - |
| `*.tsx`, `*.jsx` | react-expert | typescript-expert |
| `*.vue` | vue-expert | typescript-expert |
| `*.py` | python-expert | - |
| `*.go` | go-expert | - |
| `*.rs` | rust-expert | - |
| `*.php` | laravel-expert | - |
| `*.cy.ts`, `cypress/*` | cypress-expert | - |
| `*.spec.ts` (Playwright) | typescript-expert | - |
| `playwright/*`, `e2e/*` | typescript-expert | - |
| `*.sh`, `*.bash` | bash-expert | - |
| (--visual flag) | Chrome DevTools MCP | typescript-expert |

**Invoke via Task tool:**
```
Task tool with subagent_type: "[detected]-expert"
Prompt includes:
  - Source file content
  - Function signatures to test
  - Detected framework and conventions
  - Requested test type and focus
```

### Step 5: Generate Tests

**Test categories based on --focus:**

| Focus | What to Generate |
|-------|------------------|
| `happy` | Normal input, expected output |
| `edge` | Boundary values, empty inputs, nulls |
| `error` | Invalid inputs, exceptions, error handling |
| `all` | All of the above (default) |

**Depth levels:**

| Depth | Coverage |
|-------|----------|
| `quick` | Happy path only, 1-2 tests per function |
| `normal` | Happy + common edge cases (default) |
| `thorough` | Comprehensive: all paths, mocking, async |

### Step 6: Integration

**Auto-create task:**
```
TaskCreate:
  subject: "Run generated tests for src/auth.ts"
  description: "Verify generated tests pass and review edge cases"
  activeForm: "Running generated tests for auth.ts"
```

**Suggest next steps:**
```
Tests generated: src/auth.test.ts

Next steps:
1. Run tests: npm test src/auth.test.ts
2. Review and refine edge cases
3. Use /save to persist tasks across sessions
```

---

## Expert Routing Details

### TypeScript/JavaScript → typescript-expert
- Proper type imports
- Generic type handling
- Async/await patterns
- Mock typing

### React/JSX → react-expert
- React Testing Library patterns
- Component rendering tests
- Hook testing (renderHook)
- Accessibility queries (getByRole)

### Vue → vue-expert
- Vue Test Utils patterns
- Composition API testing
- Pinia store mocking

### Python → python-expert
- pytest fixtures
- Parametrized tests
- Mock/patch patterns
- Async test handling

### Go → go-expert
- Table-driven tests (`[]struct` pattern)
- `testing.T` and subtests (`t.Run`)
- Testify assertions (when detected)
- Benchmark functions (`testing.B`)
- Parallel tests (`t.Parallel()`)

### Rust → rust-expert
- `#[test]` attribute functions
- `#[cfg(test)]` module organization
- `#[should_panic]` for error testing
- proptest/quickcheck for property testing

### PHP/Laravel → laravel-expert
- PHPUnit/Pest patterns
- Database transactions
- Factory usage

### E2E → cypress-expert
- Page object patterns
- Custom commands
- Network stubbing

### Playwright → typescript-expert
- Page object model patterns
- Locator strategies
- Visual regression testing

---

## CLI Tool Integration

| Tool | Purpose | Fallback |
|------|---------|----------|
| `jq` | Parse package.json | Read tool |
| `rg` | Find existing tests | Grep tool |
| `ast-grep` | Parse function signatures | ripgrep patterns |
| `fd` | Find test files | Glob tool |
| Chrome DevTools MCP | Visual testing (--visual) | Playwright/Cypress |

**Graceful degradation:**
```bash
command -v jq >/dev/null 2>&1 && cat package.json | jq '.devDependencies' || cat package.json
```

---

## Reference Files

For framework-specific code examples, see:
- `frameworks.md` - Complete test examples for all supported languages
- `visual-testing.md` - Chrome DevTools integration for --visual flag

---

## Integration

| Command | Relationship |
|---------|--------------|
| `/review` | Review generated tests before committing |
| `/explain` | Understand complex code before testing |
| `/save` | Track test coverage goals |

Overview

This skill generates comprehensive tests with automatic framework detection, expert agent routing, and auto TaskCreate to verify results. It matches project conventions and creates test files in the correct locations, covering happy paths, edge cases, and error handling. The tool supports multiple languages and test runners and integrates with a task system to run and review generated tests.

How this skill works

Given a target file, function, or directory, the skill analyzes source code, finds existing tests to avoid duplicates, and extracts signatures when available. It detects the project test framework (e.g., jest, pytest, go test, cargo test) and loads project conventions to choose file naming and placement. The system routes work to language-specific expert agents (TypeScript, Python, Go, Rust, PHP, React, Vue, Cypress, etc.), generates focused tests according to requested depth and focus, and auto-creates a verification task for running and reviewing results.

When to use it

  • Add unit tests to a new or existing file without manual framework setup
  • Generate quick coverage for functions or directories before a refactor
  • Create E2E tests or visual checks (with --visual) for UI flows
  • Auto-create tasks to run and verify generated tests in CI or locally
  • Bootstrap test suites across polyglot repositories

Best practices

  • Run framework detection in the repo root so package/config files are available
  • Specify --focus (happy|edge|error|all) and --depth (quick|normal|thorough) for predictable output
  • Review generated tests before committing and refine edge/error cases manually
  • Keep test files co-located or follow the project’s established test layout to avoid clutter
  • Use the TaskCreate output to run tests and iterate (/review, /save)

Example use cases

  • Generate Jest unit tests for src/utils/helper.ts, placing files per project convention
  • Create pytest cases for a specific Python function with parameterized edge inputs
  • Produce table-driven Go tests and subtests for pkg/auth/token.go
  • Dispatch parallel expert agents to create tests across a multi-language feature directory
  • Generate Cypress E2E tests for cypress/ flows or visual checks with Chrome DevTools MCP

FAQ

What frameworks are detected automatically?

Common frameworks: Jest, Vitest, Mocha, Cypress, Playwright, pytest, unittest, go test, cargo test, PHPUnit/Pest, and related runners via project config files.

How do I control test scope and thoroughness?

Use --focus (happy, edge, error, all) to choose types of tests and --depth (quick, normal, thorough) to select coverage level and mocking detail.

Can it avoid creating duplicate tests?

Yes. It scans existing test files and naming patterns to skip duplicates and follow existing conventions.