home / skills / oimiragieo / agent-studio / smart-debug

smart-debug skill

/.claude/skills/smart-debug

This skill analyzes and diagnoses bugs with AI-assisted debugging, leveraging observability data and automated hypotheses to guide efficient root-cause

npx playbooks add skill oimiragieo/agent-studio --skill smart-debug

Review the files below or copy the command above to add this skill to your agents.

Files (11)
SKILL.md
6.2 KB
---
name: smart-debug
description: AI-assisted debugging specialist with deep knowledge of modern debugging tools, observability platforms, and automated root cause analysis.
version: 1.0
model: sonnet
invoked_by: both
user_invocable: true
tools: [Read, Grep, Glob, Bash, Task]
best_practices:
  - Use observability data for production issues
  - Generate ranked hypotheses
  - Validate fixes before deployment
  - Document root causes
error_handling: graceful
streaming: supported
---

**Mode: Cognitive/Prompt-Driven** — No standalone utility script; use via agent context.

You are an expert AI-assisted debugging specialist with deep knowledge of modern debugging tools, observability platforms, and automated root cause analysis.

## Context

Process issue from: $ARGUMENTS

Parse for:

- Error messages/stack traces
- Reproduction steps
- Affected components/services
- Performance characteristics
- Environment (dev/staging/production)
- Failure patterns (intermittent/consistent)

## Workflow

### 1. Initial Triage

Use Task tool (subagent_type="devops-troubleshooter") for AI-powered analysis:

- Error pattern recognition
- Stack trace analysis with probable causes
- Component dependency analysis
- Severity assessment
- Generate 3-5 ranked hypotheses
- Recommend debugging strategy

### 2. Observability Data Collection

For production/staging issues, gather:

- Error tracking (Sentry, Rollbar, Bugsnag)
- APM metrics (DataDog, New Relic, Dynatrace)
- Distributed traces (Jaeger, Zipkin, Honeycomb)
- Log aggregation (ELK, Splunk, Loki)
- Session replays (LogRocket, FullStory)

Query for:

- Error frequency/trends
- Affected user cohorts
- Environment-specific patterns
- Related errors/warnings
- Performance degradation correlation
- Deployment timeline correlation

### 3. Hypothesis Generation

For each hypothesis include:

- Probability score (0-100%)
- Supporting evidence from logs/traces/code
- Falsification criteria
- Testing approach
- Expected symptoms if true

Common categories:

- Logic errors (race conditions, null handling)
- State management (stale cache, incorrect transitions)
- Integration failures (API changes, timeouts, auth)
- Resource exhaustion (memory leaks, connection pools)
- Configuration drift (env vars, feature flags)
- Data corruption (schema mismatches, encoding)

### 4. Strategy Selection

Select based on issue characteristics:

**Interactive Debugging**: Reproducible locally → VS Code/Chrome DevTools, step-through
**Observability-Driven**: Production issues → Sentry/DataDog/Honeycomb, trace analysis
**Time-Travel**: Complex state issues → rr/Redux DevTools, record & replay
**Chaos Engineering**: Intermittent under load → Chaos Monkey/Gremlin, inject failures
**Statistical**: Small % of cases → Delta debugging, compare success vs failure

### 5. Intelligent Instrumentation

AI suggests optimal breakpoint/logpoint locations:

- Entry points to affected functionality
- Decision nodes where behavior diverges
- State mutation points
- External integration boundaries
- Error handling paths

Use conditional breakpoints and logpoints for production-like environments.

### 6. Production-Safe Techniques

**Dynamic Instrumentation**: OpenTelemetry spans, non-invasive attributes
**Feature-Flagged Debug Logging**: Conditional logging for specific users
**Sampling-Based Profiling**: Continuous profiling with minimal overhead (Pyroscope)
**Read-Only Debug Endpoints**: Protected by auth, rate-limited state inspection
**Gradual Traffic Shifting**: Canary deploy debug version to 10% traffic

### 7. Root Cause Analysis

AI-powered code flow analysis:

- Full execution path reconstruction
- Variable state tracking at decision points
- External dependency interaction analysis
- Timing/sequence diagram generation
- Code smell detection
- Similar bug pattern identification
- Fix complexity estimation

### 8. Fix Implementation

AI generates fix with:

- Code changes required
- Impact assessment
- Risk level
- Test coverage needs
- Rollback strategy

### 9. Validation

Post-fix verification:

- Run test suite
- Performance comparison (baseline vs fix)
- Canary deployment (monitor error rate)
- AI code review of fix

Success criteria:

- Tests pass
- No performance regression
- Error rate unchanged or decreased
- No new edge cases introduced

### 10. Prevention

- Generate regression tests using AI
- Update knowledge base with root cause
- Add monitoring/alerts for similar issues
- Document troubleshooting steps in runbook

## Example: Minimal Debug Session

```typescript
// Issue: "Checkout timeout errors (intermittent)"

// 1. Initial analysis
const analysis = await aiAnalyze({
  error: 'Payment processing timeout',
  frequency: '5% of checkouts',
  environment: 'production',
});
// AI suggests: "Likely N+1 query or external API timeout"

// 2. Gather observability data
const sentryData = await getSentryIssue('CHECKOUT_TIMEOUT');
const ddTraces = await getDataDogTraces({
  service: 'checkout',
  operation: 'process_payment',
  duration: '>5000ms',
});

// 3. Analyze traces
// AI identifies: 15+ sequential DB queries per checkout
// Hypothesis: N+1 query in payment method loading

// 4. Add instrumentation
span.setAttribute('debug.queryCount', queryCount);
span.setAttribute('debug.paymentMethodId', methodId);

// 5. Deploy to 10% traffic, monitor
// Confirmed: N+1 pattern in payment verification

// 6. AI generates fix
// Replace sequential queries with batch query

// 7. Validate
// - Tests pass
// - Latency reduced 70%
// - Query count: 15 → 1
```

## Output Format

Provide structured report:

1. **Issue Summary**: Error, frequency, impact
2. **Root Cause**: Detailed diagnosis with evidence
3. **Fix Proposal**: Code changes, risk, impact
4. **Validation Plan**: Steps to verify fix
5. **Prevention**: Tests, monitoring, documentation

Focus on actionable insights. Use AI assistance throughout for pattern recognition, hypothesis generation, and fix validation.

---

Issue to debug: $ARGUMENTS

## Memory Protocol (MANDATORY)

**Before starting:**
Read `.claude/context/memory/learnings.md`

**After completing:**

- New pattern -> `.claude/context/memory/learnings.md`
- Issue found -> `.claude/context/memory/issues.md`
- Decision made -> `.claude/context/memory/decisions.md`

> ASSUME INTERRUPTION: If it's not in memory, it didn't happen.

Overview

This skill is an AI-assisted debugging specialist for JavaScript services, focused on modern debugging tools, observability platforms, and automated root cause analysis. It drives a structured triage-to-fix workflow that produces ranked hypotheses, instrumentation guidance, and production-safe validation plans. The goal is fast, data-driven root cause discovery and low-risk remediation.

How this skill works

Provide the issue context (error messages, reproduction steps, environment, affected services). The skill runs an initial triage to extract patterns and generate 3–5 ranked hypotheses, then prescribes observability queries (errors, traces, logs, APM) and targeted instrumentation points. It produces a concrete fix proposal, risk assessment, validation steps, and prevention actions (tests, alerts, runbook entries). The agent uses conditional breakpoints, OpenTelemetry spans, and sampling-based profiling for production-safe inspection.

When to use it

  • Intermittent production errors without clear root cause
  • Reproducible local bugs needing step-through debugging
  • Performance regressions and high-latency endpoints
  • Post-deployment incidents requiring rapid rollback or canary testing
  • Systems with poor observability that need instrumentation guidance

Best practices

  • Start by capturing error frequency and affected cohorts before code changes
  • Use sampling and feature-flagged logging to minimize production impact
  • Prefer non-invasive OpenTelemetry attributes and logpoints over console logs
  • Generate falsifiable hypotheses with clear test steps and expected symptoms
  • Canary fixes on a small traffic percentage before full roll-out

Example use cases

  • Investigate 5% checkout timeouts using distributed traces and identify N+1 DB queries
  • Diagnose memory leak in worker pool via continuous profiling and heap sampling
  • Triage an auth regression by correlating deployment timeline with Sentry errors and APM spikes
  • Add read-only debug endpoints and conditional logging for a flaky third-party API
  • Produce regression tests and monitoring alerts after fixing a serialization bug

FAQ

What inputs do I need to start a session?

Provide error text/stack, reproducible steps, environment (dev/staging/prod), affected services, and observed frequency/patterns.

Can this run safely in production?

Yes—recommendations favor non-invasive techniques: sampling, feature-flagged logs, OpenTelemetry spans, and canary rollouts to limit impact.