home / skills / sickn33 / antigravity-awesome-skills / error-debugging-multi-agent-review

error-debugging-multi-agent-review skill

/skills/error-debugging-multi-agent-review

This skill coordinates multiple specialized agents to debug errors in multi-agent code review workflows and deliver actionable, consolidated guidance.

This is most likely a fork of the error-debugging-multi-agent-review skill from xfstudio
npx playbooks add skill sickn33/antigravity-awesome-skills --skill error-debugging-multi-agent-review

Review the files below or copy the command above to add this skill to your agents.

Files (1)
SKILL.md
6.6 KB
---
name: error-debugging-multi-agent-review
description: "Use when working with error debugging multi agent review"
---

# Multi-Agent Code Review Orchestration Tool

## Use this skill when

- Working on multi-agent code review orchestration tool tasks or workflows
- Needing guidance, best practices, or checklists for multi-agent code review orchestration tool

## Do not use this skill when

- The task is unrelated to multi-agent code review orchestration tool
- You need a different domain or tool outside this scope

## Instructions

- Clarify goals, constraints, and required inputs.
- Apply relevant best practices and validate outcomes.
- Provide actionable steps and verification.
- If detailed examples are required, open `resources/implementation-playbook.md`.

## Role: Expert Multi-Agent Review Orchestration Specialist

A sophisticated AI-powered code review system designed to provide comprehensive, multi-perspective analysis of software artifacts through intelligent agent coordination and specialized domain expertise.

## Context and Purpose

The Multi-Agent Review Tool leverages a distributed, specialized agent network to perform holistic code assessments that transcend traditional single-perspective review approaches. By coordinating agents with distinct expertise, we generate a comprehensive evaluation that captures nuanced insights across multiple critical dimensions:

- **Depth**: Specialized agents dive deep into specific domains
- **Breadth**: Parallel processing enables comprehensive coverage
- **Intelligence**: Context-aware routing and intelligent synthesis
- **Adaptability**: Dynamic agent selection based on code characteristics

## Tool Arguments and Configuration

### Input Parameters
- `$ARGUMENTS`: Target code/project for review
  - Supports: File paths, Git repositories, code snippets
  - Handles multiple input formats
  - Enables context extraction and agent routing

### Agent Types
1. Code Quality Reviewers
2. Security Auditors
3. Architecture Specialists
4. Performance Analysts
5. Compliance Validators
6. Best Practices Experts

## Multi-Agent Coordination Strategy

### 1. Agent Selection and Routing Logic
- **Dynamic Agent Matching**:
  - Analyze input characteristics
  - Select most appropriate agent types
  - Configure specialized sub-agents dynamically
- **Expertise Routing**:
  ```python
  def route_agents(code_context):
      agents = []
      if is_web_application(code_context):
          agents.extend([
              "security-auditor",
              "web-architecture-reviewer"
          ])
      if is_performance_critical(code_context):
          agents.append("performance-analyst")
      return agents
  ```

### 2. Context Management and State Passing
- **Contextual Intelligence**:
  - Maintain shared context across agent interactions
  - Pass refined insights between agents
  - Support incremental review refinement
- **Context Propagation Model**:
  ```python
  class ReviewContext:
      def __init__(self, target, metadata):
          self.target = target
          self.metadata = metadata
          self.agent_insights = {}

      def update_insights(self, agent_type, insights):
          self.agent_insights[agent_type] = insights
  ```

### 3. Parallel vs Sequential Execution
- **Hybrid Execution Strategy**:
  - Parallel execution for independent reviews
  - Sequential processing for dependent insights
  - Intelligent timeout and fallback mechanisms
- **Execution Flow**:
  ```python
  def execute_review(review_context):
      # Parallel independent agents
      parallel_agents = [
          "code-quality-reviewer",
          "security-auditor"
      ]

      # Sequential dependent agents
      sequential_agents = [
          "architecture-reviewer",
          "performance-optimizer"
      ]
  ```

### 4. Result Aggregation and Synthesis
- **Intelligent Consolidation**:
  - Merge insights from multiple agents
  - Resolve conflicting recommendations
  - Generate unified, prioritized report
- **Synthesis Algorithm**:
  ```python
  def synthesize_review_insights(agent_results):
      consolidated_report = {
          "critical_issues": [],
          "important_issues": [],
          "improvement_suggestions": []
      }
      # Intelligent merging logic
      return consolidated_report
  ```

### 5. Conflict Resolution Mechanism
- **Smart Conflict Handling**:
  - Detect contradictory agent recommendations
  - Apply weighted scoring
  - Escalate complex conflicts
- **Resolution Strategy**:
  ```python
  def resolve_conflicts(agent_insights):
      conflict_resolver = ConflictResolutionEngine()
      return conflict_resolver.process(agent_insights)
  ```

### 6. Performance Optimization
- **Efficiency Techniques**:
  - Minimal redundant processing
  - Cached intermediate results
  - Adaptive agent resource allocation
- **Optimization Approach**:
  ```python
  def optimize_review_process(review_context):
      return ReviewOptimizer.allocate_resources(review_context)
  ```

### 7. Quality Validation Framework
- **Comprehensive Validation**:
  - Cross-agent result verification
  - Statistical confidence scoring
  - Continuous learning and improvement
- **Validation Process**:
  ```python
  def validate_review_quality(review_results):
      quality_score = QualityScoreCalculator.compute(review_results)
      return quality_score > QUALITY_THRESHOLD
  ```

## Example Implementations

### 1. Parallel Code Review Scenario
```python
multi_agent_review(
    target="/path/to/project",
    agents=[
        {"type": "security-auditor", "weight": 0.3},
        {"type": "architecture-reviewer", "weight": 0.3},
        {"type": "performance-analyst", "weight": 0.2}
    ]
)
```

### 2. Sequential Workflow
```python
sequential_review_workflow = [
    {"phase": "design-review", "agent": "architect-reviewer"},
    {"phase": "implementation-review", "agent": "code-quality-reviewer"},
    {"phase": "testing-review", "agent": "test-coverage-analyst"},
    {"phase": "deployment-readiness", "agent": "devops-validator"}
]
```

### 3. Hybrid Orchestration
```python
hybrid_review_strategy = {
    "parallel_agents": ["security", "performance"],
    "sequential_agents": ["architecture", "compliance"]
}
```

## Reference Implementations

1. **Web Application Security Review**
2. **Microservices Architecture Validation**

## Best Practices and Considerations

- Maintain agent independence
- Implement robust error handling
- Use probabilistic routing
- Support incremental reviews
- Ensure privacy and security

## Extensibility

The tool is designed with a plugin-based architecture, allowing easy addition of new agent types and review strategies.

## Invocation

Target for review: $ARGUMENTS

Overview

This skill provides a focused playbook for debugging and orchestrating multi-agent code review workflows. It helps diagnose failures in agent selection, context propagation, execution ordering, and result synthesis so teams can restore reliable, high-confidence reviews quickly. Use it to systematically locate faults, verify fixes, and prevent recurrence.

How this skill works

The skill inspects orchestration components: agent routing logic, shared context/state management, parallel vs sequential execution, conflict resolution, and synthesis of agent outputs. It runs targeted checks and validation steps, suggests concrete fixes (routing rules, timeouts, caching), and recommends verification tests to confirm the root cause is addressed. Outputs include prioritized issues and actionable remediation steps.

When to use it

  • A multi-agent review run produces inconsistent or contradictory findings
  • Agent selection or routing seems to miss relevant specialists
  • Shared context or state is lost between agent stages
  • Long-running reviews hang or time out unexpectedly
  • You need a reproducible debugging checklist for review orchestration

Best practices

  • Clarify review goals and minimal required context before routing agents
  • Log and version the ReviewContext to trace state changes between agents
  • Prefer hybrid execution: parallelize independent checks and sequence dependent ones
  • Use weighted scoring and escalation rules to resolve conflicting recommendations
  • Cache intermediate artifacts to avoid redundant processing and speed re-runs
  • Add end-to-end tests that simulate common orchestration failure modes

Example use cases

  • Debugging why security auditors aren’t invoked for web app targets
  • Fixing state propagation bugs where later agents see stale metadata
  • Reducing review time by identifying unnecessary redundant analyses
  • Resolving conflicting recommendations between architecture and performance agents
  • Hardening orchestration against timeouts and transient agent failures

FAQ

What common root causes cause wrong agent selection?

Typical causes are incomplete input metadata, brittle routing heuristics, and insufficient feature extraction from the target code. Add richer context extraction and probabilistic routing to mitigate.

How do I verify a proposed fix?

Re-run the review with recorded inputs and a lightweight end-to-end scenario, compare agent participation and outputs, and validate quality scores and conflict resolution outcomes against baselines.