home / skills / shipshitdev / library / ai-dev-loop

This skill orchestrates autonomous AI development across Claude, Cursor, and Codex, streamlining task execution and QA gates for faster feature delivery.

npx playbooks add skill shipshitdev/library --skill ai-dev-loop

Review the files below or copy the command above to add this skill to your agents.

Files (2)
SKILL.md
6.0 KB
---
name: ai-dev-loop
description: Orchestrate autonomous AI development with task-based workflow and QA gates
auto_trigger: false
---

# AI Development Loop

Autonomous task execution with QA gates across multiple AI platforms.

## Overview

The AI Development Loop enables fully autonomous feature development where:

- AI agents pick up and implement tasks from a queue
- You do QA only (approve or reject in Testing column)
- Multiple platforms (Claude CLI, Cursor, Codex) can work in parallel
- Rate limits are maximized by switching between platforms

## Architecture

```
┌─────────────┐     ┌─────────────┐     ┌─────────────┐     ┌─────────────┐
│   BACKLOG   │────▶│   TO DO     │────▶│  TESTING    │────▶│    DONE     │
│             │     │             │     │             │     │             │
│  PRDs ready │     │ Agent picks │     │ YOU review  │     │  Shipped    │
│             │     │ & builds    │     │ & approve   │     │             │
└─────────────┘     └─────────────┘     └─────────────┘     └─────────────┘
                          │                   │
                    ┌─────┴─────┐       ┌─────┴─────┐
                    │  Claude   │       │  Reject   │
                    │  Cursor   │       │  → To Do  │
                    │  Codex    │       └───────────┘
                    └───────────┘
```

## Task Lifecycle

### 1. Task Creation

Tasks live in `.agents/TASKS/[task-name].md` with structured metadata:

```markdown
## Task: [Feature Name]

**ID:** feature-name-slug
**Status:** Backlog | To Do | Testing | Done
**Priority:** High | Medium | Low
**PRD:** [Link](../PRDS/prd-file.md)

### Agent Metadata

**Claimed-By:** [platform-session-id]
**Claimed-At:** [timestamp]
**Completed-At:** [timestamp]

### Progress

**Agent-Notes:** [real-time updates]
**QA-Checklist:**

- [ ] Code compiles/lints
- [ ] Tests pass (CI)
- [ ] User acceptance
- [ ] Visual review

### Rejection History

**Rejection-Count:** 0
**Rejections:** [list of rejection notes]
```

### 2. Task Claiming

When an agent runs `/loop`:

1. Scans `.agents/TASKS/` for `Status: To Do`
2. Sorts by priority (High > Medium > Low)
3. Skips tasks with active claims (< 30 min old)
4. Updates task with `Claimed-By` and `Claimed-At`

### 3. Implementation

Agent works on the task:

1. Reads task file and linked PRD
2. Checks `.agents/SESSIONS/` for related past work
3. Implements the feature/fix
4. Updates `Agent-Notes` with progress
5. Creates branch and commits

### 4. Quality Check

Before moving to Testing:

1. Runs qa-reviewer skill
2. Updates QA-Checklist items
3. Ensures code compiles/lints

### 5. Completion

Agent finalizes:

1. Sets `Status: Testing`
2. Sets `Completed-At` timestamp
3. Adds final summary to `Agent-Notes`
4. Prompts for next action

### 6. QA Gate (Your Turn)

In Kaiban.md:

1. Review Testing column
2. Click task to see PRD preview
3. Check linked PR
4. **Approve**: Drag to Done
5. **Reject**: Click reject, add note → returns to To Do

### 7. Rejection Handling

When rejected:

1. Status returns to To Do
2. Rejection-Count increments
3. Rejection note added to history
4. Next `/loop` picks up with full context

## Multi-Platform Strategy

### Platform Strengths

| Platform   | Best For                             |
| ---------- | ------------------------------------ |
| Claude CLI | Complex logic, backend, architecture |
| Cursor     | UI components, styling, visual work  |
| Codex      | Bulk refactoring, migrations, docs   |

### Parallel Execution

Multiple platforms can work simultaneously:

- Each claims different tasks
- Claims prevent conflicts (30-min lock)
- Shared state via task files

### Rate Limit Handling

When rate limited:

1. Agent saves progress to `Agent-Notes`
2. Releases claim (clears `Claimed-By`)
3. Suggests switching platform
4. User continues with different platform

## Daily Workflow

### Morning QA Session

1. Open Kaiban.md extension in VS Code
2. Review Testing column
3. Approve good work → Done
4. Reject with notes → To Do

### Throughout Day

```bash
# Claude CLI
claude
> /loop   # Process task
> /loop   # Next task
# Rate limited? Switch to Cursor
```

### Rate Limit Strategy

```
Claude limit? → Switch to Cursor
Cursor limit? → Switch to Codex
All limited? → QA time (review Testing)
```

## Integration Points

### Kaiban.md Extension

- Visual Kanban board for `.agents/TASKS/`
- Drag & drop status changes
- PRD preview panel
- Reject button with note input
- Agent claim status badges

### Existing Skills

- **qa-reviewer**: 6-phase quality verification
- **session-documenter**: Auto-document completed work
- **rules-capture**: Learn from rejection feedback

### Git Workflow

- Branch per task: `feature/[task-id]`
- Commits with clear messages
- PR linked in task file

## Not a Daemon

Important: `/loop` is NOT a background process.

- Each invocation handles ONE task
- Returns control to user
- User decides to continue or stop
- Respects "never run background processes" rule

## Claim Expiration

Claims expire after 30 minutes:

- Handles agent crashes
- Handles rate limit interruptions
- Previous `Agent-Notes` provide context for pickup
- Enables multi-platform handoff

## Best Practices

### For Task Creation

- Write clear, actionable task descriptions
- Link to PRD for requirements
- Set appropriate priority
- Include testing criteria

### For Agents

- Read task and PRD thoroughly before starting
- Update `Agent-Notes` regularly
- Run qa-reviewer before completing
- Create clean, focused commits

### For QA (You)

- Review PRD alongside implementation
- Provide specific rejection feedback
- Approve incrementally (don't batch)
- Keep Testing column short

Overview

This skill orchestrates an autonomous AI development loop that assigns, implements, and gates tasks across multiple AI platforms. It uses task files as the shared state, enables parallel work by Claude, Cursor, and Codex, and funnels completed work to a human QA gate for final approval. The flow emphasizes single-task execution per invocation and clear rejection handling to preserve context.

How this skill works

Agents scan a tasks directory for items marked To Do, claim a task (30-minute lock), read the linked PRD, implement changes in a task-specific branch, and update Agent-Notes and QA checklist. Before marking a task Testing, agents run an automated qa-reviewer, ensure code compiles/lints, and record completion metadata. Humans review the Testing column in the Kanban UI to approve (Done) or reject (back to To Do with notes).

When to use it

  • When you want autonomous agents to implement well-scoped features with human QA gates
  • When you need to parallelize work across Claude, Cursor, and Codex to avoid rate limits
  • When you want reproducible handoffs and clear audit trails for automated work
  • When you need to manage many small feature tasks with minimal human intervention
  • When you require strict QA checkpoints before code is merged or shipped

Best practices

  • Author clear, actionable task files that link to a concise PRD and include testing criteria
  • Keep tasks focused and small so each /loop invocation completes a single unit of work
  • Update Agent-Notes continuously and use the QA checklist to capture verifiable checks
  • Use platform strengths: Claude for complex logic, Cursor for UI, Codex for bulk refactors
  • Provide specific rejection feedback so the next agent pickup has clear remediation steps

Example use cases

  • Autonomously implement small features: create branch, commit changes, and open PRs for QA
  • Parallelize UI and backend work by assigning related tasks to Cursor and Claude concurrently
  • Handle rate-limit outages by switching platforms while preserving in-progress notes and claims
  • Run afternoon QA sessions where reviewers approve or reject Testing tasks via the Kanban board
  • Automate bulk refactors with Codex while agents document changes via session-documenter

FAQ

Does /loop run continuously in the background?

No. Each /loop invocation handles exactly one task and then returns control to the user.

How are claim conflicts prevented?

Agents write Claimed-By and Claimed-At into the task file and skip claims younger than 30 minutes to avoid collisions.

What happens when a task is rejected?

The task moves back to To Do, Rejection-Count increments, a rejection note is added, and the next /loop pickup includes full context for remediation.