home / skills / menkesu / awesome-pm-skills / ai-startup-building

ai-startup-building skill

/ai-startup-building

This skill helps you build AI-native products using modern prompt engineering, cost optimization, and rapid iteration to scale with 2025+ practices.

npx playbooks add skill menkesu/awesome-pm-skills --skill ai-startup-building

Review the files below or copy the command above to add this skill to your agents.

Files (1)
SKILL.md
3.8 KB
---
name: ai-startup-building
description: Builds AI-native products using Dan Shipper's 5-product playbook and Brandon Chu's AI product frameworks. Use when implementing prompt engineering, creating AI-native UX, scaling AI products, or optimizing costs. Focuses on 2025+ best practices.
---

# AI-Native Startup Patterns

## When This Skill Activates

Claude uses this skill when:
- Building AI-first products
- Implementing prompt engineering
- Creating AI-native workflows
- Scaling AI products efficiently

## Core Frameworks

### 1. AI-Native Startup Playbook (Source: Dan Shipper - 5 products, 7-fig revenue, 100% AI)

**Key Principles:**
- Build fast with AI
- Test with real users immediately
- Iterate based on usage
- Focus on distribution, not just product

### 2. 2025 Prompt Engineering Best Practices

**Modern Approach:**
```
- Use structured outputs (JSON)
- Implement streaming
- Design for retry logic
- Plan for model switching
- Cache aggressively
```

### 3. Cost Optimization

**Strategies:**
1. **Caching:** 80% of queries can be cached
2. **Model routing:** Simple → small model, complex → large model
3. **Batching:** Group similar requests
4. **Prompt optimization:** Minimize tokens

---

## Action Templates

### Template: AI Product Implementation

```typescript
// Modern AI product pattern (2025)

interface AIFeature {
  // Streaming for responsiveness
  async *stream(prompt: string): AsyncGenerator<string> {
    const cached = await checkCache(prompt);
    if (cached) return cached;
    
    // Route to appropriate model
    const model = this.selectModel(prompt);
    
    for await (const chunk of model.stream(prompt)) {
      yield chunk;
    }
  }
  
  // Model selection (cost optimization)
  selectModel(prompt: string): Model {
    if (this.isSimple(prompt)) {
      return this.smallModel; // Fast, cheap
    } else {
      return this.largeModel; // Smart, expensive
    }
  }
  
  // Retry logic (reliability)
  async withRetry<T>(fn: () => Promise<T>): Promise<T> {
    for (let i = 0; i < 3; i++) {
      try {
        return await fn();
      } catch (e) {
        if (i === 2) throw e;
        await sleep(Math.pow(2, i) * 1000);
      }
    }
  }
}
```

### Template: AI Cost Budget

```markdown
# AI Cost Analysis: [Feature]

## Current Usage
- Daily requests: [X]
- Model: [GPT-4/Claude/etc.]
- Cost per 1K requests: [$X]
- Monthly cost: [$Y]

## Optimization Plan

### 1. Caching (Est. 80% hit rate)
- Before: [100]% paid calls
- After: [20]% paid calls
- Savings: [80]%

### 2. Model Routing
- Simple queries ([60]%): Small model
- Complex queries ([40]%): Large model
- Savings: [50]%

### 3. Batching
- Real-time: [X]% of requests
- Batchable: [Y]% of requests
- Savings: [Z]%

## Projected Cost
- Before optimization: [$X/month]
- After optimization: [$Y/month]
- Reduction: [Z]%
```

---

## Quick Reference

### 🤖 AI Startup Checklist

**Build:**
- [ ] Streaming implemented
- [ ] Retry logic added
- [ ] Model switching supported
- [ ] Structured outputs (JSON)

**Optimize:**
- [ ] Caching implemented
- [ ] Model routing (simple vs complex)
- [ ] Prompt tokens minimized
- [ ] Batch processing where possible

**Scale:**
- [ ] Cost per user < $X
- [ ] Latency < X seconds
- [ ] Error rate < X%
- [ ] Model swappable (not locked in)

---

## Real-World Examples

### Example: Dan Shipper's AI Products

**Approach:**
- Built 5 AI products in 12 months
- All using AI end-to-end
- Revenue: 7 figures
- Team: Small, AI-augmented

**Key Insights:**
- Ship fast, learn from users
- AI makes small teams powerful
- Distribution > perfect product

---

## Key Quotes

**Dan Shipper:**
> "AI doesn't replace PMs. It makes small PM teams as powerful as large ones."

**On Prompt Engineering:**
> "The best prompts in 2025 are structured, explicit, and tested with evals."

**Brandon Chu:**
> "Build for the AI you'll have in 6 months, not the AI you have today."

Overview

This skill builds AI-native products using Dan Shipper’s 5-product playbook and Brandon Chu’s AI product frameworks, updated for 2025+ best practices. It guides product teams through prompt engineering, AI-native UX, cost optimization, and scalable model architectures. The focus is on rapid shipping, data-driven iteration, and efficient distribution so small teams can achieve outsized outcomes. Use it to turn AI capabilities into repeatable product playbooks and measurable business results.

How this skill works

The skill inspects product requirements and converts them into actionable AI patterns: streaming responses, structured JSON outputs, model routing, caching, batching, and retry logic. It provides implementation templates and cost-analysis blueprints to quantify savings from caching and model selection. It also prescribes metrics and checkpoints for latency, error rate, and per-user cost so you can iterate on both product and infrastructure.

When to use it

  • Designing an AI-first feature that needs low latency and reliable outputs
  • Implementing prompt engineering with structured outputs and evals
  • Planning cost optimization across models, caching, and batching
  • Scaling an AI product while keeping error rates and costs in check
  • Creating an AI-native UX that streams and updates in real time

Best practices

  • Return structured outputs (JSON) to make downstream logic deterministic
  • Implement streaming + retry logic for responsiveness and reliability
  • Route simple queries to smaller models and complex ones to larger models
  • Cache aggressively—target an 80% hit rate for repeat queries
  • Batch non-real-time requests to reduce per-request overhead
  • Design model swappability and monitor token usage continuously

Example use cases

  • A writing assistant that streams content as it’s generated and caches templates for repeat prompts
  • A customer-support agent that routes short FAQs to a small model and escalates complex tickets to a larger model
  • A product that batches nightly data enrichment jobs to cut per-call costs
  • Rapidly validating multiple AI product ideas by launching incremental MVPs and measuring real usage
  • A pricing and cost dashboard that projects savings after applying caching and model routing

FAQ

How much can caching actually save?

Typical implementations assume ~80% cache hit rates on repeat or templated prompts, which can reduce paid calls by a comparable percentage; real savings vary by product and user behavior.

When should I switch models dynamically?

Use simple heuristics (prompt length, complexity signals, or classifier) to route cheap requests to smaller models and reserve larger models for high-stakes or complex tasks.

What metrics should I track first?

Start with latency, error rate, cost per user, cache hit rate, and model call distribution (percent small vs large).