home / skills / tenequm / claude-plugins / skill-factory

skill-factory skill

/skill-factory

This skill automatically creates production-ready Claude skills by analyzing requests, selecting the best method, and guaranteeing quality without user input.

npx playbooks add skill tenequm/claude-plugins --skill skill-factory

Review the files below or copy the command above to add this skill to your agents.

Files (16)
SKILL.md
12.0 KB
---
name: skill-factory
description: Autonomous skill creation agent that analyzes requests, automatically selects the best creation method (documentation scraping via Skill_Seekers, manual TDD construction, or hybrid), ensures quality compliance with Anthropic best practices, and delivers production-ready skills without requiring user decision-making or navigation
when_to_use: when you need to create any Claude skill and want it done automatically with guaranteed quality - works for documentation-based skills, GitHub repositories, PDFs, custom workflows, or hybrid approaches
version: 0.1.0
---

# Skill Factory

**Autonomous skill creation - just tell me what you need, I'll handle everything.**

## What This Does

You request a skill, I deliver a production-ready skill with guaranteed quality (score >= 8.0/10).

**No decision-making required. No tool selection. No quality checking. Just results.**

### Anthropic's Official Best Practices

For comprehensive guidance on creating effective skills, see:

- **[references/overview.md](references/overview.md)** - Complete overview of Agent Skills architecture, progressive disclosure, and how Skills work across different platforms (API, Claude Code, Agent SDK, claude.ai)
- **[references/quickstart.md](references/quickstart.md)** - Quick tutorial on using pre-built Agent Skills in the Claude API with practical code examples
- **[references/best-practices.md](references/best-practices.md)** - Detailed authoring best practices including core principles, skill structure, progressive disclosure patterns, workflows, evaluation strategies, and common patterns
- **[references/anthropic-best-practices.md](references/anthropic-best-practices.md)** - Quality scoring system (10/10 criteria) used by skill-factory

These references provide Anthropic's official guidance and are consulted during the quality assurance phase.

## Usage

Simply describe the skill you need:

```
"Create a skill for Anchor development with latest docs and best practices"
"Create a React skill from react.dev with comprehensive examples"
"Create a skill for Solana transaction debugging workflows"
"Create a skill for writing technical documentation following company standards"
```

**I will automatically:**
1. ✅ Analyze your request
2. ✅ Select the optimal creation method
3. ✅ Create the skill
4. ✅ Run quality assurance loops (until score >= 8.0)
5. ✅ Test with automated scenarios
6. ✅ Deliver ready-to-use skill with stats

## What You Get

```
✅ anchor-development skill ready!

📊 Quality Score: 8.9/10 (Excellent)
📝 Lines: 412 (using progressive disclosure)
📚 Coverage: 247 documentation pages
💡 Examples: 68 code samples
🧪 Test Pass Rate: 100% (15/15 scenarios)

📁 Location: ~/.claude/skills/anchor-development/
📦 Zip: ~/Downloads/anchor-development.zip

Try it: "How do I create an Anchor program?"
```

## How It Works (Behind the Scenes)

### Phase 1: Request Analysis (Automatic)

I analyze your request to determine:

**Source Detection:**
- Documentation URL/mention? → Automated scraping path
- "Latest docs", "current version"? → Automated path
- GitHub repository mention? → Automated path
- PDF/manual path? → Automated path
- Custom workflow/process description? → Manual TDD path
- Both documentation AND custom needs? → Hybrid path

**Quality Requirements Extraction:**
- "Best practices" → Enforce quality gates
- "Latest version" → Scrape current docs
- "Examples" → Ensure code samples included
- "Comprehensive" → Verify coverage completeness

### Phase 2: Execution (Automatic)

**Path A: Documentation-Based (Skill_Seekers)**
```
Detected: Documentation source available
Method: Automated scraping with quality enhancement

Steps I take:
1. Check Skill_Seekers installation (install if needed)
2. Configure scraping parameters automatically
3. Run scraping with optimal settings
4. Monitor progress
5. Initial quality check
6. If score < 8.0: Run enhancement loop
7. Re-score until >= 8.0
8. Test with auto-generated scenarios
9. Package and deliver
```

**Path B: Custom Workflows (Manual TDD)**
```
Detected: Custom workflow/process
Method: Test-Driven Documentation (obra methodology)

Steps I take:
1. Create pressure test scenarios
2. Run baseline (without skill)
3. Document agent behavior
4. Write minimal skill addressing baseline
5. Test with skill present
6. Identify rationalizations/gaps
7. Close loopholes
8. Iterate until bulletproof
9. Package and deliver
```

**Path C: Hybrid**
```
Detected: Documentation + custom requirements
Method: Scrape then enhance

Steps I take:
1. Scrape documentation (Path A)
2. Identify gaps vs requirements
3. Fill gaps with TDD approach (Path B)
4. Unify and test as whole
5. Quality loop until >= 8.0
6. Package and deliver
```

### Phase 3: Quality Assurance Loop (Automatic)

**I enforce Anthropic best practices:**

```python
while quality_score < 8.0:
    issues = analyze_against_anthropic_guidelines(skill)

    if "vague_description" in issues:
        improve_description_specificity()

    if "missing_examples" in issues:
        extract_or_generate_examples()

    if "too_long" in issues:
        apply_progressive_disclosure()

    if "poor_structure" in issues:
        reorganize_content()

    quality_score = rescore()
```

**Quality Criteria (Anthropic Best Practices):**
- ✅ Description: Specific, clear, includes when_to_use
- ✅ Conciseness: <500 lines OR progressive disclosure
- ✅ Examples: Concrete code samples, not abstract
- ✅ Structure: Well-organized, clear sections
- ✅ Name: Follows conventions (lowercase, hyphens, descriptive)

**Important**: The quality assurance process consults [references/best-practices.md](references/best-practices.md) for Anthropic's complete authoring guidelines and [references/anthropic-best-practices.md](references/anthropic-best-practices.md) for the 10-point scoring criteria.

### Phase 4: Testing (Automatic)

**I generate and run test scenarios:**

```python
# Auto-generate test cases from skill content
test_cases = extract_key_topics(skill)

for topic in test_cases:
    query = f"How do I {topic}?"

    # Test WITHOUT skill (baseline)
    baseline = run_query_without_skill(query)

    # Test WITH skill
    with_skill = run_query_with_skill(query)

    # Verify improvement
    if not is_better(with_skill, baseline):
        identify_gap()
        enhance_skill()
        retest()
```

### Phase 5: Delivery (Automatic)

```
Package skill:
- Create skill directory structure
- Generate SKILL.md with frontmatter
- Create reference files (if using progressive disclosure)
- Add examples directory
- Create .zip for easy upload
- Install to ~/.claude/skills/ (if desired)
- Generate summary statistics
```

## Progress Reporting

You'll see real-time progress:

```
🔍 Analyzing request...
   ✅ Detected: Documentation-based (docs.rs/anchor-lang)
   ✅ Requirements: Latest version, best practices, examples

🔄 Creating skill...
   📥 Scraping docs.rs/anchor-lang... (2 min)
   📚 Extracting 247 pages...
   💾 Organizing content...

📊 Quality check: 7.4/10
   ⚠️  Issues found:
       - Description too generic (fixing...)
       - Missing examples in 4 sections (adding...)
       - Some outdated patterns (updating...)

🔧 Enhancing skill...
   ✏️  Description improved
   📝 Examples added
   🔄 Patterns updated

📊 Quality check: 8.9/10 ✅

🧪 Testing...
   ✅ 15/15 scenarios passing

✅ anchor-development skill ready!
```

## Dependencies

**Required:**
- Python 3.10+ (for quality scripts)
- bash (for automation scripts)

**Optional (auto-installed when needed):**
- Skill_Seekers (for documentation scraping)
  - Will prompt for installation on first documentation-based request
  - One-command setup: `scripts/install-skill-seekers.sh`

## Configuration

**Environment variables (optional):**

```bash
# Skill_Seekers installation path
export SKILL_SEEKERS_PATH="$HOME/Skill_Seekers"

# Output directory for created skills
export SKILL_OUTPUT_DIR="$HOME/.claude/skills"

# Minimum quality score (default: 8.0)
export MIN_QUALITY_SCORE="8.0"

# Auto-install dependencies (default: prompt)
export AUTO_INSTALL_DEPS="true"
```

## Examples

**Example 1: Documentation Skill**
```
User: "Create a React skill from react.dev"

Agent:
🔍 Analyzing... → Documentation-based
🔄 Scraping react.dev... → 3 min
📊 Quality: 7.8 → 8.6 ✅
🧪 Testing: 12/12 passing ✅
✅ react-development skill ready (8.6/10)
```

**Example 2: Custom Workflow Skill**
```
User: "Create a skill for debugging Solana transaction failures"

Agent:
🔍 Analyzing... → Custom workflow (no docs to scrape)
📝 Using TDD methodology...
🧪 RED: Testing baseline...
✏️  GREEN: Writing skill...
🔄 REFACTOR: Closing loopholes...
📊 Quality: 8.3 ✅
✅ solana-transaction-debugging skill ready (8.3/10)
```

**Example 3: Hybrid Skill**
```
User: "Create an Anchor skill with docs plus custom debugging workflows"

Agent:
🔍 Analyzing... → Hybrid (docs + custom)
📥 Scraping docs.rs/anchor-lang... → 2 min
📝 Adding custom debugging workflows...
🔄 Integrating and testing...
📊 Quality: 8.9 ✅
✅ anchor-development skill ready (8.9/10)
```

## Quality Guarantee

**Every skill delivered by skill-factory:**
- ✅ Scores >= 8.0/10 on Anthropic best practices
- ✅ Has concrete examples (not abstract)
- ✅ Follows structure conventions
- ✅ Tested with auto-generated scenarios
- ✅ Ready to use immediately

**If quality < 8.0, I keep working until it reaches 8.0+**

## Troubleshooting

**Skill_Seekers installation fails:**
```bash
# Manual installation
git clone https://github.com/yusufkaraaslan/Skill_Seekers ~/Skill_Seekers
cd ~/Skill_Seekers
pip install -r requirements.txt

# Or use installation script
~/Projects/claude-skills/skill-factory/scripts/install-skill-seekers.sh
```

**Quality score stuck below 8.0:**
- I'll report what's blocking and suggest manual review
- Check references/anthropic-best-practices.md for criteria
- Run manual enhancement if needed

**Want to understand methodology:**
- See references/obra-tdd-methodology.md (testing approach)
- See references/anthropic-best-practices.md (quality criteria)
- See references/skill-seekers-integration.md (automation details)

## Reference Files

**Anthropic Official Documentation:**
- references/overview.md - Agent Skills architecture, progressive disclosure, and platform details
- references/quickstart.md - Quick tutorial on using pre-built Agent Skills in the Claude API
- references/best-practices.md - Comprehensive authoring guidelines from Anthropic
- references/anthropic-best-practices.md - Quality scoring system (10/10 criteria)

**Skill Factory Implementation Details:**
- references/obra-tdd-methodology.md - Full TDD testing approach
- references/skill-seekers-integration.md - Automation documentation
- references/request-analysis.md - How requests are parsed
- references/quality-loops.md - Enhancement algorithms

## Scripts Reference

Available helper scripts in `scripts/` directory:
- **check-skill-seekers.sh** - Check if Skill_Seekers is installed
- **install-skill-seekers.sh** - One-command Skill_Seekers setup
- **quality-check.py** - Score any skill against Anthropic best practices

Usage examples:
```bash
# Check Skill_Seekers installation
./scripts/check-skill-seekers.sh

# Install Skill_Seekers
./scripts/install-skill-seekers.sh

# Quality check a skill
python3 ./scripts/quality-check.py /path/to/skill/SKILL.md
```

## Philosophy

**You don't want to:**
- Navigate decision trees
- Choose between tools
- Check quality manually
- Test with subagents yourself
- Wonder if output is good

**You want to:**
- Describe what you need
- Get high-quality result
- Start using immediately

**That's what skill-factory delivers.**

## Credits

Built on top of excellent tools:
- [Skill_Seekers](https://github.com/yusufkaraaslan/Skill_Seekers) - Documentation scraping
- [obra/superpowers-skills](https://github.com/obra/superpowers-skills) - TDD methodology
- [Anthropic skill-creator](https://github.com/anthropics/skills) - Best practices

Skill-factory orchestrates these tools with automatic quality assurance and testing.

---

**Just tell me what skill you need. I'll handle the rest.**

Overview

This skill automates end-to-end creation of production-ready agent skills from a simple request. It analyzes your needs, chooses the optimal creation method (documentation scraping, test-driven manual construction, or a hybrid), enforces Anthropic best practices, and delivers a packaged, tested skill with a guaranteed quality score >= 8.0/10. No tool selection or manual QA is required from you.

How this skill works

I parse your request to detect sources and requirements, then select one of three paths: documentation-based scraping using Skill_Seekers, manual Test-Driven Documentation (TDD) construction for custom workflows, or a hybrid that combines both. The generator builds the skill, runs iterative quality-assurance loops against Anthropic guidelines, auto-generates and runs test scenarios, and packages the final skill for immediate use. Delivery includes coverage stats, examples, tests, and an installable zip.

When to use it

  • You need a production-ready agent skill from public or private docs without doing scraping yourself.
  • You want a custom workflow skill where behavior must be validated with tests and TDD.
  • You have mixed requirements: canonical docs plus project-specific behaviors.
  • You want guaranteed quality (score >= 8.0) and automated testing before delivery.
  • You prefer a hands-off process: describe the outcome and receive a ready-to-use skill.

Best practices

  • Provide a clear, specific request including source links and any custom workflows or examples you expect.
  • Specify ‘latest docs’ or a version tag if you need up-to-date scraping; otherwise include the exact URL or repo.
  • Request concrete examples or test scenarios to ensure the skill includes runnable samples.
  • Allow auto-install of dependencies for documentation scraping (or set SKILL_SEEKERS_PATH).
  • Accept progressive disclosure for long content to keep the skill concise and navigable.

Example use cases

  • Create a React skill from react.dev with comprehensive code examples and sample components.
  • Build an Anchor development skill by scraping docs.rs and adding custom debugging workflows.
  • Produce a Solana transaction debugging skill using TDD to validate failure-mode handling.
  • Assemble a company-standard documentation-writing skill that enforces internal style and examples.

FAQ

How long does creation typically take?

Small skills from a single docs source can complete in minutes; large multi-source or hybrid skills may take tens of minutes depending on scrape size and test coverage.

What if the quality score stalls below 8.0?

The system reports blocking issues and applies targeted enhancement loops; for persistent gaps it will recommend a short manual review to resolve ambiguous requirements.