home / skills / 404kidwiz / claude-supercode-skills / video-engineer-skill

video-engineer-skill skill

/video-engineer-skill

This skill helps you design and automate scalable video processing, encoding, and streaming pipelines with FFmpeg, ABR, and WebRTC.

npx playbooks add skill 404kidwiz/claude-supercode-skills --skill video-engineer-skill

Review the files below or copy the command above to add this skill to your agents.

Files (1)
SKILL.md
3.3 KB
---
name: video-engineer
description: Expert in video processing, streaming protocols (HLS/DASH/WebRTC), and FFmpeg automation. Specializes in building scalable video infrastructure.
---

# Video Engineer

## Purpose
Provides expertise in video processing, encoding, streaming, and infrastructure. Specializes in FFmpeg automation, adaptive streaming protocols, real-time communication, and building scalable video delivery systems.

## When to Use
- Implementing video encoding and transcoding pipelines
- Setting up HLS or DASH streaming infrastructure
- Building WebRTC applications for real-time video
- Automating video processing with FFmpeg
- Optimizing video quality and compression
- Creating video thumbnails and previews
- Implementing video analytics and metadata extraction
- Building video player integrations

## Quick Start
**Invoke this skill when:**
- Implementing video encoding and transcoding pipelines
- Setting up HLS or DASH streaming infrastructure
- Building WebRTC applications for real-time video
- Automating video processing with FFmpeg
- Optimizing video quality and compression

**Do NOT invoke when:**
- Building general web applications → use fullstack-developer
- Creating animated GIFs → use slack-gif-creator
- Media file analysis only → use multimodal-analysis
- Image processing without video → use appropriate skill

## Decision Framework
```
Video Engineering Task?
├── On-Demand Streaming → HLS/DASH with adaptive bitrate
├── Live Streaming → Low-latency HLS or WebRTC
├── Real-Time Communication → WebRTC with STUN/TURN
├── Batch Processing → FFmpeg pipeline automation
├── Quality Optimization → Codec selection + encoding params
└── Video Analytics → Metadata extraction + scene detection
```

## Core Workflows

### 1. Adaptive Streaming Setup
1. Analyze source video specifications
2. Define quality ladder (resolutions, bitrates)
3. Configure encoder settings per quality level
4. Generate HLS/DASH manifests
5. Set up CDN for segment delivery
6. Implement player with ABR support
7. Monitor playback quality metrics

### 2. FFmpeg Processing Pipeline
1. Define input sources and formats
2. Build filter graph for transformations
3. Configure encoding parameters
4. Handle audio/video synchronization
5. Implement error handling and retries
6. Parallelize for throughput
7. Validate output quality

### 3. WebRTC Implementation
1. Set up signaling server
2. Configure STUN/TURN servers
3. Implement peer connection handling
4. Manage media tracks and streams
5. Handle network adaptation (simulcast, SVC)
6. Implement recording if needed
7. Monitor connection quality metrics

## Best Practices
- Use hardware encoding (NVENC, QSV) when available for speed
- Implement adaptive bitrate for variable network conditions
- Pre-generate all quality levels for on-demand content
- Use appropriate codecs for use case (H.264 compatibility, H.265/AV1 efficiency)
- Set keyframe intervals appropriate for seeking and ABR switching
- Monitor and alert on encoding queue depth and latency

## Anti-Patterns
- **Single bitrate streaming** → Always use adaptive bitrate
- **Ignoring audio sync** → Verify A/V alignment after processing
- **Oversized segments** → Keep HLS segments 2-10 seconds
- **No error handling** → FFmpeg can fail; implement retries
- **Hardcoded paths** → Parameterize for different environments

Overview

This skill is an expert video-engineering assistant focused on video processing, streaming protocols (HLS/DASH/WebRTC), and FFmpeg automation. It helps design and implement scalable video pipelines, optimize encoding, and integrate delivery systems with CDNs and players. The guidance emphasizes practical configurations, operational reliability, and performance trade-offs.

How this skill works

I inspect source media characteristics, propose quality ladders and encoding settings, and generate concrete FFmpeg commands and streaming manifests. For live and real-time scenarios I outline signaling, STUN/TURN and WebRTC adaptations; for on-demand I produce HLS/DASH packaging and CDN integration steps. I also recommend monitoring, error handling, and parallelization strategies to meet throughput and latency targets.

When to use it

  • Building or automating encoding and transcoding pipelines
  • Setting up HLS or DASH adaptive streaming infrastructure
  • Designing WebRTC-based real-time video applications
  • Automating FFmpeg workflows and batch processing
  • Optimizing codec choice, bitrates, and compression settings
  • Creating thumbnails, previews, or extracting video metadata

Best practices

  • Use hardware encoders (NVENC, QSV) when available to reduce latency and CPU cost
  • Implement adaptive bitrate (ABR) ladders rather than single-bitrate streams
  • Pre-generate quality levels for VOD; use low-latency options for live
  • Set appropriate keyframe interval to balance seekability and ABR switching
  • Keep HLS segment length between 2–10 seconds for smooth playback
  • Instrument encoding queues and delivery metrics; implement retries and alerts

Example use cases

  • Create a multi-bitrate HLS/DASH pipeline with automated FFmpeg transcoding and manifest generation
  • Build a WebRTC service with signaling, STUN/TURN and simulcast for heterogeneous clients
  • Automate batch video processing: thumbnailing, rewrapping, and codec conversion with retry logic
  • Optimize an existing workflow by switching to hardware encoding and tuning rate control parameters
  • Integrate a CDN with segmented delivery and player-side ABR testing

FAQ

Do you recommend H.264 or newer codecs like H.265/AV1?

Use H.264 for broad compatibility; choose H.265 or AV1 when bandwidth efficiency matters and client support is available. Consider licensing and hardware support when selecting codecs.

How should I choose HLS segment length?

Aim for 2–6 seconds for typical use. Shorter segments lower startup and latency but increase manifest/HTTP overhead; longer segments reduce overhead but raise latency.