home / skills / yuniorglez / gemini-elite-core / voice-ux-pro

voice-ux-pro skill

/skills/voice-ux-pro

This skill enables ultra-fast voice interfaces with sub-300ms latency, spatial hearing, and multimodal feedback for frictionless hands-free workflows.

npx playbooks add skill yuniorglez/gemini-elite-core --skill voice-ux-pro

Review the files below or copy the command above to add this skill to your agents.

Files (4)
SKILL.md
4.4 KB
---
name: voice-ux-pro
id: voice-ux-pro
version: 1.1.0
description: "Master of Voice-First Interfaces, specialized in sub-300ms Latency, Spatial Hearing AI, and Multimodal Voice-Haptic feedback."
last_updated: "2026-01-22"
---

# Skill: Voice UX Pro (Standard 2026)

**Role:** The Voice UX Pro is a specialized designer and engineer responsible for "Frictionless" conversational interfaces. In 2026, this role masters sub-300ms response times, Spatial Hearing AI (voice separation), and the integration of subtle haptic feedback to guide users through hands-free workflows.

## 🎯 Primary Objectives
1.  **Sub-300ms Responsiveness:** Achieving natural human-like interaction speeds using Streaming APIs and Edge Inference.
2.  **Spatial Clarity:** Implementing "Spatial Hearing AI" to isolate user voices from complex background noise.
3.  **Conversational Design:** Crafting non-linear, robust dialogues that handle interruptions and "Ums/Ahs" gracefully.
4.  **Multimodal Synergy:** Synchronizing Voice with Haptics and Visuals for a holistic, accessible experience.

---

## 🏗️ The 2026 Voice Stack

### 1. Speech Engines
- **Whisper v4 / Chirp v3:** For high-fidelity, multilingual transcription (STT).
- **Google Speech-to-Speech (S2S):** For near-instant, zero-latency response loops.
- **ElevenLabs v3:** For emotive, human-grade synthetic voices (TTS).

### 2. Interaction & Feedback
- **Native Haptics (iOS/Android):** Precise vibration patterns synchronized with speech phases.
- **Audio Shaders:** Real-time spatialization of AI voices using Shopify Skia or native audio APIs.

---

## 🛠️ Implementation Patterns

### 1. The "Listen-Ahead" Pattern (Sub-300ms)
Generating partial results while the user is still speaking to "Pre-warm" the LLM prompt.

```typescript
// 2026 Pattern: Streaming STT to LLM
const sttStream = await speechClient.createStreamingSTT();
const aiStream = await genAI.generateContentStream();

sttStream.on('partial', (text) => {
  // Pre-load context if 'intent' is detected early
  if (detectEarlyIntent(text)) aiStream.warmUp();
});
```

### 2. Voice-Haptic Synchronization
Providing "Micro-confirmation" via haptics when the AI starts/stops listening.

```tsx
import * as Haptics from 'expo-haptics';

function useVoiceInteraction() {
  const onStartListening = () => {
    // Light pulse to indicate "I am hearing you"
    Haptics.impactAsync(Haptics.ImpactFeedbackStyle.Light);
  };
  
  const onSuccess = () => {
    // Success sequence: Short, crisp double-tap
    Haptics.notificationAsync(Haptics.NotificationFeedbackType.Success);
  };
}
```

### 3. Spatial Isolation Logic
Isolating the user's voice based on 3D coordinates.

---

## 🚫 The "Do Not List" (Anti-Patterns)
1.  **NEVER** force the user to wait for a full sentence to be transcribed before acting.
2.  **NEVER** use "Robotic" monotonically generated voices. Use emotive TTS with prosody control.
3.  **NEVER** trigger loud audio confirmations in public settings without a "Silent Mode" check.
4.  **NEVER** ignore background noise. Always implement a "Noise-Floor" calibration step.

---

## 🛠️ Troubleshooting & Latency Audit

| Issue | Likely Cause | 2026 Corrective Action |
| :--- | :--- | :--- |
| **"Uncanny Valley" Delay** | Round-trip latency > 500ms | Move STT/TTS to a Regional Edge Function. |
| **Cross-Talk Failure** | Ambiguous sound sources | Implement Spatial Hearing AI (3D Beamforming). |
| **Instruction Fatigue** | Too many verbal options | Use "Contextual Shortlisting" (Only suggest relevant next steps). |
| **Accidental Triggers** | Sensitive Wake-word detection | Use "Personalized Voice Fingerprinting" for activation. |

---

## 📚 Reference Library
- **[Low-Latency Voice Stack](./references/1-low-latency-voice-stack.md):** STT, TTS, and S2S.
- **[Conversational Design](./references/2-conversational-design.md):** Beyond simple commands.
- **[Haptics & Multimodal](./references/3-haptics-and-multimodal.md):** Tactile feedback patterns.

---

## 📊 Performance Metrics
- **Interaction Latency:** < 300ms (Goal).
- **Word Error Rate (WER):** < 3% for noisy environments.
- **User Completion Rate:** > 90% for voice-only tasks.

---

## 🔄 Evolution from 2023 to 2026
- **2023:** Batch transcription, high latency, mono-visual.
- **2024:** Real-time streaming (Whisper Turbo).
- **2025-2026:** Spatial Hearing, Emotive S2S, and Haptic-Voice synchronization.

---

**End of Voice UX Pro Standard (v1.1.0)**

Overview

This skill equips engineers and designers to build frictionless voice-first interfaces with sub-300ms responsiveness, spatial hearing, and synchronized haptic feedback. It consolidates patterns for streaming STT/S2S, emotive TTS, and multimodal coordination to create natural, interruption-friendly conversational experiences. The guidance focuses on measurable outcomes like latency, WER, and completion rates.

How this skill works

The skill formalizes patterns such as Listen-Ahead streaming STT that emits partial transcripts to pre-warm LLM prompts, Spatial Hearing AI for voice separation and 3D beamforming, and tight voice-haptic synchronization for micro-confirmations. It prescribes moving inference to regional edge functions, using emotive TTS for prosody control, and implementing noise-floor calibration and silent-mode checks to avoid public disruptions.

When to use it

  • Building hands-free workflows where responsiveness must feel instantaneous (<300ms)
  • Designing systems that operate reliably in noisy, multi-speaker environments
  • Creating multimodal experiences that pair voice responses with subtle haptics and visuals
  • Implementing conversational agents that gracefully handle interruptions and filler words
  • Optimizing edge-deployed STT/TTS pipelines to reduce round-trip latency

Best practices

  • Stream partial STT results and warm LLM prompts on early intent detection
  • Deploy STT/TTS to regional edge functions to meet sub-300ms targets
  • Calibrate a noise floor and use spatial isolation before making decisions
  • Use emotive TTS with prosody controls, not monotonous synthetic voices
  • Provide Silent Mode and volume/haptics safeguards for public contexts
  • Shortlist contextual options to avoid instruction fatigue and reduce cognitive load

Example use cases

  • Voice-driven navigation in AR glasses with spatial audio and micro-haptics for turn cues
  • Hands-free industrial workflows where workers issue commands amid heavy machinery noise
  • In-car assistants that separate driver voice from passengers and maintain low latency
  • Accessibility-first apps that combine voice, haptics, and minimal visuals for blind or motor-impaired users
  • Customer service kiosks that detect and adapt to multi-speaker queues using spatial hearing

FAQ

How do I achieve sub-300ms latency in practice?

Move STT/TTS inference and short-context LLM operations to regional edge functions, stream partial transcripts, and warm downstream models on early intent signals.

What prevents accidental triggers in noisy environments?

Combine a calibrated noise-floor, personalized voice fingerprinting for wake activation, and spatial isolation to reduce false positives.