home / skills / jeremylongshore / claude-code-plugins-plus-skills / setting-up-distributed-tracing
/plugins/performance/distributed-tracing-setup/skills/setting-up-distributed-tracing
This skill helps you automate distributed tracing setup for microservices, configuring backend, instrumentation, and context propagation to enable end-to-end
npx playbooks add skill jeremylongshore/claude-code-plugins-plus-skills --skill setting-up-distributed-tracingReview the files below or copy the command above to add this skill to your agents.
---
name: setting-up-distributed-tracing
description: |
Execute this skill automates the setup of distributed tracing for microservices. it helps developers implement end-to-end request visibility by configuring context propagation, span creation, trace collection, and analysis. use this skill when the user re... Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.
allowed-tools: Read, Write, Edit, Grep, Glob, Bash(cmd:*)
version: 1.0.0
author: Jeremy Longshore <[email protected]>
license: MIT
---
# Distributed Tracing Setup
This skill provides automated assistance for distributed tracing setup tasks.
## Overview
This skill streamlines the process of setting up distributed tracing in a microservices environment. It guides you through the key steps of instrumenting your services, configuring trace context propagation, and selecting a backend for trace collection and analysis, enabling comprehensive monitoring and debugging.
## How It Works
1. **Backend Selection**: Determines the preferred tracing backend (e.g., Jaeger, Zipkin, Datadog).
2. **Instrumentation Strategy**: Designs an instrumentation strategy for each service, focusing on key operations and dependencies.
3. **Configuration Generation**: Generates the necessary configuration files and code snippets to enable distributed tracing.
## When to Use This Skill
This skill activates when you need to:
- Implement distributed tracing in a microservices application.
- Gain end-to-end visibility into request flows across multiple services.
- Troubleshoot performance bottlenecks and latency issues.
## Examples
### Example 1: Adding Tracing to a New Microservice
User request: "setup tracing for the new payment service"
The skill will:
1. Prompt for the preferred tracing backend (e.g., Jaeger).
2. Generate code snippets for OpenTelemetry instrumentation in the payment service.
### Example 2: Troubleshooting Performance Issues
User request: "implement distributed tracing to debug slow checkout process"
The skill will:
1. Guide the user through instrumenting relevant services in the checkout flow.
2. Provide configuration examples for context propagation.
## Best Practices
- **Backend Choice**: Select a tracing backend that aligns with your existing infrastructure and monitoring tools.
- **Sampling Strategy**: Implement a sampling strategy to manage trace volume and cost, especially in high-traffic environments.
- **Context Propagation**: Ensure proper context propagation across all services to maintain trace continuity.
## Integration
This skill can be used in conjunction with other plugins to automate the deployment and configuration of tracing infrastructure. For example, it can integrate with infrastructure-as-code tools to provision Jaeger or Zipkin clusters.
## Prerequisites
- Appropriate file access permissions
- Required dependencies installed
## Instructions
1. Invoke this skill when the trigger conditions are met
2. Provide necessary context and parameters
3. Review the generated output
4. Apply modifications as needed
## Output
The skill produces structured output relevant to the task.
## Error Handling
- Invalid input: Prompts for correction
- Missing dependencies: Lists required components
- Permission errors: Suggests remediation steps
## Resources
- Project documentation
- Related skills and commandsThis skill automates the setup of distributed tracing for microservices to deliver end-to-end request visibility. It helps instrument services, configure context propagation, and generate backend-specific configuration for trace collection. The goal is faster debugging, clearer performance insights, and repeatable tracing deployments.
The skill selects a tracing backend (e.g., Jaeger, Zipkin, Datadog) based on your inputs and environment. It designs an instrumentation plan for each service and creates code snippets and config files for OpenTelemetry-compatible tracing. It also produces context-propagation patterns, sampling recommendations, and deployment hints for trace collectors.
Which backends are supported?
Common backends like Jaeger, Zipkin, and Datadog are supported via OpenTelemetry exporters. I can tailor config for each target.
Will this change service code automatically?
I generate code snippets and configuration for you to apply. Automatic edits can be done only if you provide consent and file access.