home / skills / dbt-labs / dbt-agent-skills / building-dbt-semantic-layer
/skills/building-dbt-semantic-layer
This skill guides creating and modifying dbt semantic layer components, including models, entities, dimensions, and metrics, with validation and spec guidance.
npx playbooks add skill dbt-labs/dbt-agent-skills --skill building-dbt-semantic-layerReview the files below or copy the command above to add this skill to your agents.
---
name: building-dbt-semantic-layer
description: Use when creating or modifying dbt Semantic Layer components — semantic models, metrics, dimensions, entities, measures, or time spines. Covers MetricFlow configuration, metric types (simple, derived, cumulative, ratio, conversion), and validation for both latest and legacy YAML specs.
user-invocable: false
metadata:
author: dbt-labs
---
# Building the dbt Semantic Layer
This skill guides the creation and modification of dbt Semantic Layer components: semantic models, entities, dimensions, and metrics.
- **Semantic models** - Metadata configurations that define how dbt models map to business concepts
- **Entities** - Keys that identify the grain of your data and enable joins between semantic models
- **Dimensions** - Attributes used to filter or group metrics (categorical or time-based)
- **Metrics** - Business calculations defined on top of semantic models (e.g., revenue, order count)
## Additional Resources
- [Time Spine Setup](references/time-spine.md) - Required for time-based metrics and aggregations
- [Best Practices](references/best-practices.md) - Design patterns and recommendations for semantic models and metrics
- [Latest Spec Authoring Guide](references/latest-spec.md) - Full YAML reference for dbt Core 1.12+ and Fusion
- [Legacy Spec Authoring Guide](references/legacy-spec.md) - Full YAML reference for dbt Core 1.6-1.11
## Determine Which Spec to Use
There are two versions of the Semantic Layer YAML spec:
- **Latest spec** - Semantic models are configured as metadata on dbt models. Simpler authoring. Supported by dbt Core 1.12+ and Fusion.
- **Legacy spec** - Semantic models are defined as separate top-level resources. Uses measures as building blocks for metrics. Supported by dbt Core 1.6 through 1.11. Also supported by Core 1.12+ for backwards compatibility.
### Step 1: Check for Existing Semantic Layer Config
Look for existing semantic layer configuration in the project:
- Top-level `semantic_models:` key in YAML files → **legacy spec**
- `semantic_model:` block nested under a model → **latest spec**
### Step 2: Route Based on What You Found
**If semantic layer already exists:**
1. Determine which spec is currently in use (legacy or latest)
2. Check dbt version for compatibility:
- **Legacy spec + Core 1.6-1.11** → Compatible. Use [legacy spec guide](references/legacy-spec.md).
- **Legacy spec + Core 1.12+ or Fusion** → Compatible, but offer to upgrade first using `uvx dbt-autofix deprecations --semantic-layer` or the [migration guide](https://docs.getdbt.com/docs/build/latest-metrics-spec). They don't have to upgrade; continuing with legacy is fine.
- **Latest spec + Core 1.12+ or Fusion** → Compatible. Use [latest spec guide](references/latest-spec.md).
- **Latest spec + Core <1.12** → Incompatible. Help them upgrade to dbt Core 1.12+.
**If no semantic layer exists:**
1. **Core 1.12+ or Fusion** → Use [latest spec guide](references/latest-spec.md) (no need to ask).
2. **Core 1.6-1.11** → Ask if they want to upgrade to Core 1.12+ for the easier authoring experience. If yes, help upgrade. If no, use [legacy spec guide](references/legacy-spec.md).
### Step 3: Follow the Spec-Specific Guide
Once you know which spec to use, follow the corresponding guide's implementation workflow (Steps 1-4) for all YAML authoring. The guides are self-contained with full examples.
## Entry Points
Users may ask questions related to building metrics with the semantic layer in a few different ways. Here are the common entry points to look out for:
### Business Question First
When the user describes a metric or analysis need (e.g., "I need to track customer lifetime value by segment"):
1. Search project models or existing semantic models by name, description, and column names for relevant candidates
2. Present top matches with brief context (model name, description, key columns)
3. User confirms which model(s) / semantic models to build on / extend / update
4. Work backwards from users need to define entities, dimensions, and metrics
### Model First
When the user specifies a model to expose (e.g., "Add semantic layer to `customers` model"):
1. Read the model SQL and existing YAML config
2. Identify the grain (primary key / entity)
3. Suggest dimensions based on column types and names
4. Ask what metrics the user wants to define
Both paths converge on the same implementation workflow.
### Open Ended
User asks to build the semantic layer for a project or models that are not specified. ("Build the semantic layer for my project")
1. Identify high importance models in the project
2. Suggest some metrics and dimensions for those models
3. Ask the user if they want to create more metrics and dimensions or if there are any other models they want to build the semantic layer on
## Metric Types
Both specs support these metric types. For YAML syntax, see the spec-specific guides.
### Simple Metrics
Directly aggregate a single column expression. The most common metric type and the building block for all others.
- **Latest spec**: Defined under `metrics:` on the model with `type: simple`, `agg`, and `expr`
- **Legacy spec**: Defined as top-level `metrics:` referencing a measure via `type_params.measure`
### Derived Metrics
Combine multiple metrics using a mathematical expression. Use for calculations like profit (revenue - cost) or growth rates (period-over-period with `offset_window`).
### Cumulative Metrics
Aggregate a metric over a running window or grain-to-date period. Requires a [time spine](references/time-spine.md). Use for running totals, trailing windows (e.g., 7-day rolling average), or period-to-date (MTD, YTD).
Note: `window` and `grain_to_date` cannot be used together on the same cumulative metric.
### Ratio Metrics
Create a ratio between two metrics (numerator / denominator). Use for conversion rates, percentages, and proportions. Both numerator and denominator can have optional filters.
### Conversion Metrics
Measure how often one event leads to another for a specific entity within a time window. Use for funnel analysis (e.g., visit-to-purchase conversion rate). Supports `constant_properties` to ensure the same dimension value across both events.
## Filtering Metrics
Filters can be added to simple metrics or metric inputs to advanced metrics. Use Jinja template syntax:
```
filter: |
{{ Entity('entity_name') }} = 'value'
filter: |
{{ Dimension('primary_entity__dimension_name') }} > 100
filter: |
{{ TimeDimension('time_dimension', 'granularity') }} > '2026-01-01'
filter: |
{{ Metric('metric_name', group_by=['entity_name']) }} > 100
```
**Important**: Filter expressions can only reference columns that are declared as dimensions or entities in the semantic model. Raw table columns that aren't defined as dimensions cannot be used in filters — even if they appear in a measure's `expr`.
## Validation
After writing YAML, validate in two stages:
1. **Parse Validation**: Run `dbt parse` (or `dbtf parse` for Fusion) to confirm YAML syntax and references
2. **Semantic Layer Validation**:
- `dbt sl validate` (dbt Cloud CLI or Fusion CLI when using the dbt platform)
- `mf validate-configs` (MetricFlow CLI)
**Important**: `mf validate-configs` reads from the compiled manifest, not directly from YAML files. If you've edited YAML since the last parse, you must run `dbt parse` (or `dbtf parse`) again before `mf validate-configs` will see the changes.
**Note**: When using Fusion with MetricFlow locally (without the dbt platform), `dbtf parse` will show `warning: dbt1005: Skipping semantic manifest validation due to: No dbt_cloud.yml config`. This is expected — use `mf validate-configs` for semantic layer validation in this setup.
Do not consider work complete until both validations pass.
## Editing Existing Components
When modifying existing semantic layer config:
- Check which spec is in use (see "Determine Which Spec to Use" above)
- Read existing entities, dimensions, and metrics before making changes
- Preserve all existing YAML content not being modified
- After edits, run full validation to ensure nothing broke
## Common Pitfalls (Both Specs)
| Pitfall | Fix |
|---------|-----|
| Missing time dimension | Every semantic model with metrics/measures needs a default time dimension |
| Using `window` and `grain_to_date` together | Cumulative metrics can only have one |
| Mixing spec syntax | Don't use `type_params` in latest spec or direct keys in legacy spec |
| Filtering on non-dimension columns | Filter expressions can only use declared dimensions/entities, not raw columns |
| `mf validate-configs` shows stale results | Re-run `dbt parse` / `dbtf parse` first to regenerate the manifest |
| MetricFlow install breaks `dbt-semantic-interfaces` | Install `dbt-metricflow` (not bare `metricflow`) to get compatible dependency versions |
This skill helps create and modify dbt Semantic Layer components: semantic models, entities, dimensions, metrics, and time spines. It covers choosing the correct YAML spec (latest vs legacy), common metric types (simple, derived, cumulative, ratio, conversion), and validation steps to ensure correctness. Use it to author consistent, validated semantic definitions that work with dbt Core and MetricFlow.
The skill inspects your project YAML and dbt models to detect whether a legacy top-level semantic_models key or a latest nested semantic_model block is present, then routes work to the appropriate authoring guide. It suggests entities and dimensions by analyzing model SQL and column types, builds metrics per spec syntax, and enforces validation steps (dbt parse + mf validate-configs or dbt sl validate). It also checks compatibility with dbt Core versions and offers upgrade guidance when necessary.
How do I choose latest vs legacy spec?
Check YAML for a top-level semantic_models key (legacy) or semantic_model nested under models (latest). Use latest with dbt Core 1.12+; legacy is supported for Core 1.6–1.11 and for backward compatibility.
What validation steps are required after editing YAML?
Run dbt parse (or dbtf parse) to regenerate the manifest, then run dbt sl validate or mf validate-configs. mf validate-configs reads the compiled manifest, so parse must run first to include edits.