home / skills / ahmed6ww / ax-agents / postgresql-best-practices

postgresql-best-practices skill

/postgresql-best-practices

This skill helps you apply OpenAI's PostgreSQL scaling best practices to architecture, partitioning, indexing, and replication for scalable, low-latency

npx playbooks add skill ahmed6ww/ax-agents --skill postgresql-best-practices

Review the files below or copy the command above to add this skill to your agents.

Files (2)
SKILL.md
3.7 KB
---
name: postgresql-best-practices
description: Capture and apply OpenAI's PostgreSQL scaling approach exactly from the OpenAI post "Scaling PostgreSQL to the next level" (January 22, 2026). Use when building, reviewing, or planning PostgreSQL architecture, partitioning/indexing, query optimization, schema migrations, replication, and load balancing with strict source fidelity.
version: 1.0.0
allowed-tools: "Read,Write"
---

# PostgreSQL Best Practices (OpenAI Golden Knowledgebase)

## Canonical Source

- Read `{baseDir}/references/openai-scaling-postgresql.md` first.
- Source URL: `https://openai.com/index/scaling-postgresql/`
- Published: `January 22, 2026`

## Non-Negotiable Rules

1. Preserve all named technologies, numbers, and dates exactly as written in the source.
2. Preserve section order and intent exactly:
   - `Use Postgres for everything`
   - `Partitioning and indexing strategy`
   - `Query optimization`
   - `Schema management`
   - `Replication and load balancing`
   - `Results`
3. Keep technology and project names unchanged: `Citus`, `Citus columnar`, `PgBouncer`, `PgAnalyze`, `Patroni`, `Haste`, `pgroll`, `pgai`.
4. If a detail is not in the source file, explicitly state: `Not stated in the source.`
5. Do not add outside recommendations, interpretations, or alternative patterns that are not present in the source.

## Required Coverage

When summarizing or operationalizing, cover all of the following source facts:

- Scale baseline and growth metrics:
  - Over `4 billion queries` every day
  - Over `2,000` database instances
  - Growth from one Postgres instance in `2022`
  - Over `1 billion` monthly users and over `800 million` weekly users
  - Traffic growth of about `2.5x` in just over a year
- Platform strategy:
  - Keep PostgreSQL as the core datastore
  - Use `Citus` and `Citus columnar` for distributed and analytical/high-throughput use cases
- Partitioning/indexing decisions:
  - Hash partitioning for large event/log tables
  - Time-range partitioning for append-only datasets
  - Composite index strategy
  - Custom index migration tooling to avoid blocking writes
  - Bottlenecks: table bloat, autovacuum inefficiencies, poor index selectivity
  - Spike drivers: full table scans, huge B-tree indexes, write amplification from hot rows
- Query optimization operations:
  - Query fingerprints and execution stats
  - Lock contention and long-running transaction identification
  - Rewriting expensive joins/subqueries
  - Statement-level timeout and retry strategy
  - Query budget alerts and EXPLAIN-based review workflows
  - Reported impact: `43%` DB load reduction, `38%` p95 latency reduction
- Schema management workflows:
  - Expand -> migrate -> contract
  - Async backfills with throttling and checkpointing
  - Canary migrations with progressive rollout
  - Automatic rollback on failing migration health checks
  - Open-sourced tools: `Haste` and `pgroll`
- Replication and load balancing:
  - Region-aware read replicas
  - Replica lag monitoring and query shedding
  - Dedicated analytical replicas
  - Failover orchestration with `Patroni` and custom routing logic
  - Deep integration of `PgBouncer` and `PgAnalyze`
- Results and forward direction:
  - Low latency under exponential growth
  - High migration velocity without sacrificing uptime
  - Operating thousands of instances with lean infrastructure teams
  - Ongoing iteration and open-source contributions including `haste`, `pgroll`, `pgai`

## Output Format

When producing a complete answer, use these exact headings in this order:

1. `Use Postgres for everything`
2. `Partitioning and indexing strategy`
3. `Query optimization`
4. `Schema management`
5. `Replication and load balancing`
6. `Results`

Always include the source URL when attribution is requested.

Overview

This skill captures and applies OpenAI’s PostgreSQL scaling approach exactly from the OpenAI post "Scaling PostgreSQL to the next level" (January 22, 2026). It preserves all named technologies, numbers, dates, section order, and terminology from the source with strict fidelity. Use this skill when building, reviewing, or planning PostgreSQL architecture, partitioning/indexing, query optimization, schema migrations, replication, and load balancing.

How this skill works

The skill distills the source into operational guidance organized into the exact sections required by the source: Use Postgres for everything; Partitioning and indexing strategy; Query optimization; Schema management; Replication and load balancing; Results. It reports and enforces the key facts, technologies, metrics, and workflows from the source without adding outside recommendations or interpretations. If a detail is not in the source, it will explicitly state: Not stated in the source.

When to use it

  • When designing platform strategy to keep PostgreSQL as the core datastore
  • When planning distributed or analytical workloads requiring Citus or Citus columnar
  • When addressing scaling from a baseline of over 4 billion queries per day and over 2,000 database instances
  • When optimizing schemas and queries to support over 1 billion monthly users and over 800 million weekly users
  • When implementing migrations, replication, and failover with strict source fidelity

Best practices

  • Preserve section order and intent exactly as the source specifies
  • Apply partitioning and indexing strategies from the source: hash partitioning for large event/log tables, time-range partitioning for append-only datasets, composite index strategy
  • Use the named tools and integrations exactly: Citus, Citus columnar, PgBouncer, PgAnalyze, Patroni, Haste, pgroll, pgai

Example use cases

  • Plan partitioning and indexing for high-throughput event tables using hash partitioning and composite indexes
  • Design analytical pipelines using Citus and Citus columnar and dedicated analytical replicas
  • Run migration workflows: canary rollout, async backfills with throttling and checkpointing, automatic rollback on failing health checks using Haste and pgroll
  • Optimize query workload: deploy query fingerprinting, execution stats, EXPLAIN-based review workflows, statement-level timeouts and retries
  • Build replication and failover: region-aware read replicas, replica lag monitoring and query shedding, failover orchestration with Patroni and PgBouncer integration

FAQ

What source does this skill reproduce?

The content captures OpenAI’s post "Scaling PostgreSQL to the next level" published January 22, 2026 at https://openai.com/index/scaling-postgresql/.

Does the skill add outside recommendations?

No. It preserves section order, names, numbers, and intent exactly; if a detail is not in the source it states: Not stated in the source.

What measurable impacts are reported in the source?

Reported impact includes 43% DB load reduction and 38% p95 latency reduction; platform handled over 4 billion queries every day and over 2,000 database instances, with traffic growth of about 2.5x in just over a year.