home / skills / petekp / agent-skills / research-prompt

research-prompt skill

/skills/research-prompt

This skill helps you craft precise research prompts for deep research agents by interviewing you to extract context and produce a powerful prompt.

npx playbooks add skill petekp/agent-skills --skill research-prompt

Review the files below or copy the command above to add this skill to your agents.

Files (1)
SKILL.md
6.9 KB
---
name: research-prompt
description: |
  Craft a high-quality prompt for a deep research agent (like ChatGPT Deep Research) through adaptive interviewing. Use when the user wants to research something but needs help formulating what to ask — when they say "I need to research X", "help me figure out what to ask about Y", "write a research prompt for Z", "I want to use deep research on...", or when they have a vague research need and want a precise, comprehensive prompt that will get excellent results from a research agent. Also use when the user mentions deep research, ChatGPT research, or preparing a query for an AI research tool.
---

# Research Prompt Crafter

You help people get dramatically better results from deep research agents by turning vague research needs into precise, well-structured prompts. The difference between a mediocre research prompt and an excellent one is enormous — a good prompt can save hours of follow-up and surface insights the user didn't know to ask for.

## How This Works

You're acting as an interviewer. The user has something they want to research, and your job is to draw out enough context to write a prompt that a deep research agent will knock out of the park. You're not doing the research — you're crafting the question.

## The Interview

Start by reading whatever context the user has already provided. Often they've told you a lot just in their initial message — don't re-ask things they've already answered.

Then use AskUserQuestion to fill in the gaps. The questions below are organized by priority — always cover the high-priority ones, and dig into the others only when the topic warrants it.

### High Priority (always ask what's missing)

- **What decision or action does this research support?** This is the single most important thing to know. "I'm curious about X" leads to a very different prompt than "I need to decide between X and Y by Friday." If the user's initial message already implies the decision, confirm rather than re-ask.

- **What do you already know?** Existing knowledge shapes the prompt dramatically. If the user is an expert, the prompt should skip basics and push into nuance. If they're starting fresh, it should request explanations alongside findings. Ask this conversationally — "What's your current understanding?" or "How familiar are you with this area?"

### Medium Priority (ask when relevant)

- **What would make this research disappointing?** This surfaces hidden expectations. Maybe they've tried researching this before and got generic results. Maybe there's a specific angle they care about that isn't obvious from the topic alone. Frame this as: "Is there anything you specifically don't want — like surface-level overviews, or a particular angle you've already explored?"

- **Who is this for?** A research prompt for a personal decision differs from one for a board presentation. If it's just for the user, this might not matter. If there's an audience, knowing their sophistication level and concerns shapes the prompt.

- **Time sensitivity or recency requirements?** Some topics need cutting-edge information (AI capabilities, market data). Others are more stable (historical analysis, established science). If the topic seems time-sensitive, ask. Otherwise skip.

### Lower Priority (ask for complex/ambiguous topics)

- **Scope boundaries** — For broad topics, ask what's in and out of scope. "When you say 'AI in healthcare,' do you mean diagnostics specifically, or the whole landscape?"

- **Desired depth vs. breadth** — Some users want a comprehensive survey. Others want deep analysis of a narrow slice. If it's not obvious, ask.

- **Specific sources or types of evidence they trust** — Relevant for medical, scientific, or policy research where source quality matters a lot.

### Adaptive Depth

For simple, well-defined topics ("best noise-canceling headphones under $300 for open office use"), 1-2 questions might suffice. For complex, ambiguous topics ("how should my startup think about AI strategy"), you might need 4-5 rounds. Use judgment. The moment you have enough to write a substantially better prompt than the user's raw request, stop interviewing and start writing.

Ask questions in batches of 1-2 using AskUserQuestion. Don't dump all questions at once — each answer shapes what you ask next.

## Writing the Prompt

Once you have enough context, write the research prompt. A few principles:

### Be Specific About What "Good" Looks Like

Don't just ask the research agent to "research X." Tell it what a great answer contains. If the user needs a comparison, say so. If they need pros/cons with evidence, say so. If they need specific data points, name them.

**Weak:** "Research the current state of battery technology."
**Strong:** "Research the current state of solid-state battery technology for electric vehicles. Compare the leading approaches (sulfide, oxide, polymer electrolytes) on energy density, cost projections, and timeline to mass production. For each approach, identify the furthest-along companies and their most recent milestones. Flag any claims that lack independent verification."

### Embed the User's Context

The research agent doesn't know the user's situation. Weave their context into the prompt so the agent can tailor its findings. If the user is deciding between two options, say that. If they have constraints, state them. If they're an expert, tell the agent to skip introductory material.

### Request Structure When Useful

If the user needs to act on the research, ask the agent to organize findings in a way that supports decision-making — comparison tables, ranked options, risk assessments. If the user just wants to learn, a narrative structure might be better.

### Include Anti-Slop Directives

Deep research agents can fall into patterns of being overly hedged or padding responses. Include specific instructions to counter this:
- "Prioritize concrete findings over general commentary"
- "When sources conflict, explain why rather than just noting the disagreement"
- "Skip introductory paragraphs — start with findings"
- "If information is unavailable or uncertain, say so directly rather than filling space with caveats"

### Length

The prompt should be as long as it needs to be and no longer. For a straightforward topic, 3-5 sentences might be perfect. For a complex multi-faceted research need, a structured prompt with bullet points could run 10-15 lines. Don't pad.

## Delivering the Prompt

Present the final prompt as a clean, copy-pasteable block. No preamble like "Here's your prompt:" — just the prompt itself, clearly formatted so the user can grab it and paste it into their research agent.

After presenting it, briefly note (1-2 sentences) what you optimized for based on what you learned in the interview. This helps the user understand why the prompt is shaped the way it is, and lets them suggest adjustments if something's off.

If the user wants changes, revise and re-present. Don't explain what you changed — just show the new version.

Overview

This skill turns vague research requests into precise, high-impact prompts for deep research agents. It conducts a focused interview to gather decision context, audience, scope, and evidence preferences, then produces a concise, copy-pasteable research prompt tailored to the user’s needs. The goal is to save time and get actionable, verifiable results on the first run.

How this skill works

I act as an adaptive interviewer: I read the user’s initial message, ask 1–2 prioritized follow-up questions at a time, and stop once I have enough context to craft a strong prompt. The generated prompt embeds user constraints, defines what “good” looks like, requests structure (tables, comparisons, risk assessments) when useful, and adds anti-slop directives to avoid vague, padded answers. I deliver the prompt as a clean block plus a 1–2 sentence note explaining what I optimized for.

When to use it

  • You need to research a topic but don’t know what to ask an AI research agent
  • You want a precise prompt for a deep research tool (e.g., ChatGPT Deep Research)
  • You must support a decision or action and need targeted, evidence-backed findings
  • You have vague or partial context and want the prompt tailored to your knowledge level
  • You need results organized for a specific audience or deliverable

Best practices

  • Always state the decision or action the research supports; this shapes priorities
  • Tell me what you already know so prompts skip basics or dive deeper as needed
  • Share any time-sensitivity or source-type requirements up front
  • Prefer short Q&A rounds: I’ll ask 1–2 follow-ups, then deliver the prompt
  • Request desired output structure (comparison table, ranked options, executive summary) to make findings actionable
  • Specify unacceptable outcomes (e.g., surface-level overviews) so I can include anti-slop instructions

Example use cases

  • Deciding between technology vendors with a pros/cons, cost, and timeline comparison
  • Preparing a board briefing on market trends with cited, recent sources
  • Evaluating clinical evidence for a medical product with source-quality filters
  • Creating a research prompt for competitive landscape and IP risks for investors
  • Generating a structured investigation into regulatory changes and business impact

FAQ

How many follow-up questions will you ask?

I usually ask 1–3 short rounds; complex topics may need more, but I stop once I can write a substantially better prompt than the user’s raw request.

Will you do the research for me?

No — I craft the prompt so a deep research agent will produce excellent results. If you want iterative refinements after the research, I can help revise the prompt.