home / skills / orchestra-research / ai-research-skills / creative-thinking-for-research
This skill helps researchers generate genuinely novel CS and AI ideas by applying cognitive science frameworks like combinatorial creativity and constraint
npx playbooks add skill orchestra-research/ai-research-skills --skill creative-thinking-for-researchReview the files below or copy the command above to add this skill to your agents.
---
name: creative-thinking-for-research
description: Applies cognitive science frameworks for creative thinking to CS and AI research ideation. Use when seeking genuinely novel research directions by leveraging combinatorial creativity, analogical reasoning, constraint manipulation, and other empirically grounded creative strategies.
version: 1.0.0
author: Orchestra Research
license: MIT
tags: [Creative Thinking, Research Ideation, Analogical Reasoning, Problem Reformulation, Cognitive Science]
dependencies: []
---
# Creative Thinking for Research
Eight empirically grounded frameworks from cognitive science, applied to computer science and AI research. Unlike ad-hoc brainstorming, each framework here is backed by decades of creativity research — from Koestler's bisociation to Kauffman's adjacent possible. They target distinct cognitive operations: combining, reformulating, analogizing, constraining, inverting, abstracting, exploring boundaries, and holding contradictions.
## When to Use This Skill
- Generating genuinely novel ideas, not incremental extensions of prior work
- Feeling trapped in a local optimum of thinking within a single subfield
- Wanting to systematically apply creativity heuristics rather than waiting for inspiration
- Preparing for a research retreat or PhD-level ideation session
- Bridging between fields and seeking structural (not superficial) connections
**Do NOT use this skill when**:
- You need structured project-level brainstorming workflows (use `brainstorming-research-ideas`)
- You have a well-defined problem and need execution help (use domain-specific skills)
- You need a literature survey (use `scientific-skills:literature-review`)
**Relationship to Brainstorm skill**: The brainstorm skill provides operational workflows (diverge → converge → refine) and practical filters. This skill provides the deeper cognitive engines that power creative leaps. Use them together: creative-thinking to generate raw insight, brainstorm to structure and evaluate it.
---
## Framework 1: Combinatorial Creativity (Bisociation)
Novel ideas arise from combining existing concepts in unexpected ways. Arthur Koestler called this **bisociation** — connecting two previously unrelated frames of reference, as distinct from routine association within a single frame.
**Why it works**: Meta-research consistently shows that breadth of knowledge is a precursor to creative output. People who read across disciplines produce more novel work. The combination itself is the creative act.
**In CS Research**:
- Biological evolution → optimization (genetic algorithms)
- Game theory → networking (mechanism design for routing)
- Statistical physics → machine learning (Boltzmann machines, energy-based models)
- Linguistics → programming (type theory, formal grammars)
**Systematic Bisociation Workflow**:
1. **Select two domains** you have at least passing familiarity with
2. **List core primitives** in each domain (5-10 fundamental concepts per domain)
3. **Create a cross-product matrix**: row = concepts from Domain A, column = concepts from Domain B
4. **For each cell**, ask: "What would it mean to apply A's concept to B's problem?"
5. **Filter**: Which combinations produce a non-trivial, testable research question?
6. **Validate structural depth**: Is the connection mechanistic or merely metaphorical?
**Cross-Product Example**:
| | Caching | Load Balancing | Fault Tolerance |
|---|---------|---------------|-----------------|
| **Natural Selection** | Evict least-fit entries | Adaptive allocation via fitness | Population-level redundancy |
| **Immune Memory** | Learned threat signatures | Distributed detection | Self/non-self discrimination |
| **Symbiosis** | Cooperative prefetching | Mutualistic resource sharing | Co-dependent resilience |
**Quality Test**: A strong bisociation is not a surface metaphor ("the network is like a brain") but a structural mapping where the mechanism transfers ("attention mechanisms implement a form of selective gating analogous to cognitive attention filtering").
**Self-Check**:
- [ ] Is the connection structural (mechanisms map) or merely verbal (labels map)?
- [ ] Does the combination generate testable predictions?
- [ ] Would an expert in both fields find the connection non-obvious but sound?
---
## Framework 2: Problem Reformulation (Representational Change)
Gestalt psychologists identified that breakthroughs often come not from solving the problem as stated, but from **re-representing the problem itself**. Kaplan and Simon's work on insight shows that changing the problem space — the constraints, the abstraction level, the formalism — is often where creativity lives.
**The Key Shift**: From "How do I solve this problem?" to "Am I even thinking about this problem correctly?"
**Reformulation Strategies**:
| Strategy | Example |
|----------|---------|
| **Change the objective** | "Make the algorithm faster" → "Eliminate the need for this computation" |
| **Change the formalism** | Graph problem → linear algebra problem (spectral methods) |
| **Change the granularity** | Per-token prediction → per-span prediction |
| **Change the agent** | "How should the model learn?" → "How should the data teach?" (curriculum learning) |
| **Change the timescale** | Real-time optimization → amortized inference |
| **Invert the direction** | Forward simulation → inverse problem (learning from observations) |
**Workflow**:
1. State your current problem in one sentence
2. Identify the **hidden assumptions** in that statement:
- What formalism are you using? (Could you use a different one?)
- What is the objective? (Is it the right objective?)
- What level of granularity? (Could you go coarser or finer?)
- Who is the agent? (Could you shift perspective?)
3. For each assumption, **generate the alternative**: "What if [opposite assumption]?"
4. For each alternative, ask: "Does this reformulation make the problem easier, harder, or different in a useful way?"
5. A reformulation that makes a hard problem easy is often a publishable insight on its own
**Classic CS Examples**:
- **PageRank**: Reformulated "find important web pages" from content analysis to graph eigenvalue problem
- **Dropout**: Reformulated "prevent overfitting" from regularization to approximate ensemble
- **Attention**: Reformulated "handle long sequences" from remembering everything to selectively querying
---
## Framework 3: Analogical Reasoning (Structure-Mapping)
Dedre Gentner's **structure-mapping theory** and Kevin Dunbar's studies of real scientists show that analogy is the core engine of scientific creativity. The critical finding: surface-level analogies are common but weak; **structural or relational analogies** — where the deep causal/relational structure maps across domains — produce the most powerful insights.
**Dunbar's Finding**: In the most successful labs, analogies from distant domains drove the most important discoveries. Nearby analogies refined ideas; distant analogies generated them.
**Levels of Analogical Depth**:
| Level | Description | Value | Example |
|-------|-------------|-------|---------|
| **Surface** | Things look similar | Low | "A neural network is like a brain" |
| **Relational** | Relationships between entities match | Medium | "Attention allocation in models parallels resource allocation in economics" |
| **Structural** | Deep causal mechanisms map | High | "Diffusion models reverse a thermodynamic process; the math of non-equilibrium stat-mech directly applies" |
**Structure-Mapping Workflow**:
1. **Describe your problem** using only relational/causal language (strip domain-specific nouns)
- Bad: "We need to improve transformer attention efficiency"
- Good: "We have a system that must selectively aggregate information from a large set, where relevance is context-dependent and the cost scales quadratically with set size"
2. **Search for structural matches**: What other systems selectively aggregate from large sets?
- Database query optimization, visual attention in neuroscience, information retrieval, resource allocation
3. **Pick the most distant match** with genuine structural fidelity
4. **Map the solution mechanism**: How does the source domain solve this?
5. **Transfer and adapt**: What changes when you bring that mechanism into your domain?
6. **Generate predictions**: The analogy should tell you something you didn't already know
**Validation Checklist**:
- [ ] Does the mapping preserve causal/relational structure (not just labels)?
- [ ] Can I identify at least one prediction the analogy makes in my domain?
- [ ] Would an expert in the source domain confirm the mechanism is correctly understood?
- [ ] Is the analogy non-obvious to my target audience?
---
## Framework 4: Constraint Manipulation (Boden's Framework)
Margaret Boden's framework distinguishes three forms of creativity based on how they interact with constraints:
| Type | Operation | CS Example |
|------|-----------|------------|
| **Exploratory** | Search within the existing conceptual space | Hyperparameter tuning, architecture search within a fixed paradigm |
| **Combinational** | Combine elements from different spaces | Multi-task learning, neuro-symbolic methods |
| **Transformational** | Change the rules of the space itself | Dropping the assumption that training requires labels (self-supervised learning) |
**Transformational creativity is the rarest and highest-impact.** It happens when you change what is even considered a valid solution.
**Constraint Analysis Workflow**:
1. **List the constraints** of your current approach (5-10 constraints):
- Computational: "Must fit in GPU memory"
- Methodological: "Requires labeled data"
- Architectural: "Uses fixed-length context"
- Evaluative: "Measured by accuracy on benchmark X"
2. **Classify each constraint**:
- **Hard**: Physically or logically necessary (cannot violate)
- **Soft**: Convention or historical accident (can question)
- **Hidden**: Not stated but implicitly assumed (most fertile for innovation)
3. **For each soft/hidden constraint**, ask:
- What if we relaxed it? (streaming algorithms from relaxing "fits in memory")
- What if we tightened it? (efficiency research from tightening compute budgets)
- What if we replaced it with a different constraint entirely?
4. **The most productive move** is often exposing and dropping a hidden constraint
**Classic Examples of Constraint Transformation**:
- "Data must fit in memory" → dropped → streaming algorithms, external memory
- "Training requires human labels" → dropped → self-supervised learning
- "Models must be deterministic" → dropped → variational methods, diffusion
- "Inference must happen in one pass" → dropped → iterative refinement, chain-of-thought
---
## Framework 5: Negation and Inversion
Take a core assumption in your field and negate it. This is formalized in De Bono's lateral thinking and the **TRIZ methodology** from engineering.
**The Pattern**: "What if [widely held assumption] is wrong, unnecessary, or invertible?"
**Systematic Negation Workflow**:
1. **List 5-10 core assumptions** in your subfield (the things "everyone knows")
2. **Negate each one** and ask: What system would you build?
3. **Evaluate each negation**:
- Incoherent → discard
- Already explored → check if conditions have changed (see brainstorm skill, Framework 5)
- Unexplored and coherent → potential research direction
**Negation Hall of Fame in CS**:
| Assumption | Negation | Result |
|-----------|----------|--------|
| "We need strong consistency" | What if we don't? | Eventual consistency, CRDTs |
| "We need exact answers" | What if approximate is fine? | Sketches, LSH, approximate nearest neighbors |
| "Labels are necessary" | What if we learn without them? | Self-supervised learning, contrastive methods |
| "More parameters = more compute" | What if we don't use all parameters? | Mixture of Experts, sparse models |
| "Training and inference are separate" | What if the model keeps learning? | Online learning, test-time training |
| "Errors must be prevented" | What if we embrace and correct them? | Speculative decoding, self-correction |
**TRIZ-Inspired Principles for CS**:
| TRIZ Principle | CS Application |
|---------------|----------------|
| **Inversion** | Reverse the process (generative vs. discriminative) |
| **Segmentation** | Break monolithic into modular (microservices, mixture of experts) |
| **Merging** | Combine separate steps (end-to-end learning) |
| **Universality** | One component serves multiple functions (multi-task models) |
| **Nesting** | Place one system inside another (meta-learning) |
| **Dynamization** | Make static things adaptive (dynamic architectures, adaptive computation) |
---
## Framework 6: Abstraction and Generalization Laddering
Moving up and down the abstraction ladder is a fundamental creative act. Polya's heuristics formalize this: *"Can you solve a more general problem? A more specific one? An analogous one?"*
**Three Moves**:
| Move | Question | Outcome |
|------|----------|---------|
| **Generalize** | "Is my solution a special case of something broader?" | Framework papers, unifying theories |
| **Specialize** | "What happens when I add extreme constraints?" | Niche applications, surprising edge cases |
| **Analogize** | "Where else does this abstract pattern appear?" | Cross-domain transfer (see Framework 3) |
**Generalization Workflow**:
1. State your specific result
2. Replace each specific element with a variable: "ResNet works for ImageNet" → "Architecture X works for distribution Y"
3. Ask: Under what conditions does this hold? What is the general principle?
4. If the general principle is novel → that is the contribution
**Specialization Workflow**:
1. Take a general method
2. Add extreme constraints: tiny data, huge dimensionality, adversarial inputs, real-time requirements
3. Ask: Does the method still work? If not, why not?
4. The failure case often reveals the method's true assumptions
**When to Generalize vs. Specialize**:
- Generalize when you have results but no explanation
- Specialize when you have theory but no grounding
- Analogize when you are stuck in either direction
---
## Framework 7: The Adjacent Possible (Kauffman / Johnson)
Stuart Kauffman's concept, popularized by Steven Johnson: innovation happens at the boundary of what is currently reachable — the **adjacent possible**. New ideas become thinkable once their prerequisites exist. This explains why simultaneous independent discovery is so common — multiple people reach the same boundary.
**Practical Implication**: Map what has recently become possible and explore the space those enablers open.
**Adjacent Possible Mapping Workflow**:
1. **List recent enablers** (last 1-3 years):
- New hardware capabilities (longer context, faster inference, new accelerators)
- New datasets or benchmarks
- New open-source tools or frameworks
- New theoretical results
- New regulatory or social conditions
2. **For each enabler, ask**: "What was previously impossible or impractical that this now permits?"
3. **Combine enablers**: The most powerful adjacent possibles arise from the intersection of multiple new enablers
4. **Check for competition**: If many people can see the same adjacent possible, speed or a unique angle matters
**Current Adjacent Possibles (2025-2026)**:
| Enabler | Newly Possible |
|---------|---------------|
| 1M+ token context windows | Full-codebase reasoning, book-length analysis |
| Inference cost drops (100x in 2 years) | Real-time agentic loops, always-on AI assistants |
| Open-weight models at GPT-4 level | Reproducible research on frontier capabilities |
| Multimodal models (vision + language + audio) | Unified perception-reasoning systems |
| Synthetic data at scale | Training data for domains with no natural data |
| Tool-using models | Research automation, self-improving systems |
**Timing Signal**: If your idea requires technology that doesn't exist yet, it's beyond the adjacent possible — park it. If your idea could have been done 5 years ago, someone probably did — check the literature. The sweet spot is ideas that became feasible in the last 6-18 months.
---
## Framework 8: Janusian and Dialectical Thinking
Albert Rothenberg's studies of eminent creators found that **holding two contradictory ideas simultaneously** is a hallmark of creative thinking. Named after Janus, the two-faced Roman god, this mode of thinking doesn't resolve contradictions by choosing a side — it generates new frameworks that transcend the opposition.
**In CS**: The most influential results often emerge from tensions previously thought irreconcilable.
| Contradiction | Resolution | Impact |
|--------------|------------|--------|
| Consistency AND Availability (distributed systems) | CAP theorem: formalized the trade-off, then Raft/CRDTs found practical middle grounds | Foundation of distributed systems theory |
| Security AND Usability | Zero-knowledge proofs: prove knowledge without revealing it | Enabled private computation |
| Expressiveness AND Tractability | Probabilistic programming: express complex models, automate inference | New programming paradigm |
| Memorization AND Generalization | Grokking: models memorize first, then generalize with more training | New understanding of learning dynamics |
| Compression AND Quality | Neural codecs that compress beyond information-theoretic limits via learned priors | Redefined compression research |
**Dialectical Thinking Workflow**:
1. **Identify a binary** in your field: A vs. B (two approaches, goals, or paradigms treated as opposites)
2. **Resist choosing a side**. Instead ask:
- "What would a system look like that achieves both A and B?"
- "Under what conditions is the A-B trade-off not fundamental?"
- "Is the opposition an artifact of how we formalized the problem?"
3. **Seek synthesis**: The resolution often requires a new abstraction that reframes the relationship
4. **Test the synthesis**: Can you demonstrate empirically that both goals are achievable?
**Self-Check**:
- [ ] Am I holding the contradiction genuinely (not prematurely resolving it)?
- [ ] Is the synthesis a new idea, not just a compromise (splitting the difference)?
- [ ] Does the resolution change how people think about the problem, not just the solution?
---
## Combining Frameworks: A Creative Thinking Protocol
These frameworks are most powerful in combination. Here is a systematic protocol for a deep creative thinking session:
### Phase 1: Map the Space (15 min)
1. **Constraint Manipulation** (F4): List all constraints of the current paradigm. Mark which are hard, soft, hidden.
2. **Adjacent Possible** (F7): List recent enablers that change the feasibility landscape.
### Phase 2: Generate Disruptions (30 min)
3. **Negation** (F5): Negate 3 soft/hidden constraints. What systems emerge?
4. **Bisociation** (F1): Pick a distant field and create a cross-product matrix with your domain.
5. **Problem Reformulation** (F2): Restate your problem 3 different ways (change objective, formalism, agent).
### Phase 3: Deepen Promising Leads (30 min)
6. **Analogical Reasoning** (F3): For each promising idea, find a structural analogy and extract predictions.
7. **Abstraction Laddering** (F6): Move each idea up (generalize) and down (specialize).
8. **Janusian Thinking** (F8): Identify any tensions. Can you synthesize rather than choose?
### Phase 4: Evaluate (15 min)
Apply the two-sentence test (from the brainstorm skill):
> "**[Domain] currently struggles with [problem] because [reason].** We [approach] by [mechanism], which works because [insight]."
Any idea that survives all four phases and passes the two-sentence test is worth pursuing.
---
## Common Creative Blocks and Unblocking Strategies
| Block | Symptom | Framework to Apply |
|-------|---------|-------------------|
| **Fixation** | Cannot stop thinking about the problem one way | Problem Reformulation (F2) — force a different representation |
| **Tunnel vision** | All ideas come from the same subfield | Bisociation (F1) or Analogical Reasoning (F3) — import from elsewhere |
| **Self-censoring** | Dismissing ideas as "too weird" before exploring | Negation (F5) — weird is the point; evaluate after generating |
| **Incrementalism** | Every idea is "+2% on benchmark X" | Constraint Manipulation (F4) — change the rules, not the parameters |
| **Analysis paralysis** | Too many options, cannot commit | Adjacent Possible (F7) — what is feasible right now? |
| **False dichotomy** | Stuck choosing between two approaches | Janusian Thinking (F8) — seek synthesis, not selection |
---
## Usage Instructions for Agents
When a researcher asks for help with creative thinking or novel ideation:
1. **Assess the block**: What kind of thinking are they stuck in? (See Common Creative Blocks table)
2. **Select 2-3 frameworks** based on the block type
3. **Walk through each framework interactively**, asking the researcher to supply domain-specific content
4. **Push for structural depth**: If an analogy or combination is surface-level, probe deeper
5. **Maintain a running list** of all generated ideas, even unusual ones
6. **Apply the two-sentence test** to candidates that survive exploration
7. **Hand off to the brainstorm skill** for systematic evaluation (diverge → converge → refine)
**Key Principles**:
- Generative mode first, evaluative mode second — do not filter prematurely
- Distant analogies are more valuable than nearby ones, but require more validation
- The researcher's domain expertise is essential — the agent provides the cognitive scaffolding, not the domain knowledge
- Encourage the researcher to sit with contradictions rather than resolve them quickly
This skill applies eight empirically grounded cognitive-science frameworks to spark genuinely novel computer science and AI research directions. It packages structured heuristics—combinatorial combination, analogical mapping, constraint manipulation, inversion, abstraction laddering, and more—so you can move beyond incremental tweaks. Use it to generate research hypotheses that are mechanistic, testable, and non-obvious.
The skill inspects your current problem statement and cognitive framing, then systematically applies different creative engines: bisociation (cross-domain combination), representational change, structural analogy, constraint analysis, negation/inversion, abstraction/generalization, and exploration of the adjacent possible. Each framework provides a short workflow: how to transform assumptions, produce candidate hypotheses, and validate structural depth and testability. Outputs are filtered by novelty, mechanistic plausibility, and feasibility for follow-up experiments.
Is this a replacement for systematic literature review?
No. Use this skill to generate structural hypotheses; pair it with a literature review to check novelty and prior work.
How do I judge whether a cross-domain mapping is substantive?
Require a structural mapping that yields at least one clear, testable prediction and check with experts in both domains.