home / skills / leegonzales / aiskills / claimify

claimify skill

/Claimify/claimify

This skill extracts and maps claims from discourse into structured argument networks with supports, opposes, and assumptions to aid analysis.

npx playbooks add skill leegonzales/aiskills --skill claimify

Review the files below or copy the command above to add this skill to your agents.

Files (5)
SKILL.md
5.8 KB
---
name: claimify
description: Extract and structure claims from discourse into analyzable argument maps with logical relationships and assumptions. Use when analyzing arguments, red-teaming reasoning, synthesizing debates, or transforming conversations into structured claim networks. Triggers include "what are the claims," "analyze this argument," "map the logic," or "find contradictions."
---

# Claimify

Extract claims from text and map their logical relationships into structured argument networks.

## Overview

Claimify transforms messy discourse (conversations, documents, debates, meeting notes) into analyzable claim structures that reveal:
- Explicit and implicit claims
- Logical relationships (supports/opposes/assumes/contradicts)
- Evidence chains
- Argument structure
- Tension points and gaps

## Workflow

1. **Ingest**: Read source material (conversation, document, transcript)
2. **Extract**: Identify atomic claims (one assertion per claim)
3. **Classify**: Label claim types (factual/normative/definitional/causal/predictive)
4. **Map**: Build relationship graph (which claims support/oppose/assume others)
5. **Analyze**: Identify structure, gaps, contradictions, implicit assumptions
6. **Output**: Format as requested (table/graph/narrative/JSON)

## Claim Extraction Guidelines

### Atomic Claims
Each claim should be a single, testable assertion.

**Good:**
- "AI adoption increases productivity by 15-30%"
- "Psychological safety enables team learning"
- "Current training methods fail to build AI fluency"

**Bad (not atomic):**
- "AI is useful and everyone should use it" → Split into 2 claims

### Claim Types

| Type | Definition | Example |
|------|------------|---------|
| **Factual** | Empirical statement about reality | "Remote work increased 300% since 2020" |
| **Normative** | Value judgment or prescription | "Organizations should invest in AI training" |
| **Definitional** | Establishes meaning | "AI fluency = ability to shape context and evaluate output" |
| **Causal** | X causes Y | "Lack of training causes AI underutilization" |
| **Predictive** | Future-oriented | "AI adoption will plateau without culture change" |
| **Assumption** | Unstated premise | [implicit] "Humans resist change" |

### Relationship Types

- **Supports**: Claim A provides evidence/reasoning for claim B
- **Opposes**: Claim A undermines or contradicts claim B
- **Assumes**: Claim A requires claim B to be true (often implicit)
- **Refines**: Claim A specifies/clarifies claim B
- **Contradicts**: Claims are mutually exclusive
- **Independent**: No logical relationship

## Output Formats

### Table Format (default)

```markdown
| ID | Claim | Type | Supports | Opposes | Assumes | Evidence |
|----|-------|------|----------|---------|---------|----------|
| C1 | [claim text] | Factual | - | - | C5 | [source/reasoning] |
| C2 | [claim text] | Normative | C1 | C4 | - | [source/reasoning] |
```

### Graph Format

Use Mermaid for visualization:

```mermaid
graph TD
    C1[Claim 1: AI increases productivity]
    C2[Claim 2: Training is insufficient]
    C3[Claim 3: Organizations should invest]
    
    C1 -->|supports| C3
    C2 -->|supports| C3
    C2 -.->|assumes| C4[Implicit: Change requires structure]
```

### Narrative Format

Write as structured prose with clear transitions showing logical flow:

```markdown
## Core Argument

The author argues that [main claim]. This rests on three supporting claims:

1. [Factual claim] - This is supported by [evidence]
2. [Causal claim] - However, this assumes [implicit assumption]
3. [Normative claim] - This follows if we accept [prior claims]

## Tensions

The argument contains internal tensions:
- Claims C2 and C5 appear contradictory because...
- The causal chain from C3→C7 has a missing premise...
```

### JSON Format

For programmatic processing:

```json
{
  "claims": [
    {
      "id": "C1",
      "text": "AI adoption increases productivity",
      "type": "factual",
      "explicit": true,
      "supports": ["C3"],
      "opposed_by": [],
      "assumes": ["C4"],
      "evidence": "Multiple case studies cited"
    }
  ],
  "relationships": [
    {"from": "C1", "to": "C3", "type": "supports", "strength": "strong"}
  ],
  "meta_analysis": {
    "completeness": "Missing link between C2 and C5",
    "contradictions": ["C4 vs C7"],
    "key_assumptions": ["C4", "C9"]
  }
}
```

## Analysis Depth Levels

**Level 1: Surface**
- Extract only explicit claims
- Basic support/oppose relationships
- No implicit assumption mining

**Level 2: Standard** (default)
- Extract explicit claims
- Identify clear logical relationships
- Surface obvious implicit assumptions
- Flag apparent contradictions

**Level 3: Deep**
- Extract all claims (explicit + implicit)
- Map full logical structure
- Identify hidden assumptions
- Analyze argument completeness
- Red-team reasoning
- Suggest strengthening moves

## Best Practices

1. **Be charitable**: Steelman arguments before critique
2. **Distinguish**: Separate what's claimed from what's implied
3. **Be atomic**: One claim per line, no compound assertions
4. **Track evidence**: Note source/support for each claim
5. **Flag uncertainty**: Mark inferential leaps
6. **Mind the gaps**: Identify missing premises explicitly
7. **Stay neutral**: Describe structure before evaluating strength

## Common Patterns

### Argument Chains
```
Premise 1 (factual) → Premise 2 (causal) → Conclusion (normative)
```

### Implicit Assumptions
Often found by asking: "What must be true for this conclusion to follow?"

### Contradictions
Watch for:
- Same speaker, different times
- Different speakers, same topic
- Explicit vs implicit claims

### Weak Links
- Unsupported factual claims
- Causal claims without mechanism
- Normative leaps (is → ought)
- Definitional ambiguity

## Examples

See `references/examples.md` for detailed worked examples.

Overview

This skill extracts and structures claims from conversations, documents, or debates and turns them into analyzable argument maps. It reveals explicit and implicit assertions, logical relationships (supports/opposes/assumes/contradicts), evidence chains, and tension points so you can evaluate and improve reasoning.

How this skill works

Claimify ingests source text, isolates atomic claims (one testable assertion per claim), and classifies each by type (factual, normative, causal, definitional, predictive, assumption). It then builds a relationship graph linking claims (supports, opposes, assumes, refines, contradicts) and produces outputs in table, graph, narrative, or JSON formats. Optional depth levels control whether the analysis is surface-level, standard, or deep with red-teaming and hidden-assumption mining.

When to use it

  • Analyzing a debate or policy paper to surface core claims and gaps
  • Red-teaming reasoning or checking for contradictions across a thread or transcript
  • Synthesizing meeting notes into a clear argument structure and actionables
  • Preparing rebuttals or briefings by mapping supporting and opposing claims
  • Transforming unstructured conversation into JSON or graph for downstream tools

Best practices

  • Extract atomic claims: one clear assertion per claim
  • Be charitable first: steelman arguments before critique
  • Label claim types to clarify role in reasoning (factual, normative, causal, etc.)
  • Document evidence or source for each claim and flag uncertainty
  • Explicitly note assumptions and missing premises that enable claims
  • Use appropriate analysis depth: start standard, go deep for strategic red-teaming

Example use cases

  • Turn a debate transcript into a claim graph to find contradictions and leverage points
  • Convert stakeholder meeting notes into a prioritized claim table with evidence and assumptions
  • Evaluate a policy memo: map causal chains and surface missing premises before recommending changes
  • Create JSON outputs for programmatic argument analysis or visualization tools
  • Run a red-team pass on a product launch argument to expose weak links and hidden assumptions

FAQ

What output formats are available?

Table, graph (Mermaid), structured narrative, and JSON suitable for programmatic use.

How does depth affect results?

Surface extracts only explicit claims; Standard adds clear relationships and obvious assumptions; Deep mines implicit claims, hidden assumptions, and runs red-team style critiques.