Back to skills
SkillHub ClubWrite Technical DocsTech Writer

claimify

Extract and structure claims from discourse into analyzable argument maps with logical relationships and assumptions. Use when analyzing arguments, red-teaming reasoning, synthesizing debates, or transforming conversations into structured claim networks. Triggers include "what are the claims," "analyze this argument," "map the logic," or "find contradictions."

Packaged view

This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.

Stars
24
Hot score
88
Updated
March 20, 2026
Overall rating
C2.9
Composite score
2.9
Best-practice grade
N/A

Install command

npx @skill-hub/cli install leegonzales-aiskills-claimify
analysislogicstructuredebate

Repository

leegonzales/AISkills

Skill path: Claimify/claimify

Extract and structure claims from discourse into analyzable argument maps with logical relationships and assumptions. Use when analyzing arguments, red-teaming reasoning, synthesizing debates, or transforming conversations into structured claim networks. Triggers include "what are the claims," "analyze this argument," "map the logic," or "find contradictions."

Open repository

Best for

Primary workflow: Write Technical Docs.

Technical facets: Tech Writer.

Target audience: everyone.

License: Unknown.

Original source

Catalog source: SkillHub Club.

Repository owner: leegonzales.

This is still a mirrored public skill entry. Review the repository before installing into production workflows.

What it helps with

  • Install claimify into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
  • Review https://github.com/leegonzales/AISkills before adding claimify to shared team environments
  • Use claimify for writing workflows

Works across

Claude CodeCodex CLIGemini CLIOpenCode

Favorites: 0.

Sub-skills: 0.

Aggregator: No.

Original source / Raw SKILL.md

---
name: claimify
description: Extract and structure claims from discourse into analyzable argument maps with logical relationships and assumptions. Use when analyzing arguments, red-teaming reasoning, synthesizing debates, or transforming conversations into structured claim networks. Triggers include "what are the claims," "analyze this argument," "map the logic," or "find contradictions."
---

# Claimify

Extract claims from text and map their logical relationships into structured argument networks.

## Overview

Claimify transforms messy discourse (conversations, documents, debates, meeting notes) into analyzable claim structures that reveal:
- Explicit and implicit claims
- Logical relationships (supports/opposes/assumes/contradicts)
- Evidence chains
- Argument structure
- Tension points and gaps

## Workflow

1. **Ingest**: Read source material (conversation, document, transcript)
2. **Extract**: Identify atomic claims (one assertion per claim)
3. **Classify**: Label claim types (factual/normative/definitional/causal/predictive)
4. **Map**: Build relationship graph (which claims support/oppose/assume others)
5. **Analyze**: Identify structure, gaps, contradictions, implicit assumptions
6. **Output**: Format as requested (table/graph/narrative/JSON)

## Claim Extraction Guidelines

### Atomic Claims
Each claim should be a single, testable assertion.

**Good:**
- "AI adoption increases productivity by 15-30%"
- "Psychological safety enables team learning"
- "Current training methods fail to build AI fluency"

**Bad (not atomic):**
- "AI is useful and everyone should use it" → Split into 2 claims

### Claim Types

| Type | Definition | Example |
|------|------------|---------|
| **Factual** | Empirical statement about reality | "Remote work increased 300% since 2020" |
| **Normative** | Value judgment or prescription | "Organizations should invest in AI training" |
| **Definitional** | Establishes meaning | "AI fluency = ability to shape context and evaluate output" |
| **Causal** | X causes Y | "Lack of training causes AI underutilization" |
| **Predictive** | Future-oriented | "AI adoption will plateau without culture change" |
| **Assumption** | Unstated premise | [implicit] "Humans resist change" |

### Relationship Types

- **Supports**: Claim A provides evidence/reasoning for claim B
- **Opposes**: Claim A undermines or contradicts claim B
- **Assumes**: Claim A requires claim B to be true (often implicit)
- **Refines**: Claim A specifies/clarifies claim B
- **Contradicts**: Claims are mutually exclusive
- **Independent**: No logical relationship

## Output Formats

### Table Format (default)

```markdown
| ID | Claim | Type | Supports | Opposes | Assumes | Evidence |
|----|-------|------|----------|---------|---------|----------|
| C1 | [claim text] | Factual | - | - | C5 | [source/reasoning] |
| C2 | [claim text] | Normative | C1 | C4 | - | [source/reasoning] |
```

### Graph Format

Use Mermaid for visualization:

```mermaid
graph TD
    C1[Claim 1: AI increases productivity]
    C2[Claim 2: Training is insufficient]
    C3[Claim 3: Organizations should invest]
    
    C1 -->|supports| C3
    C2 -->|supports| C3
    C2 -.->|assumes| C4[Implicit: Change requires structure]
```

### Narrative Format

Write as structured prose with clear transitions showing logical flow:

```markdown
## Core Argument

The author argues that [main claim]. This rests on three supporting claims:

1. [Factual claim] - This is supported by [evidence]
2. [Causal claim] - However, this assumes [implicit assumption]
3. [Normative claim] - This follows if we accept [prior claims]

## Tensions

The argument contains internal tensions:
- Claims C2 and C5 appear contradictory because...
- The causal chain from C3→C7 has a missing premise...
```

### JSON Format

For programmatic processing:

```json
{
  "claims": [
    {
      "id": "C1",
      "text": "AI adoption increases productivity",
      "type": "factual",
      "explicit": true,
      "supports": ["C3"],
      "opposed_by": [],
      "assumes": ["C4"],
      "evidence": "Multiple case studies cited"
    }
  ],
  "relationships": [
    {"from": "C1", "to": "C3", "type": "supports", "strength": "strong"}
  ],
  "meta_analysis": {
    "completeness": "Missing link between C2 and C5",
    "contradictions": ["C4 vs C7"],
    "key_assumptions": ["C4", "C9"]
  }
}
```

## Analysis Depth Levels

**Level 1: Surface**
- Extract only explicit claims
- Basic support/oppose relationships
- No implicit assumption mining

**Level 2: Standard** (default)
- Extract explicit claims
- Identify clear logical relationships
- Surface obvious implicit assumptions
- Flag apparent contradictions

**Level 3: Deep**
- Extract all claims (explicit + implicit)
- Map full logical structure
- Identify hidden assumptions
- Analyze argument completeness
- Red-team reasoning
- Suggest strengthening moves

## Best Practices

1. **Be charitable**: Steelman arguments before critique
2. **Distinguish**: Separate what's claimed from what's implied
3. **Be atomic**: One claim per line, no compound assertions
4. **Track evidence**: Note source/support for each claim
5. **Flag uncertainty**: Mark inferential leaps
6. **Mind the gaps**: Identify missing premises explicitly
7. **Stay neutral**: Describe structure before evaluating strength

## Common Patterns

### Argument Chains
```
Premise 1 (factual) → Premise 2 (causal) → Conclusion (normative)
```

### Implicit Assumptions
Often found by asking: "What must be true for this conclusion to follow?"

### Contradictions
Watch for:
- Same speaker, different times
- Different speakers, same topic
- Explicit vs implicit claims

### Weak Links
- Unsupported factual claims
- Causal claims without mechanism
- Normative leaps (is → ought)
- Definitional ambiguity

## Examples

See `references/examples.md` for detailed worked examples.


---

## Referenced Files

> The following files are referenced in this skill and included for context.

### references/examples.md

```markdown
# Claimify Examples

## Example 1: Simple Argument Analysis

**Input text:**
"Remote work saves companies money because it reduces real estate costs. Therefore, all companies should adopt remote-first policies."

**Claimify output (Table format):**

| ID | Claim | Type | Supports | Opposes | Assumes | Evidence |
|----|-------|------|----------|---------|---------|----------|
| C1 | Remote work saves companies money | Factual | C3 | - | C2 | Stated premise |
| C2 | [Implicit] Cost savings are the primary decision factor | Assumption | C3 | - | - | Required for conclusion |
| C3 | All companies should adopt remote-first policies | Normative | - | - | C1, C2 | Conclusion |
| C4 | Remote work reduces real estate costs | Factual | C1 | - | - | Stated premise |

**Analysis:**
- Missing premise: C2 is never stated but required for the argument
- Weak link: "All companies" is absolute—no consideration of exceptions
- Implicit assumption: Real estate is significant cost for all companies

---

## Example 2: Debate Transcript

**Input:** Two speakers debating AI safety

**Speaker A:** "We need AI regulation now because current systems lack safety guardrails. Without regulation, we risk catastrophic outcomes."

**Speaker B:** "Premature regulation would stifle innovation. We should wait until the technology matures before imposing constraints."

**Claimify output (Graph format):**

```mermaid
graph TD
    A1[Current AI systems lack safety guardrails]
    A2[Lack of guardrails creates catastrophic risk]
    A3[We need AI regulation now]
    A4[Regulation prevents catastrophic outcomes]
    
    B1[Premature regulation stifles innovation]
    B2[Technology needs time to mature]
    B3[We should wait before regulating]
    B4[Innovation benefit > regulation benefit today]
    
    A1 -->|supports| A2
    A2 -->|supports| A3
    A4 -->|supports| A3
    
    B1 -->|supports| B3
    B2 -->|supports| B3
    B4 -.->|implicit assumption| B3
    
    A3 -.->|opposes| B3
    A2 -.->|contradicts| B4
    
    style A3 fill:#f9a,stroke:#333
    style B3 fill:#9af,stroke:#333
```

**Meta-analysis:**
- **Core disagreement**: Timing—regulate now vs. later
- **Implicit assumption (A)**: Regulation won't stifle beneficial innovation
- **Implicit assumption (B)**: Current risks are acceptable/manageable
- **Missing claims**: Neither addresses enforcement mechanisms or what "mature" means

---

## Example 3: Strategic Document

**Input:** Excerpt from AI transformation strategy doc

"Organizations adopting AI report 20-30% productivity gains, but most implementations fail due to inadequate training. AI Flight School addresses this by building fluency through practice-based learning rather than passive instruction. This approach creates sustainable capability because it develops mental models, not just surface skills."

**Claimify output (Narrative format):**

## Core Argument

The document argues that **practice-based AI training creates sustainable organizational capability** (C7 - normative/predictive). This rests on a chain of supporting claims:

### Factual Foundation
1. **C1**: Organizations adopting AI report 20-30% productivity gains (Factual)
   - Establishes baseline opportunity
   - Evidence: "report 20-30%" (likely citing studies)

2. **C2**: Most AI implementations fail (Factual)
   - Tension with C1—both can't fully capture reality
   - Needs definition of "fail"

3. **C3**: Implementation failures stem from inadequate training (Causal)
   - Critical link but unsupported
   - Assumes training is primary failure mode

### Proposed Solution
4. **C4**: AI Flight School builds fluency through practice-based learning (Definitional/Factual)
   - Describes approach

5. **C5**: Practice-based learning differs from passive instruction (Definitional)
   - Distinction claimed but not explained

### Mechanism Claims
6. **C6**: Practice-based learning develops mental models, not just surface skills (Causal)
   - Assumes mental models = better outcomes
   - Missing: what mental models? How measured?

7. **C7**: This approach creates sustainable capability (Predictive)
   - Main conclusion
   - Assumes: C3 (training is key) + C6 (mental models work) + [implicit] sustainability = mental models

## Implicit Assumptions

- **I1**: Mental models lead to better long-term performance than surface skills
- **I2**: Practice-based learning reliably builds mental models
- **I3**: The specific practice-based approach of AI Flight School is effective
- **I4**: Training is the primary bottleneck (vs. culture, tooling, incentives)

## Structural Analysis

**Strengths:**
- Clear causal chain from problem → solution → outcome
- Empirical grounding (C1 productivity stats)

**Weaknesses:**
- C2 and C1 create tension—are implementations succeeding or failing?
- C3 (training causes failure) is unsupported—correlation ≠ causation
- Mechanism (practice → mental models → sustainability) needs evidence
- Missing: comparison to alternatives, success metrics, failure modes

**Missing links:**
- What mental models specifically?
- How do we know practice builds them?
- Why is this superior to other training approaches?
- How do we measure sustainability?

---

## Example 4: Meeting Notes → Decision Map

**Input:** Meeting transcript about Q4 planning

**Output:**

| ID | Claim | Type | Speaker | Decision Impact |
|----|-------|------|---------|-----------------|
| C1 | Q4 revenue will miss target by 15% | Predictive | CFO | High - drives next 3 decisions |
| C2 | Sales team lacks AI tools | Factual | VP Sales | Medium |
| C3 | AI tools would close gap | Causal | VP Sales | High - contested by C5 |
| C4 | We should invest in AI tooling | Normative | VP Sales | **DECISION** |
| C5 | Problem is pipeline, not tools | Causal | CEO | High - opposes C3 |
| C6 | [Implicit] We can only afford one initiative | Assumption | - | Critical - unstated constraint |

**Decision structure:**
```
C1 (problem) → {C2+C3} vs {C5} → C4 (proposed action)
                  ↓                    ↓
              Sales view          CEO view
```

**Resolution needed:** C3 vs C5 contradiction—is it tools or pipeline?

---

## Example 5: Research Synthesis

**Input:** Synthesizing 3 papers on AI adoption

**Output (JSON format):**

```json
{
  "papers": ["Paper A", "Paper B", "Paper C"],
  "claims": [
    {
      "id": "C1",
      "text": "AI adoption correlates with productivity gains",
      "type": "factual",
      "sources": ["Paper A", "Paper C"],
      "agreement": "strong",
      "evidence_quality": "medium - correlation studies"
    },
    {
      "id": "C2", 
      "text": "Productivity gains require cultural change",
      "type": "causal",
      "sources": ["Paper B"],
      "agreement": "single source",
      "evidence_quality": "weak - case study only"
    },
    {
      "id": "C3",
      "text": "Technical training alone is insufficient",
      "type": "factual",
      "sources": ["Paper A", "Paper B"],
      "agreement": "moderate",
      "contradicts": ["C4"]
    },
    {
      "id": "C4",
      "text": "Technical skills predict AI usage",
      "type": "causal",
      "sources": ["Paper C"],
      "agreement": "contested",
      "contradicts": ["C3"]
    }
  ],
  "synthesis": {
    "consensus": ["C1"],
    "contested": ["C3 vs C4 - mechanism unclear"],
    "gaps": [
      "No longitudinal studies of sustainability",
      "Cultural change operationalized differently across papers",
      "Missing: measurement of 'fluency' vs 'skills'"
    ],
    "research_needed": [
      "RCT comparing training approaches",
      "Longitudinal tracking of capability retention",
      "Operationalize 'cultural change' construct"
    ]
  }
}
```

---

## Example 6: Red-Teaming Your Own Argument

**Input:** "AI will transform work fundamentally because it automates cognitive tasks, unlike previous technologies that only automated physical tasks."

**Claimify analysis:**

### Your Claims
| ID | Claim | Type | Status |
|----|-------|------|--------|
| C1 | AI automates cognitive tasks | Factual | ✓ Supported |
| C2 | Previous technologies only automated physical tasks | Factual | ✗ **FALSE** |
| C3 | Cognitive automation is fundamentally different | Causal | ⚠️ Needs support |
| C4 | AI will transform work fundamentally | Predictive | ⚠️ Depends on C2, C3 |

### Red-Team Findings

**Fatal flaw:** C2 is demonstrably false
- Spreadsheets automated cognitive work (calculation)
- Databases automated cognitive work (memory/retrieval)
- Compilers automated cognitive work (translation)

**Weakened conclusion:** If C2 falls, C4 needs new support. The argument becomes:
"AI automates cognitive tasks *at a different scale/scope* than previous technologies"

**Strengthen by:**
1. Replace C2 with: "AI automates a broader range of cognitive tasks"
2. Add: "AI handles ambiguity in ways previous automation couldn't"
3. Specify: What does "fundamentally transform" mean? (Scope ambiguity)

### Revised Argument
"AI will transform work in unprecedented ways because it automates cognitive tasks that require judgment and ambiguity-handling—capabilities previous automation couldn't replicate at scale."

**New structure:**
- More specific claim (ambiguity + judgment, not just "cognitive")
- Historical comparison that holds up
- "Unprecedented" instead of "fundamental" (less absolute)

```

claimify | SkillHub