Claude Code Plugins

Community-maintained marketplace

Feedback

Extract and structure claims from discourse into analyzable argument maps with logical relationships and assumptions. Use when analyzing arguments, red-teaming reasoning, synthesizing debates, or transforming conversations into structured claim networks. Triggers include "what are the claims," "analyze this argument," "map the logic," or "find contradictions.

Install Skill

1Download skill
2Enable skills in Claude

Open claude.ai/settings/capabilities and find the "Skills" section

3Upload to Claude

Click "Upload skill" and select the downloaded ZIP file

Note: Please verify skill by going through its instructions before using it.

SKILL.md

name claimify
description Extract and structure claims from discourse into analyzable argument maps with logical relationships and assumptions. Use when analyzing arguments, red-teaming reasoning, synthesizing debates, or transforming conversations into structured claim networks. Triggers include "what are the claims," "analyze this argument," "map the logic," or "find contradictions."

Claimify

Extract claims from text and map their logical relationships into structured argument networks.

Overview

Claimify transforms messy discourse (conversations, documents, debates, meeting notes) into analyzable claim structures that reveal:

  • Explicit and implicit claims
  • Logical relationships (supports/opposes/assumes/contradicts)
  • Evidence chains
  • Argument structure
  • Tension points and gaps

Workflow

  1. Ingest: Read source material (conversation, document, transcript)
  2. Extract: Identify atomic claims (one assertion per claim)
  3. Classify: Label claim types (factual/normative/definitional/causal/predictive)
  4. Map: Build relationship graph (which claims support/oppose/assume others)
  5. Analyze: Identify structure, gaps, contradictions, implicit assumptions
  6. Output: Format as requested (table/graph/narrative/JSON)

Claim Extraction Guidelines

Atomic Claims

Each claim should be a single, testable assertion.

Good:

  • "AI adoption increases productivity by 15-30%"
  • "Psychological safety enables team learning"
  • "Current training methods fail to build AI fluency"

Bad (not atomic):

  • "AI is useful and everyone should use it" → Split into 2 claims

Claim Types

Type Definition Example
Factual Empirical statement about reality "Remote work increased 300% since 2020"
Normative Value judgment or prescription "Organizations should invest in AI training"
Definitional Establishes meaning "AI fluency = ability to shape context and evaluate output"
Causal X causes Y "Lack of training causes AI underutilization"
Predictive Future-oriented "AI adoption will plateau without culture change"
Assumption Unstated premise [implicit] "Humans resist change"

Relationship Types

  • Supports: Claim A provides evidence/reasoning for claim B
  • Opposes: Claim A undermines or contradicts claim B
  • Assumes: Claim A requires claim B to be true (often implicit)
  • Refines: Claim A specifies/clarifies claim B
  • Contradicts: Claims are mutually exclusive
  • Independent: No logical relationship

Output Formats

Table Format (default)

| ID | Claim | Type | Supports | Opposes | Assumes | Evidence |
|----|-------|------|----------|---------|---------|----------|
| C1 | [claim text] | Factual | - | - | C5 | [source/reasoning] |
| C2 | [claim text] | Normative | C1 | C4 | - | [source/reasoning] |

Graph Format

Use Mermaid for visualization:

graph TD
    C1[Claim 1: AI increases productivity]
    C2[Claim 2: Training is insufficient]
    C3[Claim 3: Organizations should invest]
    
    C1 -->|supports| C3
    C2 -->|supports| C3
    C2 -.->|assumes| C4[Implicit: Change requires structure]

Narrative Format

Write as structured prose with clear transitions showing logical flow:

## Core Argument

The author argues that [main claim]. This rests on three supporting claims:

1. [Factual claim] - This is supported by [evidence]
2. [Causal claim] - However, this assumes [implicit assumption]
3. [Normative claim] - This follows if we accept [prior claims]

## Tensions

The argument contains internal tensions:
- Claims C2 and C5 appear contradictory because...
- The causal chain from C3→C7 has a missing premise...

JSON Format

For programmatic processing:

{
  "claims": [
    {
      "id": "C1",
      "text": "AI adoption increases productivity",
      "type": "factual",
      "explicit": true,
      "supports": ["C3"],
      "opposed_by": [],
      "assumes": ["C4"],
      "evidence": "Multiple case studies cited"
    }
  ],
  "relationships": [
    {"from": "C1", "to": "C3", "type": "supports", "strength": "strong"}
  ],
  "meta_analysis": {
    "completeness": "Missing link between C2 and C5",
    "contradictions": ["C4 vs C7"],
    "key_assumptions": ["C4", "C9"]
  }
}

Analysis Depth Levels

Level 1: Surface

  • Extract only explicit claims
  • Basic support/oppose relationships
  • No implicit assumption mining

Level 2: Standard (default)

  • Extract explicit claims
  • Identify clear logical relationships
  • Surface obvious implicit assumptions
  • Flag apparent contradictions

Level 3: Deep

  • Extract all claims (explicit + implicit)
  • Map full logical structure
  • Identify hidden assumptions
  • Analyze argument completeness
  • Red-team reasoning
  • Suggest strengthening moves

Best Practices

  1. Be charitable: Steelman arguments before critique
  2. Distinguish: Separate what's claimed from what's implied
  3. Be atomic: One claim per line, no compound assertions
  4. Track evidence: Note source/support for each claim
  5. Flag uncertainty: Mark inferential leaps
  6. Mind the gaps: Identify missing premises explicitly
  7. Stay neutral: Describe structure before evaluating strength

Common Patterns

Argument Chains

Premise 1 (factual) → Premise 2 (causal) → Conclusion (normative)

Implicit Assumptions

Often found by asking: "What must be true for this conclusion to follow?"

Contradictions

Watch for:

  • Same speaker, different times
  • Different speakers, same topic
  • Explicit vs implicit claims

Weak Links

  • Unsupported factual claims
  • Causal claims without mechanism
  • Normative leaps (is → ought)
  • Definitional ambiguity

Examples

See references/examples.md for detailed worked examples.