Back to skills
SkillHub ClubShip Full StackFull Stack

confidence-honesty

Force honest confidence assessment before claiming conclusions. Triggers on 'root cause identified', 'problem identified', 'complete clarity'. Express confidence as percentage, explain what's stopping 100%, validate assumptions before presenting.

Packaged view

This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.

Stars
0
Hot score
74
Updated
March 20, 2026
Overall rating
C2.5
Composite score
2.5
Best-practice grade
C61.2

Install command

npx @skill-hub/cli install audunstrand-status-app-confidence-honesty

Repository

audunstrand/status-app

Skill path: .github/skills/confidence-honesty

Force honest confidence assessment before claiming conclusions. Triggers on 'root cause identified', 'problem identified', 'complete clarity'. Express confidence as percentage, explain what's stopping 100%, validate assumptions before presenting.

Open repository

Best for

Primary workflow: Ship Full Stack.

Technical facets: Full Stack.

Target audience: everyone.

License: Unknown.

Original source

Catalog source: SkillHub Club.

Repository owner: audunstrand.

This is still a mirrored public skill entry. Review the repository before installing into production workflows.

What it helps with

  • Install confidence-honesty into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
  • Review https://github.com/audunstrand/status-app before adding confidence-honesty to shared team environments
  • Use confidence-honesty for development workflows

Works across

Claude CodeCodex CLIGemini CLIOpenCode

Favorites: 0.

Sub-skills: 0.

Aggregator: No.

Original source / Raw SKILL.md

---
name: Confidence Honesty
description: Force honest confidence assessment before claiming conclusions. Triggers on 'root cause identified', 'problem identified', 'complete clarity'. Express confidence as percentage, explain what's stopping 100%, validate assumptions before presenting.
---

# GitHub Copilot Skill: confidence-honesty

> **Note:** This skill has been adapted from [claude-skillz](https://github.com/NTCoding/claude-skillz) 
> for use with GitHub Copilot Agent Skills.

---

# Confidence Honesty

## The Problem

Claude builds detailed, well-structured analyses that *look* thoroughโ€”then presents them with phrases like "complete clarity" or "root cause identified." Users reasonably trust this confidence. They act on it, communicate it to stakeholders, make decisions.

Then new evidence appears and invalidates the entire hypothesis.

**The harm:**
- User trusted a conclusion that was actually ~40% confident
- Time wasted on wrong direction
- Stakeholders were misinformed
- Trust in Claude's analysis erodes

**Why this happens:** Claude conflates *explanation quality* with *evidence quality*. A thorough, well-reasoned analysis feels like certaintyโ€”but reasoning without verified evidence is just speculation with extra steps.

## The Solution

Force explicit confidence assessment before claiming conclusions:
1. Express confidence as a percentage (not vague certainty)
2. Show the math: what evidence adds confidence, what gaps subtract it
3. Mandatory "Why not 100%?" for anything below 95%
4. Self-validate: if you can gather more evidence yourself, do it before presenting

## Critical Rules

๐Ÿšจ **EXPRESS CONFIDENCE AS A PERCENTAGE.** Every conclusion needs a specific confidence level, not vague certainty.

๐Ÿšจ **EXPLAIN WHAT'S STOPPING 100%.** For any confidence below 95%, you MUST explain the gaps. Non-negotiable.

๐Ÿšจ **VALIDATE BEFORE PRESENTING.** If you can gather more evidence yourself, DO IT. Don't return to user with unvalidated hypotheses.

## When This Triggers

Auto-invoke when you're about to claim:
- "root cause is", "the problem is", "root cause identified"
- "complete clarity", "definitely", "certainly", "clearly the issue"
- Any conclusive claim during investigation

## Confidence Levels

| Range | Icon | Meaning |
|-------|------|---------|
| 0-30% | ๐Ÿ”ด | Speculation - needs significant validation |
| 31-60% | ๐ŸŸก | Plausible - evidence exists but gaps remain |
| 61-85% | ๐ŸŸ  | Likely - strong evidence, minor gaps |
| 86-94% | ๐ŸŸข | High confidence - validated, minor uncertainty |
| 95-100% | ๐Ÿ’ฏ | Confirmed - fully validated |

**Calibration:**
- **20%**: One possibility among several
- **40%**: Evidence points this direction but key assumptions unverified
- **60%**: Evidence supports this, alternatives not ruled out
- **80%**: Strong evidence, assumptions verified, alternatives less likely
- **95%**: Validated with direct evidence, alternatives ruled out
- **100%**: Mathematical/logical certainty only

## Pre-Conclusion Checkpoint

**Before claiming ANY conclusion, complete this:**

### 1. Evidence Inventory
- What hard evidence supports this?
- Direct evidence (code/logs that prove it)?
- Circumstantial evidence (patterns consistent)?
- What's the strongest piece of evidence?

### 2. Falsifiability Check
- What would INVALIDATE this theory?
- What data would prove me wrong?
- Have I looked for that data?
- If no: WHY NOT?

### 3. Assumption Audit
- What am I assuming WITHOUT verification?
- List each assumption explicitly
- Mark: [VERIFIED] or [ASSUMED]

### 4. Alternative Possibilities
- What else could explain these symptoms?
- List at least 2 alternatives
- Why is my conclusion more likely?

### 5. Validation Opportunities
- Can I fetch/check the actual data?
- Can I search the codebase for confirming/denying evidence?
- Should I ask user for confirming data?

## Confidence Scoring

**Start at 50% (neutral) and adjust:**

| Factor | Adjustment |
|--------|------------|
| Direct evidence (code/logs proving it) | +15-25% |
| Verified assumptions (checked, not assumed) | +10-15% |
| Alternatives ruled out with evidence | +10-15% |
| Falsifiability check completed | +5-10% |
| Unverified critical assumptions | -15-25% |
| Plausible alternatives not ruled out | -10-15% |
| No falsifiability check | -10% |
| Circumstantial evidence only | -10% |

**Example calculation:**
```
Starting: 50%
+ Code analysis shows the pattern: +15%
+ Error logs match signature: +10%
- CRITICAL: Assumed config unchanged (NOT VERIFIED): -25%
- Alternative not ruled out: -10%

Final: 40% ๐ŸŸก
```

## Response Format

```
[Icon] [X%] Confidence: [One sentence finding]

Evidence:
- [+X%] [Evidence 1]
- [+X%] [Evidence 2]

Why not 100%:
- [-X%] [What's stopping full confidence]

To increase confidence: [What would raise % - doing it yourself or need user input]
```

## Examples

### ๐Ÿ”ด 25% - Low confidence
```
๐Ÿ”ด 25% Confidence: Database connection pool may be exhausted

Evidence:
- [+15%] Error pattern matches connection timeout
- [+10%] Load spike occurred around failure time

Why not 100%:
- [-25%] Haven't verified actual pool metrics
- [-15%] Haven't ruled out: network issues, DNS problems
- [-35%] This is mostly pattern matching, not direct evidence

To increase confidence: I need the connection pool metrics from the time of failure.
```

### ๐ŸŸ  75% - Moderate-high confidence
```
๐ŸŸ  75% Confidence: API request failing due to missing required field

Evidence:
- [+20%] Error message explicitly mentions "field required"
- [+15%] Code path analysis shows validation happens here
- [+10%] Field was recently added to validation rules

Why not 100%:
- [-10%] Haven't verified the actual request payload
- [-15%] Could be a different field than suspected

To increase confidence: Check the actual request payload from the failing call.
```

### ๐Ÿ’ฏ 98% - Near certain
```
๐Ÿ’ฏ 98% Confidence: Root cause is misconfigured environment variable

Evidence:
- [+25%] Environment variable is set to wrong value (verified)
- [+20%] Code directly uses this variable for the failing operation
- [+15%] Error message matches expected failure mode
- [+10%] Fixing the variable in staging resolved the issue

Why not 100%: Haven't deployed fix to production yet to confirm.
```

## Anti-Patterns

### โŒ Premature Certainty

**What happens:**
```
Claude: "I have complete clarity on the root cause"
Claude: [200 lines of detailed report]
User: [provides data that contradicts analysis]
Claude: "This changes everything..."
```

**What should happen:**
```
๐ŸŸก 40% Confidence: The issue appears to be X

Evidence:
- [+15%] Code path analysis suggests this pattern

Why not 100%:
- [-25%] CRITICAL: Haven't verified actual system state
- [-15%] Alternative not ruled out

To increase confidence: Before I finalize, can you provide [specific data]?
```

### โŒ Confidence in Explanation Quality

Building a detailed report โ‰  having valid evidence.

Thoroughness of presentation has zero correlation with correctness.

**Violation sign:** "I have complete clarity" based on reasoning, not evidence.

### โŒ Skipping Falsifiability

If you can't answer "what would prove me wrong?", you don't understand your own theory.

## Self-Validation Rule

**Don't return to user with questions you can answer yourself.**

Before presenting, ask:
```
Can I gather more evidence myself?
โ”œโ”€ Search codebase for confirming/denying data?
โ”œโ”€ Fetch a file that validates an assumption?
โ”œโ”€ Spawn an agent to investigate further?
โ””โ”€ Check actual state vs assumed state?

If YES โ†’ DO IT. Then reassess confidence.
If NO โ†’ Present with honest confidence + what you need from user.
```

**Critical:** If confidence is below 80% and you CAN gather more evidence โ†’ DO IT.

## Summary

๐Ÿšจ **Confidence is a percentage, not a feeling.**

๐Ÿšจ **Below 95%? Explain what's stopping 100%.**

๐Ÿšจ **Can validate yourself? Do it before presenting.**

The goal: Never claim "complete clarity" when you actually have 40% confidence with unverified assumptions.
confidence-honesty | SkillHub