Epistemic Uncertainty
Understanding the difference between what models don't know (epistemic) and what's inherently uncertain (aleatoric)
What is Epistemic Uncertainty?
Epistemic uncertainty (Q2) represents reducible model ignorance — the uncertainty that arises from insufficient training data, out-of-distribution queries, or knowledge gaps that the model could theoretically learn given more information.
Key Characteristics
- • Reducible: Can be decreased with more training data
- • Signals ignorance: Model doesn't have enough knowledge
- • Detects hallucination: High Q2 = high hallucination risk
- • Out-of-distribution: Identifies queries outside training domain
Q1 (Aleatoric) vs Q2 (Epistemic)
Q1 - Aleatoric Uncertainty
Irreducible data noise and inherent ambiguity in the question itself.
(Amsterdam vs The Hague - both valid answers)
Q2 - Epistemic Uncertainty
Reducible uncertainty from model ignorance and knowledge gaps.
(Model lacks knowledge - fictional place)
| Aspect | Aleatoric (Q1) | Epistemic (Q2) |
|---|---|---|
| Source | Data ambiguity | Model ignorance |
| Reducibility | ❌ Irreducible | ✅ Reducible |
| When High | Question is ambiguous | Model lacks knowledge |
| Verdict | "MAYBE" | "REFUSED" |
| Action | Ask for clarification | Retrieve more context |
How Q1 and Q2 Are Measured
Neural Network Gates
AletheionGuard uses specialized neural networks to predict Q1 and Q2 from sentence embeddings.
Why Q2 is Conditioned on Q1
Conditioning Q2 on Q1 improves calibration by 21%. If a question is very ambiguous (high Q1), the model should account for that when assessing its own knowledge (Q2).
Derived Metrics
Height (Proximity to Truth)
- • Range: [0, 1]
- • 0 = Base (completely uncertain)
- • 1 = Apex (perfect confidence)
Total Uncertainty
- • Range: [0, ~1.41]
- • Combines both sources of uncertainty
- • Used in verdict decision logic
Verdict Decision Logic
AletheionGuard uses Q1, Q2, and total uncertainty to make verdicts:
ACCEPT
Low Q1 and Q2. Model is confident and likely correct.
MAYBE
High Q1 (aleatoric). Question is ambiguous, needs clarification.
REFUSED
High Q2 (epistemic). Model lacks knowledge, high hallucination risk.
Why Epistemic Uncertainty Matters
1. Enables Safe Automation
Without calibration, you cannot safely automate decisions. With epistemic uncertainty, you can:
2. Detects Hallucinations
High Q2 signals hallucination risk. AletheionGuard achieves ROC-AUC 0.94 for hallucination detection.
3. Enables RAG Optimization
Trigger additional retrieval when epistemic uncertainty is high:
4. Compliance & Auditability
Critical for regulated industries (medical, legal, financial). Know when AI cannot make a decision with confidence.
Real-World Example
Healthcare Q&A System
Low Epistemic Uncertainty
Q: "What is the normal heart rate?"
✅ Safe to answer automatically
High Epistemic Uncertainty
Q: "Should I take this new experimental drug?"
⚠️ Escalate to medical professional
Impact
- • Reduced hallucination rate from 18% to 4%
- • Improved user trust scores by 37%
- • 60% of queries auto-approved safely
The Pyramidal Architecture
AletheionGuard uses a geometric pyramid model to represent uncertainty:
Apex (Height = 1)
Perfect knowledge. Q1 = 0, Q2 = 0. Model is certain and correct.
Middle (0.3 < Height < 0.7)
Moderate uncertainty. Some Q1 or Q2. Requires human review.
Base (Height = 0)
Maximum uncertainty. High Q1 and/or Q2. Cannot make reliable prediction.
Next Steps
Want to Learn More?
Explore our research paper and technical documentation