Q1 vs Q2: Deep Dive
Mathematical foundations, calibration metrics, and practical interpretation of aleatoric and epistemic uncertainty
Quick Overview
Q1 - Aleatoric Uncertainty
- • Quantile: 25% (pessimistic)
- • Measures: Data noise and ambiguity
- • Reducibility: Irreducible
- • Threshold: 0.35
Q2 - Epistemic Uncertainty
- • Quantile: 75% (optimistic)
- • Measures: Model ignorance
- • Reducibility: Reducible with more data
- • Threshold: 0.35
Mathematical Definition
Q1 - Quantile 25%
Q1 represents the 25th percentile of the uncertainty distribution. It's a pessimistic estimate of uncertainty.
Example
Claim: "The Earth is flat"
Q1 prediction: 0.02 (very low aleatoric uncertainty)
In 100 claims similar to this with Q1=0.02:
→ ~25 have veracity < 0.02 (very false)
→ ~75 have veracity > 0.02 (somewhat true or completely true)
Q2 - Quantile 75%
Q2 represents the 75th percentile of the uncertainty distribution. It's an optimistic estimate of confidence.
Example
Claim: "Vaccines prevent diseases"
Q2 prediction: 0.92 (high epistemic confidence)
In 100 claims similar to this with Q2=0.92:
→ ~75 have veracity < 0.92
→ ~25 have veracity > 0.92 (highly accurate)
Why Q2 is Conditioned on Q1
Key Insight
Conditioning Q2 on Q1 improves calibration by 21%. When a question is very ambiguous (high Q1), the model should account for that when assessing its own knowledge (Q2).
❌ Without Conditioning
Q1 and Q2 predicted independently:
Problem: Q2 doesn't know about question ambiguity, leading to poor calibration.
✅ With Conditioning
Q2 conditioned on Q1:
Benefit: Q2 adjusts based on Q1, improving calibration by 21%.
Calibration Metrics
AletheionGuard uses multiple metrics to ensure Q1 and Q2 are well-calibrated:
ECE (Expected Calibration Error)
Measures the average gap between predicted confidence and observed accuracy across bins.
RCE (Relative Calibration Error)
Measures relative error of calibration as a percentage of observed accuracy.
Brier Score
Mean squared difference between predicted probabilities and actual outcomes.
Uncertainty Correlation
Correlation between epistemic uncertainty and actual error rate.
Target Thresholds
| Metric | Target | Level 0 | Level 1 |
|---|---|---|---|
| Q1 MSE | < 0.05 | ~0.06 | ~0.048 |
| Q2 MSE | < 0.05 | ~0.057 | ~0.045 |
| RCE | < 0.05 | ~0.06 | ~0.042 |
| ECE | < 0.10 | ~0.10-0.15 | ~0.07-0.10 |
| Uncertainty Corr. | > 0.5 | ~0.52 | ~0.61 |
Practical Interpretation
Reading Q1 Values
Reading Q2 Values
Combined Interpretation
Code Example
Performance Characteristics
Latency
- • Embedding: ~10ms
- • Q1/Q2 inference: ~5ms
- • Calibration: ~3ms
- • Total: 20-30ms per response
Throughput
- • Single: 50 req/sec
- • Batch 32: 500 req/sec
- • Batch 128: 1000+ req/sec
- • Production: ~200-400 req/sec sustained
Training Loss Function
AletheionGuard Level 1 uses Pyramidal VARO loss to train Q1 and Q2 gates:
Component Breakdown
- • λ₁: Q1 accuracy weight
- • λ₂: Q2 accuracy weight
- • λ₃: Height regression weight
- • λ₄: Calibration weight (RCE)
- • λ₅: Fractal constraint weight
Typical Values
- • λ₁: 1.0
- • λ₂: 1.2 (slightly higher)
- • λ₃: 0.8
- • λ₄: 1.5 (prioritize calibration)
- • λ₅: 0.5
Next Steps
Questions about Q1 and Q2?
Our team can help you understand uncertainty quantification for your use case