Back to Explorer
Research PaperResearchia:202602.26014[Data Science > Machine Learning]

Not Just How Much, But Where: Decomposing Epistemic Uncertainty into Per-Class Contributions

Mame Diarra Toure

Abstract

In safety-critical classification, the cost of failure is often asymmetric, yet Bayesian deep learning summarises epistemic uncertainty with a single scalar, mutual information (MI), that cannot distinguish whether a model's ignorance involves a benign or safety-critical class. We decompose MI into a per-class vector Ck(x)=σk2/(2μk)C_k(x)=σ_k^{2}/(2μ_k), with μk=E[pk]μ_k{=}\mathbb{E}[p_k] and σk2=Var[pk]σ_k^2{=}\mathrm{Var}[p_k] across posterior samples. The decomposition follows from a second-order Taylor expansion of the entropy; the 1/μk1/μ_k weighting corrects boundary suppression and makes CkC_k comparable across rare and common classes. By construction kCkMI\sum_k C_k \approx \mathrm{MI}, and a companion skewness diagnostic flags inputs where the approximation degrades. After characterising the axiomatic properties of CkC_k, we validate it on three tasks: (i) selective prediction for diabetic retinopathy, where critical-class CkC_k reduces selective risk by 34.7% over MI and 56.2% over variance baselines; (ii) out-of-distribution detection on clinical and image benchmarks, where kCk\sum_k C_k achieves the highest AUROC and the per-class view exposes asymmetric shifts invisible to MI; and (iii) a controlled label-noise study in which kCk\sum_k C_k shows less sensitivity to injected aleatoric noise than MI under end-to-end Bayesian training, while both metrics degrade under transfer learning. Across all tasks, the quality of the posterior approximation shapes uncertainty at least as strongly as the choice of metric, suggesting that how uncertainty is propagated through the network matters as much as how it is measured.


Source: arXiv:2602.21160v1 - http://arxiv.org/abs/2602.21160v1 PDF: https://arxiv.org/pdf/2602.21160v1 Original Link: http://arxiv.org/abs/2602.21160v1

Submission:2/26/2026
Comments:0 comments
Subjects:Machine Learning; Data Science
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Not Just How Much, But Where: Decomposing Epistemic Uncertainty into Per-Class Contributions | Researchia | Researchia