44
v1v2 (latest)

Decomposing Probabilistic Scores: Reliability, Information Loss and Uncertainty

Arthur Charpentier
Agathe Fernandes Machado
Main:8 Pages
6 Figures
Bibliography:2 Pages
4 Tables
Appendix:8 Pages
Abstract

Calibration is a conditional property that depends on the information retained by a predictor. We develop decomposition identities for arbitrary proper losses that make this dependence explicit. At any information level A\mathcal A, the expected loss of an A\mathcal A-measurable predictor splits into a proper-regret (reliability) term and a conditional entropy (residual uncertainty) term. For nested levels AB\mathcal A\subseteq\mathcal B, a chain decomposition quantifies the information gain from A\mathcal A to B\mathcal B. Applied to classification with features X\boldsymbol{X} and score S=s(X)S=s(\boldsymbol{X}), this yields a three-term identity: miscalibration, a {\em grouping} term measuring information loss from X\boldsymbol{X} to SS, and irreducible uncertainty at the feature level. We leverage the framework to analyze post-hoc recalibration, aggregation of calibrated models, and stagewise/boosting constructions, with explicit forms for Brier and log-loss.

View on arXiv
Comments on this paper