Uncertainty Quantification with Proper Scoring Rules: Adjusting Measures to Prediction Tasks

We address the problem of uncertainty quantification and propose measures of total, aleatoric, and epistemic uncertainty based on a known decomposition of (strictly) proper scoring rules, a specific type of loss function, into a divergence and an entropy component. This leads to a flexible framework for uncertainty quantification that can be instantiated with different losses (scoring rules), which makes it possible to tailor uncertainty quantification to the use case at hand. We show that this flexibility is indeed advantageous. In particular, we analyze the task of selective prediction and show that the scoring rule should ideally match the task loss. In addition, we perform experiments on two other common tasks. For out-of-distribution detection, our results confirm that a widely used measure of epistemic uncertainty, mutual information, performs best. Moreover, in the setting of active learning, our measure of epistemic uncertainty based on the zero-one-loss consistently outperforms other uncertainty measures.
View on arXiv@article{hofman2025_2505.22538, title={ Uncertainty Quantification with Proper Scoring Rules: Adjusting Measures to Prediction Tasks }, author={ Paul Hofman and Yusuf Sale and Eyke Hüllermeier }, journal={arXiv preprint arXiv:2505.22538}, year={ 2025 } }