ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2405.15047
72
6

Credal Wrapper of Model Averaging for Uncertainty Estimation in Classification

23 May 2024
Kaizheng Wang
Fabio Cuzzolin
Keivan K1 Shariatmadar
David Moens
Hans Hallez
    UQCV
    BDL
ArXivPDFHTML
Abstract

This paper presents an innovative approach, called credal wrapper, to formulating a credal set representation of model averaging for Bayesian neural networks (BNNs) and deep ensembles (DEs), capable of improving uncertainty estimation in classification tasks. Given a finite collection of single predictive distributions derived from BNNs or DEs, the proposed credal wrapper approach extracts an upper and a lower probability bound per class, acknowledging the epistemic uncertainty due to the availability of a limited amount of distributions. Such probability intervals over classes can be mapped on a convex set of probabilities (a credal set) from which, in turn, a unique prediction can be obtained using a transformation called intersection probability transformation. In this article, we conduct extensive experiments on several out-of-distribution (OOD) detection benchmarks, encompassing various dataset pairs (CIFAR10/100 vs SVHN/Tiny-ImageNet, CIFAR10 vs CIFAR10-C, CIFAR100 vs CIFAR100-C and ImageNet vs ImageNet-O) and using different network architectures (such as VGG16, ResNet-18/50, EfficientNet B2, and ViT Base). Compared to the BNN and DE baselines, the proposed credal wrapper method exhibits superior performance in uncertainty estimation and achieves a lower expected calibration error on corrupted data.

View on arXiv
@article{wang2025_2405.15047,
  title={ Credal Wrapper of Model Averaging for Uncertainty Estimation in Classification },
  author={ Kaizheng Wang and Fabio Cuzzolin and Keivan Shariatmadar and David Moens and Hans Hallez },
  journal={arXiv preprint arXiv:2405.15047},
  year={ 2025 }
}
Comments on this paper