ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.05938
45
0

Uncertainty Quantification From Scaling Laws in Deep Neural Networks

7 March 2025
Ibrahim Elsharkawy
Yonatan Kahn
Benjamin Hooberman
    UQCV
ArXivPDFHTML
Abstract

Quantifying the uncertainty from machine learning analyses is critical to their use in the physical sciences. In this work we focus on uncertainty inherited from the initialization distribution of neural networks. We compute the mean μL\mu_{\mathcal{L}}μL​ and variance σL2\sigma_{\mathcal{L}}^2σL2​ of the test loss L\mathcal{L}L for an ensemble of multi-layer perceptrons (MLPs) with neural tangent kernel (NTK) initialization in the infinite-width limit, and compare empirically to the results from finite-width networks for three example tasks: MNIST classification, CIFAR classification and calorimeter energy regression. We observe scaling laws as a function of training set size NDN_\mathcal{D}ND​ for both μL\mu_{\mathcal{L}}μL​ and σL\sigma_{\mathcal{L}}σL​, but find that the coefficient of variation ϵL≡σL/μL\epsilon_{\mathcal{L}} \equiv \sigma_{\mathcal{L}}/\mu_{\mathcal{L}}ϵL​≡σL​/μL​ becomes independent of NDN_\mathcal{D}ND​ at both infinite and finite width for sufficiently large NDN_\mathcal{D}ND​. This implies that the coefficient of variation of a finite-width network may be approximated by its infinite-width value, and may in principle be calculable using finite-width perturbation theory.

View on arXiv
@article{elsharkawy2025_2503.05938,
  title={ Uncertainty Quantification From Scaling Laws in Deep Neural Networks },
  author={ Ibrahim Elsharkawy and Yonatan Kahn and Benjamin Hooberman },
  journal={arXiv preprint arXiv:2503.05938},
  year={ 2025 }
}
Comments on this paper