ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1908.05355
49
626

The generalization error of random features regression: Precise asymptotics and double descent curve

14 August 2019
Song Mei
Andrea Montanari
ArXivPDFHTML
Abstract

Deep learning methods operate in regimes that defy the traditional statistical mindset. Neural network architectures often contain more parameters than training samples, and are so rich that they can interpolate the observed labels, even if the latter are replaced by pure noise. Despite their huge complexity, the same architectures achieve small generalization error on real data. This phenomenon has been rationalized in terms of a so-called `double descent' curve. As the model complexity increases, the test error follows the usual U-shaped curve at the beginning, first decreasing and then peaking around the interpolation threshold (when the model achieves vanishing training error). However, it descends again as model complexity exceeds this threshold. The global minimum of the test error is found above the interpolation threshold, often in the extreme overparametrization regime in which the number of parameters is much larger than the number of samples. Far from being a peculiar property of deep neural networks, elements of this behavior have been demonstrated in much simpler settings, including linear regression with random covariates. In this paper we consider the problem of learning an unknown function over the ddd-dimensional sphere Sd−1\mathbb S^{d-1}Sd−1, from nnn i.i.d. samples (xi,yi)∈Sd−1×R(\boldsymbol x_i, y_i)\in \mathbb S^{d-1} \times \mathbb R(xi​,yi​)∈Sd−1×R, i≤ni\le ni≤n. We perform ridge regression on NNN random features of the form σ(waTx)\sigma(\boldsymbol w_a^{\mathsf T} \boldsymbol x)σ(waT​x), a≤Na\le Na≤N. This can be equivalently described as a two-layers neural network with random first-layer weights. We compute the precise asymptotics of the test error, in the limit N,n,d→∞N,n,d\to \inftyN,n,d→∞ with N/dN/dN/d and n/dn/dn/d fixed. This provides the first analytically tractable model that captures all the features of the double descent phenomenon without assuming ad hoc misspecification structures.

View on arXiv
Comments on this paper