ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1903.08560
204
746
v1v2v3v4v5 (latest)

Surprises in High-Dimensional Ridgeless Least Squares Interpolation

19 March 2019
Trevor Hastie
Andrea Montanari
Saharon Rosset
Robert Tibshirani
ArXiv (abs)PDFHTML
Abstract

Interpolators -- estimators that achieve zero training error -- have attracted growing attention in machine learning, mainly because state-of-the art neural networks appear to be models of this type. In this paper, we study minimum ℓ2\ell_2ℓ2​-norm interpolation in high-dimensional linear regression. Motivated by the connection with overparametrized neural networks, we consider the case of random features. We study two distinct models for the features' distribution: a linear model in which the feature vectors xi∈Rpx_i\in{\mathbb R}^pxi​∈Rp are obtained by applying a linear transform to vectors of i.i.d. entries, xi=Σ1/2zix_i = \Sigma^{1/2}z_ixi​=Σ1/2zi​ (with zi∈Rpz_i\in{\mathbb R}^pzi​∈Rp); a nonlinear model, in which the features are obtained by passing the input through a random one-layer neural network xi=φ(Wzi)x_i = \varphi(Wz_i)xi​=φ(Wzi​) (with zi∈Rdz_i\in{\mathbb R}^dzi​∈Rd, and φ\varphiφ an activation function acting independently on the coordinates of WziWz_iWzi​). We recover -- in a precise quantitative way -- several phenomena that have been observed in large scale neural networks and kernel machines, including the `double descent' behavior of the generalization error and the potential benefit of overparametrization.

View on arXiv
Comments on this paper