ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2202.09889
13
10

Memorize to Generalize: on the Necessity of Interpolation in High Dimensional Linear Regression

20 February 2022
Chen Cheng
John C. Duchi
Rohith Kuditipudi
ArXivPDFHTML
Abstract

We examine the necessity of interpolation in overparameterized models, that is, when achieving optimal predictive risk in machine learning problems requires (nearly) interpolating the training data. In particular, we consider simple overparameterized linear regression y=Xθ+wy = X \theta + wy=Xθ+w with random design X∈Rn×dX \in \mathbb{R}^{n \times d}X∈Rn×d under the proportional asymptotics d/n→γ∈(1,∞)d/n \to \gamma \in (1, \infty)d/n→γ∈(1,∞). We precisely characterize how prediction (test) error necessarily scales with training error in this setting. An implication of this characterization is that as the label noise variance σ2→0\sigma^2 \to 0σ2→0, any estimator that incurs at least cσ4\mathsf{c}\sigma^4cσ4 training error for some constant c\mathsf{c}c is necessarily suboptimal and will suffer growth in excess prediction error at least linear in the training error. Thus, optimal performance requires fitting training data to substantially higher accuracy than the inherent noise floor of the problem.

View on arXiv
Comments on this paper