ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2004.04328
25
68

A Brief Prehistory of Double Descent

7 April 2020
Marco Loog
T. Viering
A. Mey
Jesse H. Krijthe
David Tax
ArXivPDFHTML
Abstract

In their thought-provoking paper [1], Belkin et al. illustrate and discuss the shape of risk curves in the context of modern high-complexity learners. Given a fixed training sample size nnn, such curves show the risk of a learner as a function of some (approximate) measure of its complexity NNN. With NNN the number of features, these curves are also referred to as feature curves. A salient observation in [1] is that these curves can display, what they call, double descent: with increasing NNN, the risk initially decreases, attains a minimum, and then increases until NNN equals nnn, where the training data is fitted perfectly. Increasing NNN even further, the risk decreases a second and final time, creating a peak at N=nN=nN=n. This twofold descent may come as a surprise, but as opposed to what [1] reports, it has not been overlooked historically. Our letter draws attention to some original, earlier findings, of interest to contemporary machine learning.

View on arXiv
Comments on this paper