Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2106.02100
Cited By
Double Descent Optimization Pattern and Aliasing: Caveats of Noisy Labels
3 June 2021
Florian Dubost
Erin Hong
Max Pike
Siddharth Sharma
Siyi Tang
Nandita Bhaskhar
Christopher Lee-Messer
D. Rubin
NoLa
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Double Descent Optimization Pattern and Aliasing: Caveats of Noisy Labels"
18 / 18 papers shown
Title
Fine-Tuning Pretrained Language Models: Weight Initializations, Data Orders, and Early Stopping
Jesse Dodge
Gabriel Ilharco
Roy Schwartz
Ali Farhadi
Hannaneh Hajishirzi
Noah A. Smith
93
595
0
15 Feb 2020
Deep Double Descent: Where Bigger Models and More Data Hurt
Preetum Nakkiran
Gal Kaplun
Yamini Bansal
Tristan Yang
Boaz Barak
Ilya Sutskever
119
935
0
04 Dec 2019
How does Early Stopping Help Generalization against Label Noise?
Hwanjun Song
Minseok Kim
Dongmin Park
Jae-Gil Lee
NoLa
62
75
0
19 Nov 2019
The generalization error of random features regression: Precise asymptotics and double descent curve
Song Mei
Andrea Montanari
80
634
0
14 Aug 2019
Benign Overfitting in Linear Regression
Peter L. Bartlett
Philip M. Long
Gábor Lugosi
Alexander Tsigler
MLT
67
777
0
26 Jun 2019
Understanding overfitting peaks in generalization error: Analytical risk curves for
l
2
l_2
l
2
and
l
1
l_1
l
1
penalized interpolation
P. Mitra
48
50
0
09 Jun 2019
A New Look at an Old Problem: A Universal Learning Approach to Linear Regression
Koby Bibas
Yaniv Fogel
M. Feder
36
34
0
12 May 2019
Harmless interpolation of noisy data in regression
Vidya Muthukumar
Kailas Vodrahalli
Vignesh Subramanian
A. Sahai
68
204
0
21 Mar 2019
Surprises in High-Dimensional Ridgeless Least Squares Interpolation
Trevor Hastie
Andrea Montanari
Saharon Rosset
Robert Tibshirani
159
743
0
19 Mar 2019
Two models of double descent for weak features
M. Belkin
Daniel J. Hsu
Ji Xu
85
376
0
18 Mar 2019
Reconciling modern machine learning practice and the bias-variance trade-off
M. Belkin
Daniel J. Hsu
Siyuan Ma
Soumik Mandal
201
1,638
0
28 Dec 2018
A jamming transition from under- to over-parametrization affects loss landscape and generalization
S. Spigler
Mario Geiger
Stéphane dÁscoli
Levent Sagun
Giulio Biroli
Matthieu Wyart
58
153
0
22 Oct 2018
The jamming transition as a paradigm to understand the loss landscape of deep neural networks
Mario Geiger
S. Spigler
Stéphane dÁscoli
Levent Sagun
Marco Baity-Jesi
Giulio Biroli
Matthieu Wyart
51
143
0
25 Sep 2018
High-dimensional dynamics of generalization error in neural networks
Madhu S. Advani
Andrew M. Saxe
AI4CE
128
469
0
10 Oct 2017
Deep Residual Learning for Image Recognition
Kaiming He
Xinming Zhang
Shaoqing Ren
Jian Sun
MedIm
1.9K
193,426
0
10 Dec 2015
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
Sergey Ioffe
Christian Szegedy
OOD
419
43,234
0
11 Feb 2015
Adam: A Method for Stochastic Optimization
Diederik P. Kingma
Jimmy Ba
ODL
1.4K
149,842
0
22 Dec 2014
ADADELTA: An Adaptive Learning Rate Method
Matthew D. Zeiler
ODL
132
6,623
0
22 Dec 2012
1