Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2110.06910
Cited By
On the Double Descent of Random Features Models Trained with SGD
13 October 2021
Fanghui Liu
Johan A. K. Suykens
Volkan Cevher
MLT
Re-assign community
ArXiv
PDF
HTML
Papers citing
"On the Double Descent of Random Features Models Trained with SGD"
11 / 11 papers shown
Title
Conditioning of Random Feature Matrices: Double Descent and Generalization Error
Zhijun Chen
Hayden Schaeffer
64
12
0
21 Oct 2021
Memorizing without overfitting: Bias, variance, and interpolation in over-parameterized models
J. Rocks
Pankaj Mehta
60
41
0
26 Oct 2020
Multiple Descent: Design Your Own Generalization Curve
Lin Chen
Yifei Min
M. Belkin
Amin Karbasi
DRL
46
61
0
03 Aug 2020
On the Optimal Weighted
ℓ
2
\ell_2
ℓ
2
Regularization in Overparameterized Linear Regression
Denny Wu
Ji Xu
47
122
0
10 Jun 2020
A Random Matrix Analysis of Random Fourier Features: Beyond the Gaussian Kernel, a Precise Phase Transition, and the Corresponding Double Descent
Zhenyu Liao
Romain Couillet
Michael W. Mahoney
43
88
0
09 Jun 2020
Rethinking Bias-Variance Trade-off for Generalization of Neural Networks
Zitong Yang
Yaodong Yu
Chong You
Jacob Steinhardt
Yi-An Ma
38
182
0
26 Feb 2020
Generalisation error in learning with random features and the hidden manifold model
Federica Gerace
Bruno Loureiro
Florent Krzakala
M. Mézard
Lenka Zdeborová
48
168
0
21 Feb 2020
Surprises in High-Dimensional Ridgeless Least Squares Interpolation
Trevor Hastie
Andrea Montanari
Saharon Rosset
Robert Tibshirani
94
737
0
19 Mar 2019
Reconciling modern machine learning practice and the bias-variance trade-off
M. Belkin
Daniel J. Hsu
Siyuan Ma
Soumik Mandal
150
1,628
0
28 Dec 2018
Fast and Faster Convergence of SGD for Over-Parameterized Models and an Accelerated Perceptron
Sharan Vaswani
Francis R. Bach
Mark Schmidt
42
297
0
16 Oct 2018
Learning with SGD and Random Features
Luigi Carratino
Alessandro Rudi
Lorenzo Rosasco
28
77
0
17 Jul 2018
1