ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2101.10588
  4. Cited By
Generalization error of random features and kernel methods:
  hypercontractivity and kernel matrix concentration

Generalization error of random features and kernel methods: hypercontractivity and kernel matrix concentration

26 January 2021
Song Mei
Theodor Misiakiewicz
Andrea Montanari
ArXivPDFHTML

Papers citing "Generalization error of random features and kernel methods: hypercontractivity and kernel matrix concentration"

31 / 31 papers shown
Title
Sobolev norm inconsistency of kernel interpolation
Sobolev norm inconsistency of kernel interpolation
Yunfei Yang
34
0
0
29 Apr 2025
Beyond Benign Overfitting in Nadaraya-Watson Interpolators
Beyond Benign Overfitting in Nadaraya-Watson Interpolators
Daniel Barzilai
Guy Kornowski
Ohad Shamir
80
0
0
11 Feb 2025
Overfitting Behaviour of Gaussian Kernel Ridgeless Regression: Varying
  Bandwidth or Dimensionality
Overfitting Behaviour of Gaussian Kernel Ridgeless Regression: Varying Bandwidth or Dimensionality
Marko Medvedev
Gal Vardi
Nathan Srebro
70
3
0
05 Sep 2024
Universal randomised signatures for generative time series modelling
Universal randomised signatures for generative time series modelling
Francesca Biagini
Lukas Gonon
Niklas Walter
42
4
0
14 Jun 2024
Characterizing Overfitting in Kernel Ridgeless Regression Through the
  Eigenspectrum
Characterizing Overfitting in Kernel Ridgeless Regression Through the Eigenspectrum
Tin Sum Cheng
Aurelien Lucchi
Anastasis Kratsios
David Belius
45
8
0
02 Feb 2024
Universal Approximation Theorem and error bounds for quantum neural networks and quantum reservoirs
Universal Approximation Theorem and error bounds for quantum neural networks and quantum reservoirs
Lukas Gonon
A. Jacquier
35
13
0
24 Jul 2023
Error Bounds for Learning with Vector-Valued Random Features
Error Bounds for Learning with Vector-Valued Random Features
S. Lanthaler
Nicholas H. Nelsen
27
12
0
26 May 2023
Least Squares Regression Can Exhibit Under-Parameterized Double Descent
Least Squares Regression Can Exhibit Under-Parameterized Double Descent
Xinyue Li
Rishi Sonthalia
44
3
0
24 May 2023
Provable Guarantees for Nonlinear Feature Learning in Three-Layer Neural Networks
Provable Guarantees for Nonlinear Feature Learning in Three-Layer Neural Networks
Eshaan Nichani
Alexandru Damian
Jason D. Lee
MLT
47
13
0
11 May 2023
Online Learning for the Random Feature Model in the Student-Teacher
  Framework
Online Learning for the Random Feature Model in the Student-Teacher Framework
Roman Worschech
B. Rosenow
48
0
0
24 Mar 2023
Learning time-scales in two-layers neural networks
Learning time-scales in two-layers neural networks
Raphael Berthier
Andrea Montanari
Kangjie Zhou
41
33
0
28 Feb 2023
Bayes-optimal Learning of Deep Random Networks of Extensive-width
Bayes-optimal Learning of Deep Random Networks of Extensive-width
Hugo Cui
Florent Krzakala
Lenka Zdeborová
BDL
30
35
0
01 Feb 2023
Strong inductive biases provably prevent harmless interpolation
Strong inductive biases provably prevent harmless interpolation
Michael Aerni
Marco Milanta
Konstantin Donhauser
Fanny Yang
42
9
0
18 Jan 2023
Random Feature Models for Learning Interacting Dynamical Systems
Random Feature Models for Learning Interacting Dynamical Systems
Yuxuan Liu
S. McCalla
Hayden Schaeffer
26
12
0
11 Dec 2022
Learning Single-Index Models with Shallow Neural Networks
Learning Single-Index Models with Shallow Neural Networks
A. Bietti
Joan Bruna
Clayton Sanford
M. Song
170
68
0
27 Oct 2022
Identifying good directions to escape the NTK regime and efficiently
  learn low-degree plus sparse polynomials
Identifying good directions to escape the NTK regime and efficiently learn low-degree plus sparse polynomials
Eshaan Nichani
Yunzhi Bai
Jason D. Lee
29
10
0
08 Jun 2022
Sharp Asymptotics of Kernel Ridge Regression Beyond the Linear Regime
Sharp Asymptotics of Kernel Ridge Regression Beyond the Linear Regime
Hong Hu
Yue M. Lu
53
15
0
13 May 2022
High-dimensional Asymptotics of Feature Learning: How One Gradient Step
  Improves the Representation
High-dimensional Asymptotics of Feature Learning: How One Gradient Step Improves the Representation
Jimmy Ba
Murat A. Erdogdu
Taiji Suzuki
Zhichao Wang
Denny Wu
Greg Yang
MLT
42
121
0
03 May 2022
SRMD: Sparse Random Mode Decomposition
SRMD: Sparse Random Mode Decomposition
Nicholas Richardson
Hayden Schaeffer
Giang Tran
27
11
0
12 Apr 2022
HARFE: Hard-Ridge Random Feature Expansion
HARFE: Hard-Ridge Random Feature Expansion
Esha Saha
Hayden Schaeffer
Giang Tran
43
14
0
06 Feb 2022
Learning with convolution and pooling operations in kernel methods
Learning with convolution and pooling operations in kernel methods
Theodor Misiakiewicz
Song Mei
MLT
15
29
0
16 Nov 2021
Conditioning of Random Feature Matrices: Double Descent and
  Generalization Error
Conditioning of Random Feature Matrices: Double Descent and Generalization Error
Zhijun Chen
Hayden Schaeffer
37
12
0
21 Oct 2021
Deformed semicircle law and concentration of nonlinear random matrices
  for ultra-wide neural networks
Deformed semicircle law and concentration of nonlinear random matrices for ultra-wide neural networks
Zhichao Wang
Yizhe Zhu
37
18
0
20 Sep 2021
Reconstruction on Trees and Low-Degree Polynomials
Reconstruction on Trees and Low-Degree Polynomials
Frederic Koehler
Elchanan Mossel
35
9
0
14 Sep 2021
A Farewell to the Bias-Variance Tradeoff? An Overview of the Theory of
  Overparameterized Machine Learning
A Farewell to the Bias-Variance Tradeoff? An Overview of the Theory of Overparameterized Machine Learning
Yehuda Dar
Vidya Muthukumar
Richard G. Baraniuk
38
71
0
06 Sep 2021
Deep Networks Provably Classify Data on Curves
Deep Networks Provably Classify Data on Curves
Tingran Wang
Sam Buchanan
D. Gilboa
John N. Wright
23
9
0
29 Jul 2021
Random feature neural networks learn Black-Scholes type PDEs without
  curse of dimensionality
Random feature neural networks learn Black-Scholes type PDEs without curse of dimensionality
Lukas Gonon
26
35
0
14 Jun 2021
Learning with invariances in random features and kernel models
Learning with invariances in random features and kernel models
Song Mei
Theodor Misiakiewicz
Andrea Montanari
OOD
55
89
0
25 Feb 2021
The Interpolation Phase Transition in Neural Networks: Memorization and
  Generalization under Lazy Training
The Interpolation Phase Transition in Neural Networks: Memorization and Generalization under Lazy Training
Andrea Montanari
Yiqiao Zhong
49
95
0
25 Jul 2020
Random Features for Kernel Approximation: A Survey on Algorithms,
  Theory, and Beyond
Random Features for Kernel Approximation: A Survey on Algorithms, Theory, and Beyond
Fanghui Liu
Xiaolin Huang
Yudong Chen
Johan A. K. Suykens
BDL
44
172
0
23 Apr 2020
Sharp analysis of low-rank kernel matrix approximations
Sharp analysis of low-rank kernel matrix approximations
Francis R. Bach
86
280
0
09 Aug 2012
1