ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2007.02486
14
51

Regularization Matters: A Nonparametric Perspective on Overparametrized Neural Network

6 July 2020
Tianyang Hu
Wei Cao
Cong Lin
Guang Cheng
ArXivPDFHTML
Abstract

Overparametrized neural networks trained by gradient descent (GD) can provably overfit any training data. However, the generalization guarantee may not hold for noisy data. From a nonparametric perspective, this paper studies how well overparametrized neural networks can recover the true target function in the presence of random noises. We establish a lower bound on the L2L_2L2​ estimation error with respect to the GD iterations, which is away from zero without a delicate scheme of early stopping. In turn, through a comprehensive analysis of ℓ2\ell_2ℓ2​-regularized GD trajectories, we prove that for overparametrized one-hidden-layer ReLU neural network with the ℓ2\ell_2ℓ2​ regularization: (1) the output is close to that of the kernel ridge regression with the corresponding neural tangent kernel; (2) minimax {optimal} rate of L2L_2L2​ estimation error can be achieved. Numerical experiments confirm our theory and further demonstrate that the ℓ2\ell_2ℓ2​ regularization approach improves the training robustness and works for a wider range of neural networks.

View on arXiv
Comments on this paper