ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2005.10815
21
48

Can Shallow Neural Networks Beat the Curse of Dimensionality? A mean field training perspective

21 May 2020
Stephan Wojtowytsch
E. Weinan
    MLT
ArXivPDFHTML
Abstract

We prove that the gradient descent training of a two-layer neural network on empirical or population risk may not decrease population risk at an order faster than t−4/(d−2)t^{-4/(d-2)}t−4/(d−2) under mean field scaling. Thus gradient descent training for fitting reasonably smooth, but truly high-dimensional data may be subject to the curse of dimensionality. We present numerical evidence that gradient descent training with general Lipschitz target functions becomes slower and slower as the dimension increases, but converges at approximately the same rate in all dimensions when the target function lies in the natural function space for two-layer ReLU networks.

View on arXiv
Comments on this paper