ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2209.01173
20
1

Optimal bump functions for shallow ReLU networks: Weight decay, depth separation and the curse of dimensionality

2 September 2022
Stephan Wojtowytsch
ArXivPDFHTML
Abstract

In this note, we study how neural networks with a single hidden layer and ReLU activation interpolate data drawn from a radially symmetric distribution with target labels 1 at the origin and 0 outside the unit ball, if no labels are known inside the unit ball. With weight decay regularization and in the infinite neuron, infinite data limit, we prove that a unique radially symmetric minimizer exists, whose weight decay regularizer and Lipschitz constant grow as ddd and d\sqrt{d}d​ respectively. We furthermore show that the weight decay regularizer grows exponentially in ddd if the label 111 is imposed on a ball of radius ε\varepsilonε rather than just at the origin. By comparison, a neural networks with two hidden layers can approximate the target function without encountering the curse of dimensionality.

View on arXiv
Comments on this paper