ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2209.11395
9
18

Achieve the Minimum Width of Neural Networks for Universal Approximation

23 September 2022
Yongqiang Cai
ArXivPDFHTML
Abstract

The universal approximation property (UAP) of neural networks is fundamental for deep learning, and it is well known that wide neural networks are universal approximators of continuous functions within both the LpL^pLp norm and the continuous/uniform norm. However, the exact minimum width, wmin⁡w_{\min}wmin​, for the UAP has not been studied thoroughly. Recently, using a decoder-memorizer-encoder scheme, \citet{Park2021Minimum} found that wmin⁡=max⁡(dx+1,dy)w_{\min} = \max(d_x+1,d_y)wmin​=max(dx​+1,dy​) for both the LpL^pLp-UAP of ReLU networks and the CCC-UAP of ReLU+STEP networks, where dx,dyd_x,d_ydx​,dy​ are the input and output dimensions, respectively. In this paper, we consider neural networks with an arbitrary set of activation functions. We prove that both CCC-UAP and LpL^pLp-UAP for functions on compact domains share a universal lower bound of the minimal width; that is, wmin⁡∗=max⁡(dx,dy)w^*_{\min} = \max(d_x,d_y)wmin∗​=max(dx​,dy​). In particular, the critical width, wmin⁡∗w^*_{\min}wmin∗​, for LpL^pLp-UAP can be achieved by leaky-ReLU networks, provided that the input or output dimension is larger than one. Our construction is based on the approximation power of neural ordinary differential equations and the ability to approximate flow maps by neural networks. The nonmonotone or discontinuous activation functions case and the one-dimensional case are also discussed.

View on arXiv
Comments on this paper