237
294

Optimal approximation of continuous functions by very deep ReLU networks

Abstract

We prove that deep ReLU neural networks with conventional fully-connected architectures with WW weights can approximate continuous ν\nu-variate functions ff with uniform error not exceeding aνωf(cνW2/ν),a_\nu\omega_f(c_\nu W^{-2/\nu}), where ωf\omega_f is the modulus of continuity of ff and aν,cνa_\nu, c_\nu are some ν\nu-dependent constants. This bound is tight. Our construction is inherently deep and nonlinear: the obtained approximation rate cannot be achieved by networks with fewer than Ω(W/lnW)\Omega(W/\ln W) layers or by networks with weights continuously depending on ff.

View on arXiv
Comments on this paper