Optimal approximation of continuous functions by very deep ReLU networks

Abstract
We prove that deep ReLU neural networks with conventional fully-connected architectures with weights can approximate continuous -variate functions with uniform error not exceeding where is the modulus of continuity of and are some -dependent constants. This bound is tight. Our construction is inherently deep and nonlinear: the obtained approximation rate cannot be achieved by networks with fewer than layers or by networks with weights continuously depending on .
View on arXivComments on this paper