Approximating Continuous Functions by ReLU Nets of Minimal Width

This article concerns the expressive power of depth in deep feed-forward neural nets with ReLU activations. Specifically, we answer the following question: for a fixed what is the minimal width so that neural nets with ReLU activations, input dimension , hidden layer widths at most and arbitrary depth can approximate any continuous, real-valued function of variables arbitrarily well? It turns out that this minimal width is exactly equal to That is, if all the hidden layer widths are bounded by , then even in the infinite depth limit, ReLU nets can only express a very limited class of functions, and, on the other hand, any continuous function on the -dimensional unit cube can be approximated to arbitrary precision by ReLU nets in which all hidden layers have width exactly Our construction in fact shows that any continuous function can be approximated by a net of width . We obtain quantitative depth estimates for such an approximation in terms of the modulus of continuity of .
View on arXiv