On overcoming the Curse of Dimensionality in Neural Networks
Let be a set, a Hilbert space. Let be a Hilbert space of functions such that we have . For , let comprise our dataset. Let be the unique global minimizer of the functional \begin{equation*} u(f) = \frac{\lambda}{2}\Vert f\Vert_{H}^{2} +\frac{1}{2} \frac{1}{n}\sum_{i=1}^{n}\Vert f(x_i)-y_i\Vert_{V}^{2}. \end{equation*} In this paper we show that for each there exists a two layer network where the first layer has number of basis functions for and the second layer takes a weighted summation of the first layer, such that the functions realized by these networks satisfy \begin{equation*} E\left[ \Vert F_{k}-f^*\Vert_{H}^{2} \right] \leq \bigl(o(1) + \frac{C}{\lambda^2} (\frac{1}{\lambda}+M^2)u(f^*) \bigr) \frac{1}{k} . \end{equation*} Thus the do not need to be in a linear space and are in a possibly infinite dimensional Hilbert space. The error rate is independent of the dimension of and data size .
View on arXiv