83
1

Overcoming the Curse of Dimensionality in Neural Networks

Abstract

Let AA be a set, VV a real Hilbert space. Let HH be a real Hilbert space of functions f:AVf:A\to V for which there exists M>0M>0 such that for all fHf\in H, supxAf(x)VMfH\sup_{x\in A}\Vert f(x)\Vert_{V}\leq M \Vert f\Vert_H. For i=1,,ni=1,\cdots,n, let (xi,yi)A×V(x_i,y_i)\in A\times V comprise our dataset. Let 0<q<10<q<1 and fHf^*\in H be the unique global minimizer of the functional \begin{equation*} u(f) = \frac{q}{2}\Vert f\Vert_{H}^{2} + \frac{1-q}{2n}\sum_{i=1}^{n}\Vert f(x_i)-y_i\Vert_{V}^{2}. \end{equation*} In this paper we show that for each kNk\in\mathbb{N} there exists a two layer network where the first layer has kk basis functions associated with the Hilbert space HH and the second layer is a weighted sum of the first layer, such that the functions fkf_k realized by these networks satisfy \begin{equation*} \Vert f_{k}-f^*\Vert_{H}^{2} \leq \Bigl( o(1) + \frac{C}{q^2} E\bigl[ \Vert Du_{I}(f^*)\Vert_{H^{*}}^{2} \bigr] \Bigr)\frac{1}{k}. \end{equation*} Let us note that xix_i do not need to be in a linear space and yiy_i are in a possibly infinite dimensional Hilbert space VV. The error estimate is independent of the data size nn and in the case VV is finite dimensional the error estimate is also independent of the dimension of VV.

View on arXiv
Comments on this paper