75
1
v1v2v3v4v5v6 (latest)

Overcoming the Curse of Dimensionality in Neural Networks

Abstract

Let AA be a set and VV a real Hilbert space. Let HH be a real Hilbert space of functions f:AVf:A\to V and assume HH is continuously embedded in the Banach space of bounded functions. For i=1,,ni=1,\cdots,n, let (xi,yi)A×V(x_i,y_i)\in A\times V comprise our dataset. Let 0<q<10<q<1 and fHf^*\in H be the unique global minimizer of the functional \begin{equation*} u(f) = \frac{q}{2}\Vert f\Vert_{H}^{2} + \frac{1-q}{2n}\sum_{i=1}^{n}\Vert f(x_i)-y_i\Vert_{V}^{2}. \end{equation*} In this paper we show that for each kNk\in\mathbb{N} there exists a two layer network where the first layer has kk functions which are Riesz representations in the Hilbert space HH of point evaluation functionals and the second layer is a weighted sum of the first layer, such that the functions fkf_k realized by these networks satisfy \begin{equation*} \Vert f_{k}-f^*\Vert_{H}^{2} \leq \Bigl( o(1) + \frac{C}{q^2} E\bigl[ \Vert Du_{I}(f^*)\Vert_{H^{*}}^{2} \bigr] \Bigr)\frac{1}{k}. \end{equation*} %Let us note that xix_i do not need to be in a linear space and yiy_i are in a possibly infinite dimensional Hilbert space VV. %The error estimate is independent of the data size nn and in the case VV is finite dimensional %the error estimate is also independent of the dimension of VV. By choosing the Hilbert space HH appropriately, the computational complexity of evaluating the Riesz representations of point evaluations might be small and thus the network has low computational complexity.

View on arXiv
Comments on this paper