169
88

Improved Approximation Properties of Dictionaries and Applications to Neural Networks

Abstract

This article addresses the problem of approximating a function in a Hilbert space by an expansion over a dictionary D\mathbb{D}. We introduce the notion of a smoothly parameterized dictionary and give upper bounds on the approximation rates, metric entropy and nn-widths of the absolute convex hull, which we denote B1(D)B_1(\mathbb{D}), of such dictionaries. The upper bounds depend upon the order of smoothness of the parameterization, and improve upon existing results in many cases. The main applications of these results is to the dictionaries D={σ(ωx+b)}L2\mathbb{D} = \{\sigma(\omega\cdot x + b)\}\subset L^2 corresponding to shallow neural networks with activation function σ\sigma, and to the dictionary of decaying Fourier modes corresponding to the spectral Barron space. This improves upon existing approximation rates for shallow neural networks when σ=ReLUk\sigma = \text{ReLU}^k for k2k\geq 2, sharpens bounds on the metric entropy, and provides the first bounds on the Gelfand nn-widths of the Barron space and spectral Barron space.

View on arXiv
Comments on this paper