17
8

Approximation of functions with one-bit neural networks

Abstract

The celebrated universal approximation theorems for neural networks roughly state that any reasonable function can be arbitrarily well-approximated by a network whose parameters are appropriately chosen real numbers. This paper examines the approximation capabilities of one-bit neural networks -- those whose nonzero parameters are ±a\pm a for some fixed a0a\not=0. One of our main theorems shows that for any fCs([0,1]d)f\in C^s([0,1]^d) with f<1\|f\|_\infty<1 and error ε\varepsilon, there is a fNNf_{NN} such that f(x)fNN(x)ε|f(\boldsymbol{x})-f_{NN}(\boldsymbol{x})|\leq \varepsilon for all x\boldsymbol{x} away from the boundary of [0,1]d[0,1]^d, and fNNf_{NN} is either implementable by a {±1}\{\pm 1\} quadratic network with O(ε2d/s)O(\varepsilon^{-2d/s}) parameters or a {±12}\{\pm \frac 1 2 \} ReLU network with O(ε2d/slog(1/ε))O(\varepsilon^{-2d/s}\log (1/\varepsilon)) parameters, as ε0\varepsilon\to0. We establish new approximation results for iterated multivariate Bernstein operators, error estimates for noise-shaping quantization on the Bernstein basis, and novel implementation of the Bernstein polynomials by one-bit quadratic and ReLU neural networks.

View on arXiv
Comments on this paper