255

Minimax Optimal Quantization of Linear Models: Information-Theoretic Limits and Efficient Algorithms

Main:15 Pages
8 Figures
Bibliography:1 Pages
12 Tables
Appendix:34 Pages
Abstract

We consider the problem of quantizing a linear model learned from measurements X=Wθ+v\mathbf{X} = \mathbf{W}\boldsymbol{\theta} + \mathbf{v}. The model is constrained to be representable using only dBdB-bits, where B(0,)B \in (0, \infty) is a pre-specified budget and dd is the dimension of the model. We derive an information-theoretic lower bound for the minimax risk under this setting and show that it is tight with a matching upper bound. This upper bound is achieved using randomized embedding based algorithms. We propose randomized Hadamard embeddings that are computationally efficient while performing near-optimally. We also show that our method and upper-bounds can be extended for two-layer ReLU neural networks. Numerical simulations validate our theoretical claims.

View on arXiv
Comments on this paper