Minimax Optimal Quantization of Linear Models: Information-Theoretic
Limits and Efficient Algorithms
- MQ
Main:15 Pages
8 Figures
Bibliography:1 Pages
12 Tables
Appendix:34 Pages
Abstract
We consider the problem of quantizing a linear model learned from measurements . The model is constrained to be representable using only -bits, where is a pre-specified budget and is the dimension of the model. We derive an information-theoretic lower bound for the minimax risk under this setting and show that it is tight with a matching upper bound. This upper bound is achieved using randomized embedding based algorithms. We propose randomized Hadamard embeddings that are computationally efficient while performing near-optimally. We also show that our method and upper-bounds can be extended for two-layer ReLU neural networks. Numerical simulations validate our theoretical claims.
View on arXivComments on this paper
