13
2

Distributed Sparse Normal Means Estimation with Sublinear Communication

Abstract

We consider the problem of sparse normal means estimation in a distributed setting with communication constraints. We assume there are MM machines, each holding dd-dimensional observations of a KK-sparse vector μ\mu corrupted by additive Gaussian noise. The MM machines are connected in a star topology to a fusion center, whose goal is to estimate the vector μ\mu with a low communication budget. Previous works have shown that to achieve the centralized minimax rate for the 2\ell_2 risk, the total communication must be high - at least linear in the dimension dd. This phenomenon occurs, however, at very weak signals. We show that at signal-to-noise ratios (SNRs) that are sufficiently high - but not enough for recovery by any individual machine - the support of μ\mu can be correctly recovered with significantly less communication. Specifically, we present two algorithms for distributed estimation of a sparse mean vector corrupted by either Gaussian or sub-Gaussian noise. We then prove that above certain SNR thresholds, with high probability, these algorithms recover the correct support with total communication that is sublinear in the dimension dd. Furthermore, the communication decreases exponentially as a function of signal strength. If in addition KMdlogdKM\ll \tfrac{d}{\log d}, then with an additional round of sublinear communication, our algorithms achieve the centralized rate for the 2\ell_2 risk. Finally, we present simulations that illustrate the performance of our algorithms in different parameter regimes.

View on arXiv
Comments on this paper