Communication Complexity of Estimating Correlations

We characterize the communication complexity of the following distributed estimation problem. Alice and Bob observe infinitely many iid copies of -correlated unit-variance (Gaussian or binary) random variables, with unknown . By interactively exchanging bits, Bob wants to produce an estimate of . We show that the best possible performance (optimized over interaction protocol and estimator ) satisfies . Furthermore, we show that the best possible unbiased estimator achieves performance of . Curiously, thus, restricting communication to bits results in (order-wise) similar minimax estimation error as restricting to samples. Our results also imply an lower bound on the information complexity of the Gap-Hamming problem, for which we show a direct information-theoretic proof. Notably, the protocol achieving (almost) optimal performance is one-way (non-interactive). For one-way protocols we also prove the bound even when is restricted to any small open sub-interval of (i.e. a local minimax lower bound). %We do not know if this local behavior remains true in the interactive setting. Our proof techniques rely on symmetric strong data-processing inequalities, various tensorization techniques from information-theoretic interactive common-randomness extraction, and (for the local lower bound) on the Otto-Villani estimate for the Wasserstein-continuity of trajectories of the Ornstein-Uhlenbeck semigroup.
View on arXiv