51
11

Better and Simpler Lower Bounds for Differentially Private Statistical Estimation

Abstract

We provide improved lower bounds for two well-known high-dimensional private estimation tasks. First, we prove that for estimating the covariance of a Gaussian up to spectral error α\alpha with approximate differential privacy, one needs Ω~(d3/2αε+dα2)\tilde{\Omega}\left(\frac{d^{3/2}}{\alpha \varepsilon} + \frac{d}{\alpha^2}\right) samples for any αO(1)\alpha \le O(1), which is tight up to logarithmic factors. This improves over previous work which established this for αO(1d)\alpha \le O\left(\frac{1}{\sqrt{d}}\right), and is also simpler than previous work. Next, we prove that for estimating the mean of a heavy-tailed distribution with bounded kkth moments with approximate differential privacy, one needs Ω~(dαk/(k1)ε+dα2)\tilde{\Omega}\left(\frac{d}{\alpha^{k/(k-1)} \varepsilon} + \frac{d}{\alpha^2}\right) samples. This matches known upper bounds and improves over the best known lower bound for this problem, which only hold for pure differential privacy, or when k=2k = 2. Our techniques follow the method of fingerprinting and are generally quite simple. Our lower bound for heavy-tailed estimation is based on a black-box reduction from privately estimating identity-covariance Gaussians. Our lower bound for covariance estimation utilizes a Bayesian approach to show that, under an Inverse Wishart prior distribution for the covariance matrix, no private estimator can be accurate even in expectation, without sufficiently many samples.

View on arXiv
Comments on this paper