We consider the problem of learning mixtures of Gaussians under the constraint of approximate differential privacy. We prove that samples are sufficient to learn a mixture of axis-aligned Gaussians in to within total variation distance while satisfying -differential privacy. This is the first result for privately learning mixtures of unbounded axis-aligned (or even unbounded univariate) Gaussians. If the covariance matrices of each of the Gaussians is the identity matrix, we show that samples are sufficient. Recently, the "local covering" technique of Bun, Kamath, Steinke, and Wu has been successfully used for privately learning high-dimensional Gaussians with a known covariance matrix and extended to privately learning general high-dimensional Gaussians by Aden-Ali, Ashtiani, and Kamath. Given these positive results, this approach has been proposed as a promising direction for privately learning mixtures of Gaussians. Unfortunately, we show that this is not possible. We design a new technique for privately learning mixture distributions. A class of distributions is said to be list-decodable if there is an algorithm that, given "heavily corrupted" samples from , outputs a list of distributions, , such that one of the distributions in approximates . We show that if is privately list-decodable, then we can privately learn mixtures of distributions in . Finally, we show axis-aligned Gaussian distributions are privately list-decodable, thereby proving mixtures of such distributions are privately learnable.
View on arXiv