15
16

Polynomial-Time Sum-of-Squares Can Robustly Estimate Mean and Covariance of Gaussians Optimally

Abstract

In this work, we revisit the problem of estimating the mean and covariance of an unknown dd-dimensional Gaussian distribution in the presence of an ε\varepsilon-fraction of adversarial outliers. The pioneering work of [DKK+16] gave a polynomial time algorithm for this task with optimal O~(ε)\tilde{O}(\varepsilon) error using n=poly(d,1/ε)n = \textrm{poly}(d, 1/\varepsilon) samples. On the other hand, [KS17b] introduced a general framework for robust moment estimation via a canonical sum-of-squares relaxation that succeeds for the more general class of certifiably subgaussian and certifiably hypercontractive [BK20] distributions. When specialized to Gaussians, this algorithm obtains the same O~(ε)\tilde{O}(\varepsilon) error guarantee as [DKK+16] but incurs a super-polynomial sample complexity (n=dO(log(1/ε)n = d^{O(\log(1/\varepsilon)}) and running time (nO(log(1/ε))n^{O(\log(1/\varepsilon))}). This cost appears inherent to their analysis as it relies only on sum-of-squares certificates of upper bounds on directional moments while the analysis in [DKK+16] relies on lower bounds on directional moments inferred from algebraic relationships between moments of Gaussian distributions. We give a new, simple analysis of the same canonical sum-of-squares relaxation used in [KS17b, BK20] and show that for Gaussian distributions, their algorithm achieves the same error, sample complexity and running time guarantees as of the specialized algorithm in [DKK+16]. Our key innovation is a new argument that allows using moment lower bounds without having sum-of-squares certificates for them. We believe that our proof technique will likely be useful in developing further robust estimation algorithms.

View on arXiv
Comments on this paper