List-Decodable Subspace Recovery via Sum-of-Squares
We give the first efficient algorithm for the problem of list-decodable subspace recovery. Our algorithm takes input samples () are generated i.i.d. from Gaussian distribution on with covariance of rank and the rest are arbitrary, potentially adversarial outliers. It outputs a list of projection matrices guaranteed to contain a projection matrix such that , where hides polylogarithmic factors in . Here, is the projection matrix to the range space of . The algorithm needs samples and runs in time time where is the ratio of the largest to smallest non-zero eigenvalues of . Our algorithm builds on the recently developed framework for list-decodable learning via the sum-of-squares (SoS) method [KKK'19, RY'20] with some key technical and conceptual advancements. Our key conceptual contribution involves showing a (SoS "certified") lower bound on the eigenvalues of covariances of arbitrary small subsamples of an i.i.d. sample of a certifiably anti-concentrated distribution. One of our key technical contributions gives a new method that allows error reduction "within SoS" with only a logarithmic cost in the exponent in the running time (in contrast to polynomial cost in [KKK'19, RY'20]. In a concurrent and independent work, Raghavendra and Yau proved related results for list-decodable subspace recovery [RY'20].
View on arXiv