32
13

On the Gap Between Strict-Saddles and True Convexity: An Omega(log d) Lower Bound for Eigenvector Approximation

Max Simchowitz
A. Alaoui
Benjamin Recht
Abstract

We prove a \emph{query complexity} lower bound on rank-one principal component analysis (PCA). We consider an oracle model where, given a symmetric matrix MRd×dM \in \mathbb{R}^{d \times d}, an algorithm is allowed to make TT \emph{exact} queries of the form w(i)=Mv(i)w^{(i)} = Mv^{(i)} for i{1,,T}i \in \{1,\dots,T\}, where v(i)v^{(i)} is drawn from a distribution which depends arbitrarily on the past queries and measurements {v(j),w(j)}1ji1\{v^{(j)},w^{(j)}\}_{1 \le j \le i-1}. We show that for a small constant ϵ\epsilon, any adaptive, randomized algorithm which can find a unit vector v^\widehat{v} for which v^Mv^(1ϵ)M\widehat{v}^{\top}M\widehat{v} \ge (1-\epsilon)\|M\|, with even small probability, must make T=Ω(logd)T = \Omega(\log d) queries. In addition to settling a widely-held folk conjecture, this bound demonstrates a fundamental gap between convex optimization and "strict-saddle" non-convex optimization of which PCA is a canonical example: in the former, first-order methods can have dimension-free iteration complexity, whereas in PCA, the iteration complexity of gradient-based methods must necessarily grow with the dimension. Our argument proceeds via a reduction to estimating the rank-one spike in a deformed Wigner model. We establish lower bounds for this model by developing a "truncated" analogue of the χ2\chi^2 Bayes-risk lower bound of Chen et al.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.