23
71

Escaping from saddle points on Riemannian manifolds

Abstract

We consider minimizing a nonconvex, smooth function ff on a Riemannian manifold M\mathcal{M}. We show that a perturbed version of Riemannian gradient descent algorithm converges to a second-order stationary point (and hence is able to escape saddle points on the manifold). The rate of convergence depends as 1/ϵ21/\epsilon^2 on the accuracy ϵ\epsilon, which matches a rate known only for unconstrained smooth minimization. The convergence rate depends polylogarithmically on the manifold dimension dd, hence is almost dimension-free. The rate also has a polynomial dependence on the parameters describing the curvature of the manifold and the smoothness of the function. While the unconstrained problem (Euclidean setting) is well-studied, our result is the first to prove such a rate for nonconvex, manifold-constrained problems.

View on arXiv
Comments on this paper