11
53

Escaping Saddle Points in Constrained Optimization

Abstract

In this paper, we study the problem of escaping from saddle points in smooth nonconvex optimization problems subject to a convex set C\mathcal{C}. We propose a generic framework that yields convergence to a second-order stationary point of the problem, if the convex set C\mathcal{C} is simple for a quadratic objective function. Specifically, our results hold if one can find a ρ\rho-approximate solution of a quadratic program subject to C\mathcal{C} in polynomial time, where ρ<1\rho<1 is a positive constant that depends on the structure of the set C\mathcal{C}. Under this condition, we show that the sequence of iterates generated by the proposed framework reaches an (ϵ,γ)(\epsilon,\gamma)-second order stationary point (SOSP) in at most O(max{ϵ2,ρ3γ3})\mathcal{O}(\max\{\epsilon^{-2},\rho^{-3}\gamma^{-3}\}) iterations. We further characterize the overall complexity of reaching an SOSP when the convex set C\mathcal{C} can be written as a set of quadratic constraints and the objective function Hessian has a specific structure over the convex set C\mathcal{C}. Finally, we extend our results to the stochastic setting and characterize the number of stochastic gradient and Hessian evaluations to reach an (ϵ,γ)(\epsilon,\gamma)-SOSP.

View on arXiv
Comments on this paper