15
80

Natasha: Faster Non-Convex Stochastic Optimization Via Strongly Non-Convex Parameter

Abstract

Given a nonconvex function that is an average of nn smooth functions, we design stochastic first-order methods to find its approximate stationary points. The convergence of our new methods depends on the smallest (negative) eigenvalue σ-\sigma of the Hessian, a parameter that describes how nonconvex the function is. Our methods outperform known results for a range of parameter σ\sigma, and can be used to find approximate local minima. Our result implies an interesting dichotomy: there exists a threshold σ0\sigma_0 so that the currently fastest methods for σ>σ0\sigma>\sigma_0 and for σ<σ0\sigma<\sigma_0 have different behaviors: the former scales with n2/3n^{2/3} and the latter scales with n3/4n^{3/4}.

View on arXiv
Comments on this paper