73
173

Dropping Convexity for Faster Semi-definite Optimization

Srinadh Bhojanapalli
Anastasios Kyrillidis
Sujay Sanghavi
Abstract

In this paper, we study the minimization of a convex function f(X)f(X) over the space of n×nn\times n positive semidefinite matrices X0X\succeq 0, but when the problem is recast as the non-convex problem minUg(U)\min_U g(U) where g(U):=f(UU)g(U) := f(UU^\top), with UU being an n×rn\times r matrix and rnr\leq n. We study the performance of gradient descent on gg -- which we refer to as Factored Gradient Descent (FGD) -- under standard assumptions on the original function ff. We provide a rule for selecting the step size, and with this choice show that the local convergence rate of FGD mirrors that of standard gradient descent on the original ff -- the error after kk steps is O(1/k)O(1/k) for smooth ff, and exponentially small in kk when ff is (restricted) strongly convex. Note that gg is not locally convex. In addition, we provide a procedure to initialize FGD for (restricted) strongly convex objectives and when one only has access to ff via a first-order oracle. FGD and similar procedures are widely used in practice for problems that can be posed as matrix factorization; to the best of our knowledge, ours is the first paper to provide precise convergence rate guarantees for general convex functions under standard convex assumptions.

View on arXiv
Comments on this paper