98
20

On the Powerball Method

Claire J. Tomlin
Abstract

We propose a new method to accelerate the convergence of optimization algorithms. This method adds a power coefficient γ(0,1)\gamma\in(0,1) to the gradient during optimization. We call this the Powerball method after the well-known Heavy-ball method \cite{heavyball}. We prove that the Powerball method can achieve ϵ\epsilon accuracy for strongly convex functions by using O((1γ)1ϵγ1)O\left((1-\gamma)^{-1}\epsilon^{\gamma-1}\right) iterations. We also demonstrate that the Powerball method provides a 1010-fold speed up of the convergence of both gradient descent and L-BFGS on multiple real datasets.

View on arXiv
Comments on this paper