Katyusha: The First Truly Accelerated Stochastic Gradient Method
- ODL

We introduce , the first direct stochastic gradient method that has an accelerated convergence rate. Given an objective that is an average of convex and smooth functions, converges to an -approximate minimizer using stochastic iterations, where is the condition number. also resolves the following open questions in optimization and machine learning For weakly convex and smooth objectives (e.g., Lasso, Logistic Regression), is the first stochastic method that achieves the optimal rate. For strongly-convex but non-smooth ERM objectives (e.g., SVM), gives the first stochastic method that achieves the optimal rate. For weakly convex and non-smooth ERM objectives (e.g., L1SVM), gives the first stochastic method that achieves the optimal rate.
View on arXiv