AutoSGD: Automatic Learning Rate Selection for Stochastic Gradient Descent

Main:9 Pages
18 Figures
Bibliography:3 Pages
2 Tables
Appendix:30 Pages
Abstract
The learning rate is an important tuning parameter for stochastic gradient descent (SGD) and can greatly influence its performance. However, appropriate selection of a learning rate schedule across all iterations typically requires a non-trivial amount of user tuning effort. To address this, we introduce AutoSGD: an SGD method that automatically determines whether to increase or decrease the learning rate at a given iteration and then takes appropriate action. We introduce theory supporting the convergence of AutoSGD, along with its deterministic counterpart for standard gradient descent. Empirical results suggest strong performance of the method on a variety of traditional optimization problems and machine learning tasks.
View on arXiv@article{surjanovic2025_2505.21651, title={ AutoSGD: Automatic Learning Rate Selection for Stochastic Gradient Descent }, author={ Nikola Surjanovic and Alexandre Bouchard-Côté and Trevor Campbell }, journal={arXiv preprint arXiv:2505.21651}, year={ 2025 } }
Comments on this paper