ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1712.07424
13
16

ADINE: An Adaptive Momentum Method for Stochastic Gradient Descent

20 December 2017
Vishwak Srinivasan
Adepu Ravi Sankar
V. Balasubramanian
    ODL
ArXivPDFHTML
Abstract

Two major momentum-based techniques that have achieved tremendous success in optimization are Polyak's heavy ball method and Nesterov's accelerated gradient. A crucial step in all momentum-based methods is the choice of the momentum parameter mmm which is always suggested to be set to less than 111. Although the choice of m<1m < 1m<1 is justified only under very strong theoretical assumptions, it works well in practice even when the assumptions do not necessarily hold. In this paper, we propose a new momentum based method ADINE\textit{ADINE}ADINE, which relaxes the constraint of m<1m < 1m<1 and allows the learning algorithm to use adaptive higher momentum. We motivate our hypothesis on mmm by experimentally verifying that a higher momentum (≥1\ge 1≥1) can help escape saddles much faster. Using this motivation, we propose our method ADINE\textit{ADINE}ADINE that helps weigh the previous updates more (by setting the momentum parameter >1> 1>1), evaluate our proposed algorithm on deep neural networks and show that ADINE\textit{ADINE}ADINE helps the learning algorithm to converge much faster without compromising on the generalization error.

View on arXiv
Comments on this paper