ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.16020
48
0

AlphaGrad: Non-Linear Gradient Normalization Optimizer

22 April 2025
Soham Sane
    ODL
ArXivPDFHTML
Abstract

We introduce AlphaGrad, a memory-efficient, conditionally stateless optimizer addressing the memory overhead and hyperparameter complexity of adaptive methods like Adam. AlphaGrad enforces scale invariance via tensor-wise L2 gradient normalization followed by a smooth hyperbolic tangent transformation, g′=tanh⁡(α⋅g~)g' = \tanh(\alpha \cdot \tilde{g})g′=tanh(α⋅g~​), controlled by a single steepness parameter α\alphaα. Our contributions include: (1) the AlphaGrad algorithm formulation; (2) a formal non-convex convergence analysis guaranteeing stationarity; (3) extensive empirical evaluation on diverse RL benchmarks (DQN, TD3, PPO). Compared to Adam, AlphaGrad demonstrates a highly context-dependent performance profile. While exhibiting instability in off-policy DQN, it provides enhanced training stability with competitive results in TD3 (requiring careful α\alphaα tuning) and achieves substantially superior performance in on-policy PPO. These results underscore the critical importance of empirical α\alphaα selection, revealing strong interactions between the optimizer's dynamics and the underlying RL algorithm. AlphaGrad presents a compelling alternative optimizer for memory-constrained scenarios and shows significant promise for on-policy learning regimes where its stability and efficiency advantages can be particularly impactful.

View on arXiv
@article{sane2025_2504.16020,
  title={ AlphaGrad: Non-Linear Gradient Normalization Optimizer },
  author={ Soham Sane },
  journal={arXiv preprint arXiv:2504.16020},
  year={ 2025 }
}
Comments on this paper