ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2209.03282
36
2
v1v2 (latest)

Quadratic Gradient: Uniting Gradient Algorithm and Newton Method as One

3 September 2022
Jonathan Z. Chiang
ArXiv (abs)PDFHTML
Abstract

It might be inadequate for the line search technique for Newton's method to use only one floating point number. A column vector of the same size as the gradient might be better than a mere float number to accelerate each of the gradient elements with different rates. Moreover, a square matrix of the same order as the Hessian matrix might be helpful to correct the Hessian matrix. Chiang applied something between a column vector and a square matrix, namely a diagonal matrix, to accelerate the gradient and further proposed a faster gradient variant called quadratic gradient. In this paper, we present a new way to build a new version of the quadratic gradient. This new quadratic gradient doesn't satisfy the convergence conditions of the fixed Hessian Newton's method. However, experimental results show that it sometimes has a better performance than the original one in convergence rate. Also, Chiang speculates that there might be a relation between the Hessian matrix and the learning rate for the first-order gradient descent method. We prove that the floating number 1ϵ+max⁡{∣λi∣}\frac{1}{\epsilon + \max \{| \lambda_i | \}}ϵ+max{∣λi​∣}1​ can be a good learning rate of the gradient methods, where ϵ\epsilonϵ is a number to avoid division by zero and λi\lambda_iλi​ the eigenvalues of the Hessian matrix.

View on arXiv
Comments on this paper