ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2109.11395
9
0

Generalisations and improvements of New Q-Newton's method Backtracking

23 September 2021
T. Truong
ArXivPDFHTML
Abstract

In this paper, we propose a general framework for the algorithm New Q-Newton's method Backtracking, developed in the author's previous work. For a symmetric, square real matrix AAA, we define minsp(A):=min⁡∣∣e∣∣=1∣∣Ae∣∣minsp(A):=\min _{||e||=1} ||Ae||minsp(A):=min∣∣e∣∣=1​∣∣Ae∣∣. Given a C2C^2C2 cost function f:Rm→Rf:\mathbb{R}^m\rightarrow \mathbb{R}f:Rm→R and a real number 0<τ0<\tau 0<τ, as well as m+1m+1m+1 fixed real numbers δ0,…,δm\delta _0,\ldots ,\delta _mδ0​,…,δm​, we define for each x∈Rmx\in \mathbb{R}^mx∈Rm with ∇f(x)≠0\nabla f(x)\not= 0∇f(x)=0 the following quantities: κ:=min⁡i≠j∣δi−δj∣\kappa :=\min _{i\not= j}|\delta _i-\delta _j|κ:=mini=j​∣δi​−δj​∣; A(x):=∇2f(x)+δ∣∣∇f(x)∣∣τIdA(x):=\nabla ^2f(x)+\delta ||\nabla f(x)||^{\tau}IdA(x):=∇2f(x)+δ∣∣∇f(x)∣∣τId, where δ\deltaδ is the first element in the sequence {δ0,…,δm}\{\delta _0,\ldots ,\delta _m\}{δ0​,…,δm​} for which minsp(A(x))≥κ∣∣∇f(x)∣∣τminsp(A(x))\geq \kappa ||\nabla f(x)||^{\tau}minsp(A(x))≥κ∣∣∇f(x)∣∣τ; e1(x),…,em(x)e_1(x),\ldots ,e_m(x)e1​(x),…,em​(x) are an orthonormal basis of Rm\mathbb{R}^mRm, chosen appropriately; w(x)=w(x)=w(x)= the step direction, given by the formula: w(x)=\sum _{i=1}^m\frac{<\nabla f(x),e_i(x)>}{||A(x)e_i(x)||}e_i(x); (we can also normalise by w(x)/max⁡{1,∣∣w(x)∣∣}w(x)/\max \{1,||w(x)||\}w(x)/max{1,∣∣w(x)∣∣} when needed) γ(x)>0\gamma (x)>0γ(x)>0 learning rate chosen by Backtracking line search so that Armijo's condition is satisfied: f(x-\gamma (x)w(x))-f(x)\leq -\frac{1}{3}\gamma (x)<\nabla f(x),w(x)>. The update rule for our algorithm is x↦H(x)=x−γ(x)w(x)x\mapsto H(x)=x-\gamma (x)w(x)x↦H(x)=x−γ(x)w(x). In New Q-Newton's method Backtracking, the choices are τ=1+α>1\tau =1+\alpha >1τ=1+α>1 and e1(x),…,em(x)e_1(x),\ldots ,e_m(x)e1​(x),…,em​(x)'s are eigenvectors of ∇2f(x)\nabla ^2f(x)∇2f(x). In this paper, we allow more flexibility and generality, for example τ\tauτ can be chosen to be <1<1<1 or e1(x),…,em(x)e_1(x),\ldots ,e_m(x)e1​(x),…,em​(x)'s are not necessarily eigenvectors of ∇2f(x)\nabla ^2f(x)∇2f(x). New Q-Newton's method Backtracking (as well as Backtracking gradient descent) is a special case, and some versions have flavours of quasi-Newton's methods. Several versions allow good theoretical guarantees. An application to solving systems of polynomial equations is given.

View on arXiv
Comments on this paper