ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2411.17567
73
0

Improving the Convergence Rates of Forward Gradient Descent with Repeated Sampling

26 November 2024
Niklas Dexheimer
Johannes Schmidt-Hieber
    FedML
ArXivPDFHTML
Abstract

Forward gradient descent (FGD) has been proposed as a biologically more plausible alternative of gradient descent as it can be computed without backward pass. Considering the linear model with ddd parameters, previous work has found that the prediction error of FGD is, however, by a factor ddd slower than the prediction error of stochastic gradient descent (SGD). In this paper we show that by computing ℓ\ellℓ FGD steps based on each training sample, this suboptimality factor becomes d/(ℓ∧d)d/(\ell \wedge d)d/(ℓ∧d) and thus the suboptimality of the rate disappears if ℓ≳d.\ell \gtrsim d.ℓ≳d. We also show that FGD with repeated sampling can adapt to low-dimensional structure in the input distribution. The main mathematical challenge lies in controlling the dependencies arising from the repeated sampling process.

View on arXiv
Comments on this paper