ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2305.16569
16
7

Accelerating Value Iteration with Anchoring

26 May 2023
Jongmin Lee
Ernest K. Ryu
ArXivPDFHTML
Abstract

Value Iteration (VI) is foundational to the theory and practice of modern reinforcement learning, and it is known to converge at a O(γk)\mathcal{O}(\gamma^k)O(γk)-rate, where γ\gammaγ is the discount factor. Surprisingly, however, the optimal rate for the VI setup was not known, and finding a general acceleration mechanism has been an open problem. In this paper, we present the first accelerated VI for both the Bellman consistency and optimality operators. Our method, called Anc-VI, is based on an \emph{anchoring} mechanism (distinct from Nesterov's acceleration), and it reduces the Bellman error faster than standard VI. In particular, Anc-VI exhibits a O(1/k)\mathcal{O}(1/k)O(1/k)-rate for γ≈1\gamma\approx 1γ≈1 or even γ=1\gamma=1γ=1, while standard VI has rate O(1)\mathcal{O}(1)O(1) for γ≥1−1/k\gamma\ge 1-1/kγ≥1−1/k, where kkk is the iteration count. We also provide a complexity lower bound matching the upper bound up to a constant factor of 444, thereby establishing optimality of the accelerated rate of Anc-VI. Finally, we show that the anchoring mechanism provides the same benefit in the approximate VI and Gauss--Seidel VI setups as well.

View on arXiv
Comments on this paper