ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2403.08896
39
4

One-Shot Averaging for Distributed TD(λλλ) Under Markov Sampling

13 March 2024
Haoxing Tian
I. Paschalidis
Alexander Olshevsky
    OffRL
ArXivPDFHTML
Abstract

We consider a distributed setup for reinforcement learning, where each agent has a copy of the same Markov Decision Process but transitions are sampled from the corresponding Markov chain independently by each agent. We show that in this setting, we can achieve a linear speedup for TD(λ\lambdaλ), a family of popular methods for policy evaluation, in the sense that NNN agents can evaluate a policy NNN times faster provided the target accuracy is small enough. Notably, this speedup is achieved by ``one shot averaging,'' a procedure where the agents run TD(λ\lambdaλ) with Markov sampling independently and only average their results after the final step. This significantly reduces the amount of communication required to achieve a linear speedup relative to previous work.

View on arXiv
Comments on this paper