ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2405.14078
42
0

A finite time analysis of distributed Q-learning

23 May 2024
Han-Dong Lim
Donghwan Lee
    OffRL
ArXivPDFHTML
Abstract

Multi-agent reinforcement learning (MARL) has witnessed a remarkable surge in interest, fueled by the empirical success achieved in applications of single-agent reinforcement learning (RL). In this study, we consider a distributed Q-learning scenario, wherein a number of agents cooperatively solve a sequential decision making problem without access to the central reward function which is an average of the local rewards. In particular, we study finite-time analysis of a distributed Q-learning algorithm, and provide a new sample complexity result of O~(min⁡{1ϵ2tmix(1−γ)6dmin⁡4,1ϵ∣\gS∣∣\gA∣(1−σ2(W))(1−γ)4dmin⁡3})\tilde{\mathcal{O}}\left( \min\left\{\frac{1}{\epsilon^2}\frac{t_{\text{mix}}}{(1-\gamma)^6 d_{\min}^4 } ,\frac{1}{\epsilon}\frac{\sqrt{|\gS||\gA|}}{(1-\sigma_2(\boldsymbol{W}))(1-\gamma)^4 d_{\min}^3} \right\}\right)O~(min{ϵ21​(1−γ)6dmin4​tmix​​,ϵ1​(1−σ2​(W))(1−γ)4dmin3​∣\gS∣∣\gA∣​​}) under tabular lookup

View on arXiv
Comments on this paper