ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2405.04017
29
0

An Improved Finite-time Analysis of Temporal Difference Learning with Deep Neural Networks

7 May 2024
Zhifa Ke
Zaiwen Wen
Junyu Zhang
ArXivPDFHTML
Abstract

Temporal difference (TD) learning algorithms with neural network function parameterization have well-established empirical success in many practical large-scale reinforcement learning tasks. However, theoretical understanding of these algorithms remains challenging due to the nonlinearity of the action-value approximation. In this paper, we develop an improved non-asymptotic analysis of the neural TD method with a general LLL-layer neural network. New proof techniques are developed and an improved new O~(ϵ−1)\tilde{\mathcal{O}}(\epsilon^{-1})O~(ϵ−1) sample complexity is derived. To our best knowledge, this is the first finite-time analysis of neural TD that achieves an O~(ϵ−1)\tilde{\mathcal{O}}(\epsilon^{-1})O~(ϵ−1) complexity under the Markovian sampling, as opposed to the best known O~(ϵ−2)\tilde{\mathcal{O}}(\epsilon^{-2})O~(ϵ−2) complexity in the existing literature.

View on arXiv
Comments on this paper