ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1802.03171
  4. Cited By
A Unified Approach for Multi-step Temporal-Difference Learning with
  Eligibility Traces in Reinforcement Learning

A Unified Approach for Multi-step Temporal-Difference Learning with Eligibility Traces in Reinforcement Learning

9 February 2018
Long Yang
Minhao Shi
Qian Zheng
Wenjia Meng
Gang Pan
ArXivPDFHTML

Papers citing "A Unified Approach for Multi-step Temporal-Difference Learning with Eligibility Traces in Reinforcement Learning"

5 / 5 papers shown
Title
Constrained Update Projection Approach to Safe Policy Optimization
Constrained Update Projection Approach to Safe Policy Optimization
Long Yang
Jiaming Ji
Juntao Dai
Linrui Zhang
Binbin Zhou
Pengfei Li
Yaodong Yang
Gang Pan
41
43
0
15 Sep 2022
A Review of Safe Reinforcement Learning: Methods, Theory and
  Applications
A Review of Safe Reinforcement Learning: Methods, Theory and Applications
Shangding Gu
Longyu Yang
Yali Du
Guang Chen
Florian Walter
Jun Wang
Alois C. Knoll
OffRL
AI4TS
117
241
0
20 May 2022
Reducing Sampling Error in Batch Temporal Difference Learning
Reducing Sampling Error in Batch Temporal Difference Learning
Brahma S. Pavse
Ishan Durugkar
Josiah P. Hanna
Peter Stone
OffRL
22
12
0
15 Aug 2020
Gradient Q$(σ, λ)$: A Unified Algorithm with Function
  Approximation for Reinforcement Learning
Gradient Q(σ,λ)(σ, λ)(σ,λ): A Unified Algorithm with Function Approximation for Reinforcement Learning
Long Yang
Yu Zhang
Qian Zheng
Pengfei Li
Gang Pan
20
1
0
06 Sep 2019
Qualitative Measurements of Policy Discrepancy for Return-Based Deep
  Q-Network
Qualitative Measurements of Policy Discrepancy for Return-Based Deep Q-Network
Wenjia Meng
Qian Zheng
L. Yang
Pengfei Li
Gang Pan
20
21
0
14 Jun 2018
1