ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1802.07842
  4. Cited By
Convergent Actor-Critic Algorithms Under Off-Policy Training and
  Function Approximation

Convergent Actor-Critic Algorithms Under Off-Policy Training and Function Approximation

21 February 2018
H. Maei
    OffRL
ArXiv (abs)PDFHTML

Papers citing "Convergent Actor-Critic Algorithms Under Off-Policy Training and Function Approximation"

2 / 2 papers shown
Title
Q-Prop: Sample-Efficient Policy Gradient with An Off-Policy Critic
Q-Prop: Sample-Efficient Policy Gradient with An Off-Policy Critic
S. Gu
Timothy Lillicrap
Zoubin Ghahramani
Richard Turner
Sergey Levine
OffRLBDL
88
345
0
07 Nov 2016
An Emphatic Approach to the Problem of Off-policy Temporal-Difference
  Learning
An Emphatic Approach to the Problem of Off-policy Temporal-Difference Learning
R. Sutton
A. R. Mahmood
Martha White
91
272
0
14 Mar 2015
1