ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2011.00213
  4. Cited By
Finding the Near Optimal Policy via Adaptive Reduced Regularization in
  MDPs

Finding the Near Optimal Policy via Adaptive Reduced Regularization in MDPs

31 October 2020
Wenhao Yang
Xiang Li
Guangzeng Xie
Zhihua Zhang
ArXivPDFHTML

Papers citing "Finding the Near Optimal Policy via Adaptive Reduced Regularization in MDPs"

1 / 1 papers shown
Title
Softmax Policy Gradient Methods Can Take Exponential Time to Converge
Softmax Policy Gradient Methods Can Take Exponential Time to Converge
Gen Li
Yuting Wei
Yuejie Chi
Yuxin Chen
26
50
0
22 Feb 2021
1