ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2408.11513
  4. Cited By
Last-Iterate Convergence of General Parameterized Policies in
  Constrained MDPs

Last-Iterate Convergence of General Parameterized Policies in Constrained MDPs

21 August 2024
Washim Uddin Mondal
Vaneet Aggarwal
ArXiv (abs)PDFHTML

Papers citing "Last-Iterate Convergence of General Parameterized Policies in Constrained MDPs"

13 / 13 papers shown
Title
Improved Sample Complexity Analysis of Natural Policy Gradient Algorithm
  with General Parameterization for Infinite Horizon Discounted Reward Markov
  Decision Processes
Improved Sample Complexity Analysis of Natural Policy Gradient Algorithm with General Parameterization for Infinite Horizon Discounted Reward Markov Decision Processes
Washim Uddin Mondal
Vaneet Aggarwal
54
11
0
18 Oct 2023
Last-Iterate Convergent Policy Gradient Primal-Dual Methods for
  Constrained MDPs
Last-Iterate Convergent Policy Gradient Primal-Dual Methods for Constrained MDPs
Dongsheng Ding
Chen-Yu Wei
Kai Zhang
Alejandro Ribeiro
87
22
0
20 Jun 2023
Near-Optimal Sample Complexity Bounds for Constrained MDPs
Near-Optimal Sample Complexity Bounds for Constrained MDPs
Sharan Vaswani
Lin F. Yang
Csaba Szepesvári
75
36
0
13 Jun 2022
Achieving Zero Constraint Violation for Constrained Reinforcement
  Learning via Conservative Natural Policy Gradient Primal-Dual Algorithm
Achieving Zero Constraint Violation for Constrained Reinforcement Learning via Conservative Natural Policy Gradient Primal-Dual Algorithm
Qinbo Bai
Amrit Singh Bedi
Vaneet Aggarwal
56
23
0
12 Jun 2022
Algorithm for Constrained Markov Decision Process with Linear
  Convergence
Algorithm for Constrained Markov Decision Process with Linear Convergence
E. Gladin
Maksim Lavrik-Karmazin
K. Zainullina
Varvara Rudenko
Alexander V. Gasnikov
Martin Takáč
75
7
0
03 Jun 2022
PAGE-PG: A Simple and Loopless Variance-Reduced Policy Gradient Method
  with Probabilistic Gradient Estimation
PAGE-PG: A Simple and Loopless Variance-Reduced Policy Gradient Method with Probabilistic Gradient Estimation
Matilde Gargiani
Andrea Zanelli
Andrea Martinelli
Tyler H. Summers
John Lygeros
50
14
0
01 Feb 2022
A Dual Approach to Constrained Markov Decision Processes with Entropy
  Regularization
A Dual Approach to Constrained Markov Decision Processes with Entropy Regularization
Donghao Ying
Yuhao Ding
Javad Lavaei
31
34
0
17 Oct 2021
Learning Policies with Zero or Bounded Constraint Violation for
  Constrained MDPs
Learning Policies with Zero or Bounded Constraint Violation for Constrained MDPs
Tao-Wen Liu
Ruida Zhou
D. Kalathil
P. R. Kumar
Chao Tian
70
84
0
04 Jun 2021
On the Convergence and Sample Efficiency of Variance-Reduced Policy
  Gradient Method
On the Convergence and Sample Efficiency of Variance-Reduced Policy Gradient Method
Junyu Zhang
Chengzhuo Ni
Zheng Yu
Csaba Szepesvári
Mengdi Wang
94
68
0
17 Feb 2021
CRPO: A New Approach for Safe Reinforcement Learning with Convergence
  Guarantee
CRPO: A New Approach for Safe Reinforcement Learning with Convergence Guarantee
Tengyu Xu
Yingbin Liang
Guanghui Lan
76
126
0
11 Nov 2020
On Linear Convergence of Policy Gradient Methods for Finite MDPs
On Linear Convergence of Policy Gradient Methods for Finite MDPs
Jalaj Bhandari
Daniel Russo
88
61
0
21 Jul 2020
Provably Efficient Safe Exploration via Primal-Dual Policy Optimization
Provably Efficient Safe Exploration via Primal-Dual Policy Optimization
Dongsheng Ding
Xiaohan Wei
Zhuoran Yang
Zhaoran Wang
M. Jovanović
77
165
0
01 Mar 2020
An Improved Convergence Analysis of Stochastic Variance-Reduced Policy
  Gradient
An Improved Convergence Analysis of Stochastic Variance-Reduced Policy Gradient
Pan Xu
F. Gao
Quanquan Gu
64
96
0
29 May 2019
1