Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2408.11513
Cited By
Last-Iterate Convergence of General Parameterized Policies in Constrained MDPs
21 August 2024
Washim Uddin Mondal
Vaneet Aggarwal
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Last-Iterate Convergence of General Parameterized Policies in Constrained MDPs"
13 / 13 papers shown
Title
Improved Sample Complexity Analysis of Natural Policy Gradient Algorithm with General Parameterization for Infinite Horizon Discounted Reward Markov Decision Processes
Washim Uddin Mondal
Vaneet Aggarwal
54
11
0
18 Oct 2023
Last-Iterate Convergent Policy Gradient Primal-Dual Methods for Constrained MDPs
Dongsheng Ding
Chen-Yu Wei
Kai Zhang
Alejandro Ribeiro
87
22
0
20 Jun 2023
Near-Optimal Sample Complexity Bounds for Constrained MDPs
Sharan Vaswani
Lin F. Yang
Csaba Szepesvári
75
36
0
13 Jun 2022
Achieving Zero Constraint Violation for Constrained Reinforcement Learning via Conservative Natural Policy Gradient Primal-Dual Algorithm
Qinbo Bai
Amrit Singh Bedi
Vaneet Aggarwal
56
23
0
12 Jun 2022
Algorithm for Constrained Markov Decision Process with Linear Convergence
E. Gladin
Maksim Lavrik-Karmazin
K. Zainullina
Varvara Rudenko
Alexander V. Gasnikov
Martin Takáč
75
7
0
03 Jun 2022
PAGE-PG: A Simple and Loopless Variance-Reduced Policy Gradient Method with Probabilistic Gradient Estimation
Matilde Gargiani
Andrea Zanelli
Andrea Martinelli
Tyler H. Summers
John Lygeros
50
14
0
01 Feb 2022
A Dual Approach to Constrained Markov Decision Processes with Entropy Regularization
Donghao Ying
Yuhao Ding
Javad Lavaei
31
34
0
17 Oct 2021
Learning Policies with Zero or Bounded Constraint Violation for Constrained MDPs
Tao-Wen Liu
Ruida Zhou
D. Kalathil
P. R. Kumar
Chao Tian
70
84
0
04 Jun 2021
On the Convergence and Sample Efficiency of Variance-Reduced Policy Gradient Method
Junyu Zhang
Chengzhuo Ni
Zheng Yu
Csaba Szepesvári
Mengdi Wang
94
68
0
17 Feb 2021
CRPO: A New Approach for Safe Reinforcement Learning with Convergence Guarantee
Tengyu Xu
Yingbin Liang
Guanghui Lan
76
126
0
11 Nov 2020
On Linear Convergence of Policy Gradient Methods for Finite MDPs
Jalaj Bhandari
Daniel Russo
88
61
0
21 Jul 2020
Provably Efficient Safe Exploration via Primal-Dual Policy Optimization
Dongsheng Ding
Xiaohan Wei
Zhuoran Yang
Zhaoran Wang
M. Jovanović
77
165
0
01 Mar 2020
An Improved Convergence Analysis of Stochastic Variance-Reduced Policy Gradient
Pan Xu
F. Gao
Quanquan Gu
64
96
0
29 May 2019
1