28
52

A Sample-Efficient Algorithm for Episodic Finite-Horizon MDP with Constraints

Abstract

Constrained Markov Decision Processes (CMDPs) formalize sequential decision-making problems whose objective is to minimize a cost function while satisfying constraints on various cost functions. In this paper, we consider the setting of episodic fixed-horizon CMDPs. We propose an online algorithm which leverages the linear programming formulation of finite-horizon CMDP for repeated optimistic planning to provide a probably approximately correct (PAC) guarantee on the number of episodes needed to ensure an ϵ\epsilon-optimal policy, i.e., with resulting objective value within ϵ\epsilon of the optimal value and satisfying the constraints within ϵ\epsilon-tolerance, with probability at least 1δ1-\delta. The number of episodes needed is shown to be of the order O~(SAC2H2ϵ2log1δ)\tilde{\mathcal{O}}\big(\frac{|S||A|C^{2}H^{2}}{\epsilon^{2}}\log\frac{1}{\delta}\big), where CC is the upper bound on the number of possible successor states for a state-action pair. Therefore, if CSC \ll |S|, the number of episodes needed have a linear dependence on the state and action space sizes S|S| and A|A|, respectively, and quadratic dependence on the time horizon HH.

View on arXiv
Comments on this paper