Improved and Generalized Upper Bounds on the Complexity of Policy
Iteration
Given a Markov Decision Process (MDP) with states and actions per state, we study the number of iterations needed by Policy Iteratio n (PI) algorithms to converge. We consider two variations of PI: Howard's PI that changes all the actions with a positive advantage, and Sim plex-PI that only changes one action with maximal advantage. We show that Howard's PI terminates after at most $ n(m-1) \lceil \frac{1}{1-\gamma}\log (\frac{1}{1-\gamma}) \rceil $ iterations, improving by a factor a result by Hansen et al. (2013), while Simplex-PI terminates after at most $ n(m-1) \lceil \frac{n}{1-\gamma} \log (\frac{n}{1-\gamma})\rceil $ iterations, improving by a factor 2 a result by Ye (2011). We then consider bounds that are independent of the discount factor . When the MDP is deterministic, we show that Simplex-PI terminates after at most $ 2 n^2 m (m-1) \lceil 2 (n-1) \log n \rceil \lceil 2 n \log n \rceil = O(n^4 m^2 \log^2 n) $ iterations, improving by a factor a bound obtained by Post and Ye (2012). We generalize this result to general MDPs under some structural assumptions: given a measure of the maximal transient time and the maximal time to revisit states in recurrent classes under all policies, we show that Simplex-PI terminates after at most $ n^2 m (m-1) (\lceil \tau_r \log (n \tau_r) \rceil +\lceil \tau_r \log (n \tau_t) \rceil) \lceil {\tau_t} \log (n (\tau_t+1)) \rceil = \tilde O (n^2 \tau_t \tau_r m^2) $ iterations. We explain why similar results seem hard to derive for Howard's PI. Finally, under the additional (restrictive) assumption that the MDP is weakly-communicating, we show that Simplex-PI and Howard's PI terminate after at most iterations.
View on arXiv