ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2107.11433
74
62

A general sample complexity analysis of vanilla policy gradient

23 July 2021
Rui Yuan
Robert Mansel Gower
A. Lazaric
ArXivPDFHTML
Abstract

We adapt recent tools developed for the analysis of Stochastic Gradient Descent (SGD) in non-convex optimization to obtain convergence and sample complexity guarantees for the vanilla policy gradient (PG). Our only assumptions are that the expected return is smooth w.r.t. the policy parameters, that its HHH-step truncated gradient is close to the exact gradient, and a certain ABC assumption. This assumption requires the second moment of the estimated gradient to be bounded by A≥0A\geq 0A≥0 times the suboptimality gap, B≥0B \geq 0B≥0 times the norm of the full batch gradient and an additive constant C≥0C \geq 0C≥0, or any combination of aforementioned. We show that the ABC assumption is more general than the commonly used assumptions on the policy space to prove convergence to a stationary point. We provide a single convergence theorem that recovers the O~(ϵ−4)\widetilde{\mathcal{O}}(\epsilon^{-4})O(ϵ−4) sample complexity of PG to a stationary point. Our results also affords greater flexibility in the choice of hyper parameters such as the step size and the batch size mmm, including the single trajectory case (i.e., m=1m=1m=1). When an additional relaxed weak gradient domination assumption is available, we establish a novel global optimum convergence theory of PG with O~(ϵ−3)\widetilde{\mathcal{O}}(\epsilon^{-3})O(ϵ−3) sample complexity. We then instantiate our theorems in different settings, where we both recover existing results and obtain improved sample complexity, e.g., O~(ϵ−3)\widetilde{\mathcal{O}}(\epsilon^{-3})O(ϵ−3) sample complexity for the convergence to the global optimum for Fisher-non-degenerated parametrized policies.

View on arXiv
Comments on this paper