15
1

Gradient Q(σ,λ)(σ, λ): A Unified Algorithm with Function Approximation for Reinforcement Learning

Long Yang
Yu Zhang
Qian Zheng
Pengfei Li
Gang Pan
Abstract

Full-sampling (e.g., Q-learning) and pure-expectation (e.g., Expected Sarsa) algorithms are efficient and frequently used techniques in reinforcement learning. Q(σ,λ)(\sigma,\lambda) is the first approach unifies them with eligibility trace through the sampling degree σ\sigma. However, it is limited to the tabular case, for large-scale learning, the Q(σ,λ)(\sigma,\lambda) is too expensive to require a huge volume of tables to accurately storage value functions. To address above problem, we propose a GQ(σ,λ)(\sigma,\lambda) that extends tabular Q(σ,λ)(\sigma,\lambda) with linear function approximation. We prove the convergence of GQ(σ,λ)(\sigma,\lambda). Empirical results on some standard domains show that GQ(σ,λ)(\sigma,\lambda) with a combination of full-sampling with pure-expectation reach a better performance than full-sampling and pure-expectation methods.

View on arXiv
Comments on this paper