Convergence of Langevin MCMC in KL-divergence

Abstract
Langevin diffusion is a commonly used tool for sampling from a given distribution. In this work, we establish that when the target density is such that is smooth and strongly convex, discrete Langevin diffusion produces a distribution with in steps, where is the dimension of the sample space. We also study the convergence rate when the strong-convexity assumption is absent. By considering the Langevin diffusion as a gradient flow in the space of probability distributions, we obtain an elegant analysis that applies to the stronger property of convergence in KL-divergence and gives a conceptually simpler proof of the best-known convergence results in weaker metrics.
View on arXivComments on this paper