118
27

A Sharp Convergence Theory for The Probability Flow ODEs of Diffusion Models

Abstract

Diffusion models, which convert noise into new data instances by learning to reverse a diffusion process, have become a cornerstone in contemporary generative modeling. In this work, we develop non-asymptotic convergence theory for a popular diffusion-based sampler (i.e., the probability flow ODE sampler) in discrete time, assuming access to 2\ell_2-accurate estimates of the (Stein) score functions. For distributions in Rd\mathbb{R}^d, we prove that d/εd/\varepsilon iterations -- modulo some logarithmic and lower-order terms -- are sufficient to approximate the target distribution to within ε\varepsilon total-variation distance. This is the first result establishing nearly linear dimension-dependency (in dd) for the probability flow ODE sampler. Imposing only minimal assumptions on the target data distribution (e.g., no smoothness assumption is imposed), our results also characterize how 2\ell_2 score estimation errors affect the quality of the data generation processes. In contrast to prior works, our theory is developed based on an elementary yet versatile non-asymptotic approach without the need of resorting to SDE and ODE toolboxes.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.