27
0

A Convergence Theory for Diffusion Language Models: An Information-Theoretic Perspective

Main:14 Pages
Bibliography:3 Pages
Abstract

Diffusion models have emerged as a powerful paradigm for modern generative modeling, demonstrating strong potential for large language models (LLMs). Unlike conventional autoregressive (AR) models that generate tokens sequentially, diffusion models enable parallel token sampling, leading to faster generation and eliminating left-to-right generation constraints. Despite their empirical success, the theoretical understanding of diffusion model approaches remains underdeveloped. In this work, we develop convergence guarantees for diffusion language models from an information-theoretic perspective. Our analysis demonstrates that the sampling error, measured by the Kullback-Leibler (KL) divergence, decays inversely with the number of iterations TT and scales linearly with the mutual information between tokens in the target text sequence. In particular, we establish matching upper and lower bounds, up to some constant factor, to demonstrate the tightness of our convergence analysis. These results offer novel theoretical insights into the practical effectiveness of diffusion language models.

View on arXiv
@article{li2025_2505.21400,
  title={ A Convergence Theory for Diffusion Language Models: An Information-Theoretic Perspective },
  author={ Gen Li and Changxiao Cai },
  journal={arXiv preprint arXiv:2505.21400},
  year={ 2025 }
}
Comments on this paper