95

Mutual Information Collapse Explains Disentanglement Failure in ββ-VAEs

Minh Vu
Xiaoliang Wan
Shuangqing Wei
Main:7 Pages
8 Figures
Bibliography:3 Pages
2 Tables
Appendix:3 Pages
Abstract

The β\beta-VAE is a foundational framework for unsupervised disentanglement, using β\beta to regulate the trade-off between latent factorization and reconstruction fidelity. Empirically, however, disentanglement performance exhibits a pervasive non-monotonic trend: benchmarks such as MIG and SAP typically peak at intermediate β\beta and collapse as regularization increases. We demonstrate that this collapse is a fundamental information-theoretic failure, where strong Kullback-Leibler pressure promotes marginal independence at the expense of the latent channel's semantic informativeness. By formalizing this mechanism in a linear-Gaussian setting, we prove that for β>1\beta > 1, stationarity-induced dynamics trigger a spectral contraction of the encoder gain, driving latent-factor mutual information to zero. To resolve this, we introduce the λβ\lambda\beta-VAE, which decouples regularization pressure from informational collapse via an auxiliary L2L_2 reconstruction penalty λ\lambda. Extensive experiments on dSprites, Shapes3D, and MPI3D-real confirm that λ>0\lambda > 0 stabilizes disentanglement and restores latent informativeness over a significantly broader range of β\beta, providing a principled theoretical justification for dual-parameter regularization in variational inference backbones.

View on arXiv
Comments on this paper