Mutual Information Collapse Explains Disentanglement Failure in -VAEs
- DRLAAMLCoGe
The -VAE is a foundational framework for unsupervised disentanglement, using to regulate the trade-off between latent factorization and reconstruction fidelity. Empirically, however, disentanglement performance exhibits a pervasive non-monotonic trend: benchmarks such as MIG and SAP typically peak at intermediate and collapse as regularization increases. We demonstrate that this collapse is a fundamental information-theoretic failure, where strong Kullback-Leibler pressure promotes marginal independence at the expense of the latent channel's semantic informativeness. By formalizing this mechanism in a linear-Gaussian setting, we prove that for , stationarity-induced dynamics trigger a spectral contraction of the encoder gain, driving latent-factor mutual information to zero. To resolve this, we introduce the -VAE, which decouples regularization pressure from informational collapse via an auxiliary reconstruction penalty . Extensive experiments on dSprites, Shapes3D, and MPI3D-real confirm that stabilizes disentanglement and restores latent informativeness over a significantly broader range of , providing a principled theoretical justification for dual-parameter regularization in variational inference backbones.
View on arXiv