ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2212.03905
21
17

Multi-Rate VAE: Train Once, Get the Full Rate-Distortion Curve

7 December 2022
Juhan Bae
Michael Ruogu Zhang
Michael Ruan
Eric Wang
S. Hasegawa
Jimmy Ba
Roger C. Grosse
    DRL
ArXivPDFHTML
Abstract

Variational autoencoders (VAEs) are powerful tools for learning latent representations of data used in a wide range of applications. In practice, VAEs usually require multiple training rounds to choose the amount of information the latent variable should retain. This trade-off between the reconstruction error (distortion) and the KL divergence (rate) is typically parameterized by a hyperparameter β\betaβ. In this paper, we introduce Multi-Rate VAE (MR-VAE), a computationally efficient framework for learning optimal parameters corresponding to various β\betaβ in a single training run. The key idea is to explicitly formulate a response function that maps β\betaβ to the optimal parameters using hypernetworks. MR-VAEs construct a compact response hypernetwork where the pre-activations are conditionally gated based on β\betaβ. We justify the proposed architecture by analyzing linear VAEs and showing that it can represent response functions exactly for linear VAEs. With the learned hypernetwork, MR-VAEs can construct the rate-distortion curve without additional training and can be deployed with significantly less hyperparameter tuning. Empirically, our approach is competitive and often exceeds the performance of multiple β\betaβ-VAEs training with minimal computation and memory overheads.

View on arXiv
Comments on this paper