15
0

Refined αα-Divergence Variational Inference via Rejection Sampling

Abstract

We present an approximate inference method, based on a synergistic combination of R\ényi α\alpha-divergence variational inference (RDVI) and rejection sampling (RS). RDVI is based on minimization of R\ényi α\alpha-divergence Dα(pq)D_\alpha(p||q) between the true distribution p(x)p(x) and a variational approximation q(x)q(x); RS draws samples from a distribution p(x)=p~(x)/Zpp(x) = \tilde{p}(x)/Z_{p} using a proposal q(x)q(x), s.t. Mq(x)p~(x),xMq(x) \geq \tilde{p}(x), \forall x. Our inference method is based on a crucial observation that D(pq)D_\infty(p||q) equals logM(θ)\log M(\theta) where M(θ)M(\theta) is the optimal value of the RS constant for a given proposal qθ(x)q_\theta(x). This enables us to develop a \emph{two-stage} hybrid inference algorithm. Stage-1 performs RDVI to learn qθq_\theta by minimizing an estimator of Dα(pq)D_\alpha(p||q), and uses the learned qθq_\theta to find an (approximately) optimal M~(θ)\tilde{M}(\theta). Stage-2 performs RS using the constant M~(θ)\tilde{M}(\theta) to improve the approximate distribution qθq_\theta and obtain a sample-based approximation. We prove that this two-stage method allows us to learn considerably more accurate approximations of the target distribution as compared to RDVI. We demonstrate our method's efficacy via several experiments on synthetic and real datasets.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.