Variational Representations and Neural Network Estimation of Rényi Divergences

We derive a new variational formula for the R\ényi family of divergences, , between probability measures and . Our result generalizes the classical Donsker-Varadhan variational formula for the Kullback-Leibler divergence. We further show that this R\ényi variational formula holds over a range of function spaces; this leads to a formula for the optimizer under very weak assumptions and is also key in our development of a consistency theory for R\ényi divergence estimators. By applying this theory to neural-network estimators, we show that if a neural network family satisfies one of several strengthened versions of the universal approximation property then the corresponding R\ényi divergence estimator is consistent. In contrast to density-estimator based methods, our estimators involve only expectations under and and hence are more effective in high dimensional systems. We illustrate this via several numerical examples of neural network estimation in systems of up to 5000 dimensions.
View on arXiv