31
7

The Sample Complexity of Approximate Rejection Sampling with Applications to Smoothed Online Learning

Abstract

Suppose we are given access to nn independent samples from distribution μ\mu and we wish to output one of them with the goal of making the output distributed as close as possible to a target distribution ν\nu. In this work we show that the optimal total variation distance as a function of nn is given by Θ~(Df(n))\tilde\Theta(\frac{D}{f'(n)}) over the class of all pairs ν,μ\nu,\mu with a bounded ff-divergence Df(νμ)DD_f(\nu\|\mu)\leq D. Previously, this question was studied only for the case when the Radon-Nikodym derivative of ν\nu with respect to μ\mu is uniformly bounded. We then consider an application in the seemingly very different field of smoothed online learning, where we show that recent results on the minimax regret and the regret of oracle-efficient algorithms still hold even under relaxed constraints on the adversary (to have bounded ff-divergence, as opposed to bounded Radon-Nikodym derivative). Finally, we also study efficacy of importance sampling for mean estimates uniform over a function class and compare importance sampling with rejection sampling.

View on arXiv
Comments on this paper