47
3
v1v2 (latest)

Some Notes on the Sample Complexity of Approximate Channel Simulation

Abstract

Channel simulation algorithms can efficiently encode random samples from a prescribed target distribution QQ and find applications in machine learning-based lossy data compression. However, algorithms that encode exact samples usually have random runtime, limiting their applicability when a consistent encoding time is desirable. Thus, this paper considers approximate schemes with a fixed runtime instead. First, we strengthen a result of Agustsson and Theis and show that there is a class of pairs of target distribution QQ and coding distribution PP, for which the runtime of any approximate scheme scales at least super-polynomially in D[QP]D_\infty[Q \Vert P]. We then show, by contrast, that if we have access to an unnormalised Radon-Nikodym derivative rdQ/dPr \propto dQ/dP and knowledge of DKL[QP]D_{KL}[Q \Vert P], we can exploit global-bound, depth-limited A* coding to ensure TV[QP]ϵ\mathrm{TV}[Q \Vert P] \leq \epsilon and maintain optimal coding performance with a sample complexity of only exp2((DKL[QP]+o(1))/ϵ)\exp_2\big((D_{KL}[Q \Vert P] + o(1)) \big/ \epsilon\big).

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.