15
14

The Price of Tolerance in Distribution Testing

Abstract

We revisit the problem of tolerant distribution testing. That is, given samples from an unknown distribution pp over {1,,n}\{1, \dots, n\}, is it ε1\varepsilon_1-close to or ε2\varepsilon_2-far from a reference distribution qq (in total variation distance)? Despite significant interest over the past decade, this problem is well understood only in the extreme cases. In the noiseless setting (i.e., ε1=0\varepsilon_1 = 0) the sample complexity is Θ(n)\Theta(\sqrt{n}), strongly sublinear in the domain size. At the other end of the spectrum, when ε1=ε2/2\varepsilon_1 = \varepsilon_2/2, the sample complexity jumps to the barely sublinear Θ(n/logn)\Theta(n/\log n). However, very little is known about the intermediate regime. We fully characterize the price of tolerance in distribution testing as a function of nn, ε1\varepsilon_1, ε2\varepsilon_2, up to a single logn\log n factor. Specifically, we show the sample complexity to be \[\tilde \Theta\left(\frac{\sqrt{n}}{\varepsilon_2^{2}} + \frac{n}{\log n} \cdot \max \left\{\frac{\varepsilon_1}{\varepsilon_2^2},\left(\frac{\varepsilon_1}{\varepsilon_2^2}\right)^{\!\!2}\right\}\right),\] providing a smooth tradeoff between the two previously known cases. We also provide a similar characterization for the problem of tolerant equivalence testing, where both pp and qq are unknown. Surprisingly, in both cases, the main quantity dictating the sample complexity is the ratio ε1/ε22\varepsilon_1/\varepsilon_2^2, and not the more intuitive ε1/ε2\varepsilon_1/\varepsilon_2. Of particular technical interest is our lower bound framework, which involves novel approximation-theoretic tools required to handle the asymmetry between ε1\varepsilon_1 and ε2\varepsilon_2, a challenge absent from previous works.

View on arXiv
Comments on this paper