Likelihood-free hypothesis testing

Consider the problem of binary hypothesis testing. Given coming from either or , to decide between the two with small probability of error it is sufficient, and in many cases necessary, to have , where measures the separation between and in total variation (). Achieving this, however, requires complete knowledge of the distributions and can be done, for example, using the Neyman-Pearson test. In this paper we consider a variation of the problem which we call likelihood-free hypothesis testing, where access to and is given through i.i.d. observations from each. In the case when and are assumed to belong to a non-parametric family, we demonstrate the existence of a fundamental trade-off between and given by nm\asymp n_\sf{GoF}^2(\epsilon), where n_\sf{GoF}(\epsilon) is the minimax sample complexity of testing between the hypotheses vs . We show this for three families of distributions, in addition to the family of all discrete distributions for which we obtain a more complicated trade-off exhibiting an additional phase-transition. Our results demonstrate the possibility of testing without fully estimating and , provided .
View on arXiv