22
28

Properly Learning Poisson Binomial Distributions in Almost Polynomial Time

Abstract

We give an algorithm for properly learning Poisson binomial distributions. A Poisson binomial distribution (PBD) of order nn is the discrete probability distribution of the sum of nn mutually independent Bernoulli random variables. Given O~(1/ϵ2)\widetilde{O}(1/\epsilon^2) samples from an unknown PBD p\mathbf{p}, our algorithm runs in time (1/ϵ)O(loglog(1/ϵ))(1/\epsilon)^{O(\log \log (1/\epsilon))}, and outputs a hypothesis PBD that is ϵ\epsilon-close to p\mathbf{p} in total variation distance. The previously best known running time for properly learning PBDs was (1/ϵ)O(log(1/ϵ))(1/\epsilon)^{O(\log(1/\epsilon))}. As one of our main contributions, we provide a novel structural characterization of PBDs. We prove that, for all ϵ>0,\epsilon >0, there exists an explicit collection M\cal{M} of (1/ϵ)O(loglog(1/ϵ))(1/\epsilon)^{O(\log \log (1/\epsilon))} vectors of multiplicities, such that for any PBD p\mathbf{p} there exists a PBD q\mathbf{q} with O(log(1/ϵ))O(\log(1/\epsilon)) distinct parameters whose multiplicities are given by some element of M{\cal M}, such that q\mathbf{q} is ϵ\epsilon-close to p\mathbf{p}. Our proof combines tools from Fourier analysis and algebraic geometry. Our approach to the proper learning problem is as follows: Starting with an accurate non-proper hypothesis, we fit a PBD to this hypothesis. More specifically, we essentially start with the hypothesis computed by the computationally efficient non-proper learning algorithm in our recent work~\cite{DKS15}. Our aforementioned structural characterization allows us to reduce the corresponding fitting problem to a collection of (1/ϵ)O(loglog(1/ϵ))(1/\epsilon)^{O(\log \log(1/\epsilon))} systems of low-degree polynomial inequalities. We show that each such system can be solved in time (1/ϵ)O(loglog(1/ϵ))(1/\epsilon)^{O(\log \log(1/\epsilon))}, which yields the overall running time of our algorithm.

View on arXiv
Comments on this paper