Learning Coverage Functions

We study the problem of approximating and learning coverage functions. A function is a coverage function, if there exists a universe with non-negative weights for each and subsets of such that . Alternatively, coverage functions can be described as non-negative linear combinations of monotone disjunctions. They are a natural subclass of submodular functions and arise in a number of applications. We show that over the uniform distribution coverage functions with range [0,1] are PAC learnable to -error of in time and using random examples. We also show a proper learning algorithm for coverage functions whose running time is polynomial in the size of the universe over which the coverage function is defined. Our algorithm is based on several new structural properties of the Fourier spectrum of coverage functions and, in particular, we prove that any coverage function can be -approximated in by a coverage function that depends only on variables. In contrast, we show that, without assumptions on the distribution, learning coverage is at least as hard as learning polynomial-size disjoint DNF formulas. Our PAC learning algorithm on the uniform distribution implies the first polynomial-time differentially private algorithm for releasing monotone disjunction queries with low average error over the uniform distribution on disjunctions. This problem was first considered by Gupta et al. (2011) and the best previous algorithm runs in time , where is the accuracy of release (Cheraghchi et al., 2012). Further, our proper learning algorithm implies that the queries can be released using a synthetic database in time .
View on arXiv