Learning Coverage Functions

We study the problem of approximating and learning coverage functions. A function is a coverage function, if there exists a universe with non-negative weights for each and subsets of such that . Alternatively, coverage functions can be described as non-negative linear combinations of monotone disjunctions. They are a natural subclass of submodular functions and arise in a number of applications. We give an algorithm that for any , given random and uniform examples of an unknown coverage function , finds a function that approximates within factor on all but -fraction of the points in time . This is the first fully-polynomial algorithm for learning an interesting class of functions in the demanding PMAC model of Balcan and Harvey (2011). Our algorithm relies on first solving a simpler problem of learning coverage functions with low -error. Our algorithms are based on several new structural properties of coverage functions and, in particular, we prove that any coverage function can be -approximated in by a coverage function that depends only on variables. In contrast, we show that, without assumptions on the distribution, learning coverage functions is at least as hard as learning polynomial-size disjoint DNF formulas, a class of function for which the best known algorithm runs in time (Klivans and Servedio, 2004). As an application of our result, we give a simple polynomial-time differentially-private algorithm for releasing monotone disjunction queries with low average error over the uniform distribution on disjunctions.
View on arXiv