99
41

Learning Coverage Functions

Abstract

We study the problem of approximating and learning coverage functions. A function c:2[n]R+c: 2^{[n]} \rightarrow \R^{+} is a coverage function, if there exists a universe UU with non-negative weights w(u)w(u) for each uUu \in U and subsets A1,A2,...,AnA_1, A_2, ..., A_n of UU such that c(S)=uiSAiw(u)c(S) = \sum_{u \in \cup_{i \in S} A_i} w(u). Alternatively, coverage functions can be described as non-negative linear combinations of monotone disjunctions. They are a natural subclass of submodular functions and arise in a number of applications. We give an algorithm that for any γ,δ>0\gamma,\delta>0, given random and uniform examples of an unknown coverage function cc, finds a function hh that approximates cc within factor (1+γ)(1+\gamma) on all but δ\delta-fraction of the points in time \poly(n,1/γ,1/δ)\poly(n,1/\gamma,1/\delta). This is the first fully-polynomial algorithm for learning an interesting class of functions in the demanding PMAC model of Balcan and Harvey (2011). Our algorithm relies on first solving a simpler problem of learning coverage functions with low 1\ell_1-error. Our algorithms are based on several new structural properties of coverage functions and, in particular, we prove that any coverage function can be \eps\eps-approximated in 1\ell_1 by a coverage function that depends only on O(1/\eps2)O(1/\eps^2) variables. In contrast, we show that, without assumptions on the distribution, learning coverage functions is at least as hard as learning polynomial-size disjoint DNF formulas, a class of function for which the best known algorithm runs in time nO~(n1/3)n^{\tilde{O}(n^{1/3})} (Klivans and Servedio, 2004). As an application of our result, we give a simple polynomial-time differentially-private algorithm for releasing monotone disjunction queries with low average error over the uniform distribution on disjunctions.

View on arXiv
Comments on this paper