19
5

Agnostic proper learning of monotone functions: beyond the black-box correction barrier

Abstract

We give the first agnostic, efficient, proper learning algorithm for monotone Boolean functions. Given 2O~(n/ε)2^{\tilde{O}(\sqrt{n}/\varepsilon)} uniformly random examples of an unknown function f:{±1}n{±1}f:\{\pm 1\}^n \rightarrow \{\pm 1\}, our algorithm outputs a hypothesis g:{±1}n{±1}g:\{\pm 1\}^n \rightarrow \{\pm 1\} that is monotone and (opt+ε)(\mathrm{opt} + \varepsilon)-close to ff, where opt\mathrm{opt} is the distance from ff to the closest monotone function. The running time of the algorithm (and consequently the size and evaluation time of the hypothesis) is also 2O~(n/ε)2^{\tilde{O}(\sqrt{n}/\varepsilon)}, nearly matching the lower bound of Blais et al (RANDOM '15). We also give an algorithm for estimating up to additive error ε\varepsilon the distance of an unknown function ff to monotone using a run-time of 2O~(n/ε)2^{\tilde{O}(\sqrt{n}/\varepsilon)}. Previously, for both of these problems, sample-efficient algorithms were known, but these algorithms were not run-time efficient. Our work thus closes this gap in our knowledge between the run-time and sample complexity. This work builds upon the improper learning algorithm of Bshouty and Tamon (JACM '96) and the proper semiagnostic learning algorithm of Lange, Rubinfeld, and Vasilyan (FOCS '22), which obtains a non-monotone Boolean-valued hypothesis, then ``corrects'' it to monotone using query-efficient local computation algorithms on graphs. This black-box correction approach can achieve no error better than 2opt+ε2\mathrm{opt} + \varepsilon information-theoretically; we bypass this barrier by a) augmenting the improper learner with a convex optimization step, and b) learning and correcting a real-valued function before rounding its values to Boolean. Our real-valued correction algorithm solves the ``poset sorting'' problem of [LRV22] for functions over general posets with non-Boolean labels.

View on arXiv
Comments on this paper