14
3

Learning stochastic decision trees

Abstract

We give a quasipolynomial-time algorithm for learning stochastic decision trees that is optimally resilient to adversarial noise. Given an η\eta-corrupted set of uniform random samples labeled by a size-ss stochastic decision tree, our algorithm runs in time nO(log(s/ε)/ε2)n^{O(\log(s/\varepsilon)/\varepsilon^2)} and returns a hypothesis with error within an additive 2η+ε2\eta + \varepsilon of the Bayes optimal. An additive 2η2\eta is the information-theoretic minimum. Previously no non-trivial algorithm with a guarantee of O(η)+εO(\eta) + \varepsilon was known, even for weaker noise models. Our algorithm is furthermore proper, returning a hypothesis that is itself a decision tree; previously no such algorithm was known even in the noiseless setting.

View on arXiv
Comments on this paper