179

Learning and Testing Junta Distributions with Subcube Conditioning

Annual Conference Computational Learning Theory (COLT), 2020
Abstract

We study the problems of learning and testing junta distributions on {1,1}n\{-1,1\}^n with respect to the uniform distribution, where a distribution pp is a kk-junta if its probability mass function p(x)p(x) depends on a subset of at most kk variables. The main contribution is an algorithm for finding relevant coordinates in a kk-junta distribution with subcube conditioning [BC18, CCKLW20]. We give two applications: 1. An algorithm for learning kk-junta distributions with O~(k/ϵ2)logn+O(2k/ϵ2)\tilde{O}(k/\epsilon^2) \log n + O(2^k/\epsilon^2) subcube conditioning queries, and 2. An algorithm for testing kk-junta distributions with O~((k+n)/ϵ2)\tilde{O}((k + \sqrt{n})/\epsilon^2) subcube conditioning queries. All our algorithms are optimal up to poly-logarithmic factors. Our results show that subcube conditioning, as a natural model for accessing high-dimensional distributions, enables significant savings in learning and testing junta distributions compared to the standard sampling model. This addresses an open question posed by Aliakbarpour, Blais, and Rubinfeld [ABR17].

View on arXiv
Comments on this paper