271

Near-Optimal Cryptographic Hardness of Agnostically Learning Halfspaces and ReLU Regression under Gaussian Marginals

International Conference on Machine Learning (ICML), 2023
Main:12 Pages
Bibliography:4 Pages
Appendix:7 Pages
Abstract

We study the task of agnostically learning halfspaces under the Gaussian distribution. Specifically, given labeled examples (x,y)(\mathbf{x},y) from an unknown distribution on Rn×{±1}\mathbb{R}^n \times \{ \pm 1\}, whose marginal distribution on x\mathbf{x} is the standard Gaussian and the labels yy can be arbitrary, the goal is to output a hypothesis with 0-1 loss OPT+ϵ\mathrm{OPT}+\epsilon, where OPT\mathrm{OPT} is the 0-1 loss of the best-fitting halfspace. We prove a near-optimal computational hardness result for this task, under the widely believed sub-exponential time hardness of the Learning with Errors (LWE) problem. Prior hardness results are either qualitatively suboptimal or apply to restricted families of algorithms. Our techniques extend to yield near-optimal lower bounds for related problems, including ReLU regression.

View on arXiv
Comments on this paper