ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2006.16200
11
66

Near-Optimal SQ Lower Bounds for Agnostically Learning Halfspaces and ReLUs under Gaussian Marginals

29 June 2020
Ilias Diakonikolas
D. Kane
Nikos Zarifis
ArXivPDFHTML
Abstract

We study the fundamental problems of agnostically learning halfspaces and ReLUs under Gaussian marginals. In the former problem, given labeled examples (x,y)(\mathbf{x}, y)(x,y) from an unknown distribution on Rd×{±1}\mathbb{R}^d \times \{ \pm 1\}Rd×{±1}, whose marginal distribution on x\mathbf{x}x is the standard Gaussian and the labels yyy can be arbitrary, the goal is to output a hypothesis with 0-1 loss OPT+ϵ\mathrm{OPT}+\epsilonOPT+ϵ, where OPT\mathrm{OPT}OPT is the 0-1 loss of the best-fitting halfspace. In the latter problem, given labeled examples (x,y)(\mathbf{x}, y)(x,y) from an unknown distribution on Rd×R\mathbb{R}^d \times \mathbb{R}Rd×R, whose marginal distribution on x\mathbf{x}x is the standard Gaussian and the labels yyy can be arbitrary, the goal is to output a hypothesis with square loss OPT+ϵ\mathrm{OPT}+\epsilonOPT+ϵ, where OPT\mathrm{OPT}OPT is the square loss of the best-fitting ReLU. We prove Statistical Query (SQ) lower bounds of dpoly(1/ϵ)d^{\mathrm{poly}(1/\epsilon)}dpoly(1/ϵ) for both of these problems. Our SQ lower bounds provide strong evidence that current upper bounds for these tasks are essentially best possible.

View on arXiv
Comments on this paper