ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1806.00040
19
161

Efficient Algorithms and Lower Bounds for Robust Linear Regression

31 May 2018
Ilias Diakonikolas
Weihao Kong
Alistair Stewart
ArXivPDFHTML
Abstract

We study the problem of high-dimensional linear regression in a robust model where an ϵ\epsilonϵ-fraction of the samples can be adversarially corrupted. We focus on the fundamental setting where the covariates of the uncorrupted samples are drawn from a Gaussian distribution N(0,Σ)\mathcal{N}(0, \Sigma)N(0,Σ) on Rd\mathbb{R}^dRd. We give nearly tight upper bounds and computational lower bounds for this problem. Specifically, our main contributions are as follows: For the case that the covariance matrix is known to be the identity, we give a sample near-optimal and computationally efficient algorithm that outputs a candidate hypothesis vector β^\widehat{\beta}β​ which approximates the unknown regression vector β\betaβ within ℓ2\ell_2ℓ2​-norm O(ϵlog⁡(1/ϵ)σ)O(\epsilon \log(1/\epsilon) \sigma)O(ϵlog(1/ϵ)σ), where σ\sigmaσ is the standard deviation of the random observation noise. An error of Ω(ϵσ)\Omega (\epsilon \sigma)Ω(ϵσ) is information-theoretically necessary, even with infinite sample size. Prior work gave an algorithm for this problem with sample complexity Ω~(d2/ϵ2)\tilde{\Omega}(d^2/\epsilon^2)Ω~(d2/ϵ2) whose error guarantee scales with the ℓ2\ell_2ℓ2​-norm of β\betaβ. For the case of unknown covariance, we show that we can efficiently achieve the same error guarantee as in the known covariance case using an additional O~(d2/ϵ2)\tilde{O}(d^2/\epsilon^2)O~(d2/ϵ2) unlabeled examples. On the other hand, an error of O(ϵσ)O(\epsilon \sigma)O(ϵσ) can be information-theoretically attained with O(d/ϵ2)O(d/\epsilon^2)O(d/ϵ2) samples. We prove a Statistical Query (SQ) lower bound providing evidence that this quadratic tradeoff in the sample size is inherent. More specifically, we show that any polynomial time SQ learning algorithm for robust linear regression (in Huber's contamination model) with estimation complexity O(d2−c)O(d^{2-c})O(d2−c), where c>0c>0c>0 is an arbitrarily small constant, must incur an error of Ω(ϵσ)\Omega(\sqrt{\epsilon} \sigma)Ω(ϵ​σ).

View on arXiv
Comments on this paper