ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 0910.2042
48
572

Minimax rates of estimation for high-dimensional linear regression over ℓq\ell_qℓq​-balls

11 October 2009
Garvesh Raskutti
Martin J. Wainwright
Bin Yu
ArXivPDFHTML
Abstract

Consider the standard linear regression model \y=\Xmat\betastar+w\y = \Xmat \betastar + w\y=\Xmat\betastar+w, where \y∈ℜ\numobs\y \in \real^\numobs\y∈ℜ\numobs is an observation vector, \Xmat∈ℜ\numobs×\pdim\Xmat \in \real^{\numobs \times \pdim}\Xmat∈ℜ\numobs×\pdim is a design matrix, \betastar∈ℜ\pdim\betastar \in \real^\pdim\betastar∈ℜ\pdim is the unknown regression vector, and w∼N(0,σ2I)w \sim \mathcal{N}(0, \sigma^2 I)w∼N(0,σ2I) is additive Gaussian noise. This paper studies the minimax rates of convergence for estimation of \betastar\betastar\betastar for ℓ\rpar\ell_\rparℓ\rpar​-losses and in the ℓ2\ell_2ℓ2​-prediction loss, assuming that \betastar\betastar\betastar belongs to an ℓ\qpar\ell_{\qpar}ℓ\qpar​-ball \Ballq(\myrad)\Ballq(\myrad)\Ballq(\myrad) for some \qpar∈[0,1]\qpar \in [0,1]\qpar∈[0,1]. We show that under suitable regularity conditions on the design matrix \Xmat\Xmat\Xmat, the minimax error in ℓ2\ell_2ℓ2​-loss and ℓ2\ell_2ℓ2​-prediction loss scales as \Rq(log⁡\pdimn)1−\qpar2\Rq \big(\frac{\log \pdim}{n}\big)^{1-\frac{\qpar}{2}}\Rq(nlog\pdim​)1−2\qpar​. In addition, we provide lower bounds on minimax risks in ℓ\rpar\ell_{\rpar}ℓ\rpar​-norms, for all \rpar∈[1,+∞],\rpar≠\qpar\rpar \in [1, +\infty], \rpar \neq \qpar\rpar∈[1,+∞],\rpar=\qpar. Our proofs of the lower bounds are information-theoretic in nature, based on Fano's inequality and results on the metric entropy of the balls \Ballq(\myrad)\Ballq(\myrad)\Ballq(\myrad), whereas our proofs of the upper bounds are direct and constructive, involving direct analysis of least-squares over ℓ\qpar\ell_{\qpar}ℓ\qpar​-balls. For the special case q=0q = 0q=0, a comparison with ℓ2\ell_2ℓ2​-risks achieved by computationally efficient ℓ1\ell_1ℓ1​-relaxations reveals that although such methods can achieve the minimax rates up to constant factors, they require slightly stronger assumptions on the design matrix \Xmat\Xmat\Xmat than algorithms involving least-squares over the ℓ0\ell_0ℓ0​-ball.

View on arXiv
Comments on this paper