Consider the standard linear regression model , where is an observation vector, is a design matrix, is the unknown regression vector, and is additive Gaussian noise. This paper studies the minimax rates of convergence for estimation of for -losses and in the -prediction loss, assuming that belongs to an -ball for some . We show that under suitable regularity conditions on the design matrix , the minimax error in -loss and -prediction loss scales as . In addition, we provide lower bounds on minimax risks in -norms, for all . Our proofs of the lower bounds are information-theoretic in nature, based on Fano's inequality and results on the metric entropy of the balls , whereas our proofs of the upper bounds are direct and constructive, involving direct analysis of least-squares over -balls. For the special case , a comparison with -risks achieved by computationally efficient -relaxations reveals that although such methods can achieve the minimax rates up to constant factors, they require slightly stronger assumptions on the design matrix than algorithms involving least-squares over the -ball.
View on arXiv