10
5

Average case analysis of Lasso under ultra-sparse conditions

Abstract

We analyze the performance of the least absolute shrinkage and selection operator (Lasso) for the linear model when the number of regressors NN grows larger keeping the true support size dd finite, i.e., the ultra-sparse case. The result is based on a novel treatment of the non-rigorous replica method in statistical physics, which has been applied only to problem settings where NN ,dd and the number of observations MM tend to infinity at the same rate. Our analysis makes it possible to assess the average performance of Lasso with Gaussian sensing matrices without assumptions on the scaling of NN and MM, the noise distribution, and the profile of the true signal. Under mild conditions on the noise distribution, the analysis also offers a lower bound on the sample complexity necessary for partial and perfect support recovery when MM diverges as M=O(logN)M = O(\log N). The obtained bound for perfect support recovery is a generalization of that given in previous literature, which only considers the case of Gaussian noise and diverging dd. Extensive numerical experiments strongly support our analysis.

View on arXiv
Comments on this paper