33
25

Preconditioning to comply with the Irrepresentable Condition

Abstract

Preconditioning is a technique from numerical linear algebra that can accelerate algorithms to solve systems of equations. In this paper, we demonstrate how preconditioning can circumvent a stringent assumption for sign consistency in sparse linear regression. Given XRn×pX \in R^{n \times p} and YRnY \in R^n that satisfy the standard regression equation, this paper demonstrates that even if the design matrix XX does not satisfy the irrepresentable condition for the Lasso, the design matrix FXF X often does, where FRn×nF \in R^{n\times n} is a preconditioning matrix defined in this paper. By computing the Lasso on (FX,FY)(F X, F Y), instead of on (X,Y)(X, Y), the necessary assumptions on XX become much less stringent. Our preconditioner FF ensures that the singular values of the design matrix are either zero or one. When npn\ge p, the columns of FXF X are orthogonal and the preconditioner always circumvents the stringent assumptions. When pnp\ge n, FF projects the design matrix onto the Stiefel manifold; the rows of FXF X are orthogonal. We give both theoretical results and simulation results to show that, in the high dimensional case, the preconditioner helps to circumvent the stringent assumptions, improving the statistical performance of a broad class of model selection techniques in linear regression. Simulation results are particularly promising.

View on arXiv
Comments on this paper