171
65

Simultaneous support recovery in high dimensions: Benefits and perils of block 1/\ell_1/\ell_\infty-regularization

Abstract

Consider the use of 1/\ell_{1}/\ell_{\infty}-regularized regression for joint estimation of a \pdim×\numreg\pdim \times \numreg matrix of regression coefficients. We analyze the high-dimensional scaling of 1/\ell_1/\ell_\infty-regularized quadratic programming, considering both consistency in \ell_\infty-norm, and variable selection. We begin by establishing bounds on the \ell_\infty-error as well sufficient conditions for exact variable selection for fixed and random designs. Our second set of results applies to \numreg=2\numreg = 2 linear regression problems with standard Gaussian designs whose supports overlap in a fraction α[0,1]\alpha \in [0,1] of their entries: for this problem class, we prove that the 1/\ell_{1}/\ell_{\infty}-regularized method undergoes a phase transition--that is, a sharp change from failure to success--characterized by the rescaled sample size θ1,(n,p,s,α)=n/{(43α)slog(p(2α)s)}\theta_{1,\infty}(n, p, s, \alpha) = n/\{(4 - 3 \alpha) s \log(p-(2- \alpha) s)\}. An implication of this threshold is that use of 1/\ell_1 / \ell_{\infty}-regularization yields improved statistical efficiency if the overlap parameter is large enough (α>2/3\alpha > 2/3), but has \emph{worse} statistical efficiency than a naive Lasso-based approach for moderate to small overlap (α<2/3\alpha < 2/3). These results indicate that some caution needs to be exercised in the application of 1/\ell_1/\ell_\infty block regularization: if the data does not match its structure closely enough, it can impair statistical performance relative to computationally less expensive schemes.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.