High dimensional errors-in-variables models with dependent measurements

Suppose that we observe and in the following errors-in-variables model: \begin{eqnarray*} y & = & X_0 \beta^* + \epsilon \\ X & = & X_0 + W \end{eqnarray*} where is a design matrix with independent subgaussian row vectors, is a noise vector and is a mean zero random noise matrix with independent subgaussian column vectors, independent of and . This model is significantly different from those analyzed in the literature in the sense that we allow the measurement error for each covariate to be a dependent vector across its observations. Such error structures appear in the science literature when modeling the trial-to-trial fluctuations in response strength shared across a set of neurons. Under sparsity and restrictive eigenvalue type of conditions, we show that one is able to recover a sparse vector from the model given a single observation matrix and the response vector . We establish consistency in estimating and obtain the rates of convergence in the norm, where for the Lasso-type estimator, and for for a Dantzig-type conic programming estimator. We show error bounds which approach that of the regular Lasso and the Dantzig selector in case the errors in are tending to 0.
View on arXiv