30
11

SURE Information Criteria for Large Covariance Matrix Estimation and Their Asymptotic Properties

Abstract

Consider nn independent and identically distributed pp-dimensional Gaussian random vectors with covariance matrix Σ.\Sigma. The problem of estimating Σ\Sigma when pp is much larger than nn has received a lot of attention in recent years. Yet little is known about the information criterion for covariance matrix estimation. How to properly define such a criterion and what are the statistical properties? We attempt to answer these questions in the present paper by focusing on the estimation of bandable covariance matrices when p>np>n but log(p)=o(n)\log(p)=o(n). Motivated by the deep connection between Stein's unbiased risk estimation (SURE) and AIC in regression models, we propose a family of generalized SURE (SUREc\text{SURE}_c) indexed by cc for covariance matrix estimation, where cc is some constant. When cc is 2, SURE2\text{SURE}_2 provides an unbiased estimator of the Frobenious risk of the covariance matrix estimator. Furthermore, we show that by minimizing SURE2\text{SURE}_2 over all possible banding covariance matrix estimators we attain the minimax optimal rate of convergence and the resulting estimator behaves like the covariance matrix estimator obtained by the so-called oracle tuning. On the other hand, we also show that SURE2\text{SURE}_2 is selection inconsistent when the true covariance matrix is exactly banded. To fix the selection inconsistency, we consider using SURE with c=log(n)c=\log(n) and prove that by minimizing SURElog(n)\text{SURE}_{\log(n)} we select the true bandwith with probability tending to one. Therefore, our analysis indicates that SURE2\text{SURE}_2 and SURElog(n)\text{SURE}_{\log(n)} can be regarded as the AIC and BIC for large covariance matrix estimation, respectively.

View on arXiv
Comments on this paper