A Random Matrix--Theoretic Approach to Handling Singular Covariance Estimates

In many practical situations we would like to estimate the covariance matrix of a set of variables from an insufficient amount of data. More specifically, if we have a set of independent, identically distributed measurements of an dimensional random vector the maximum likelihood estimate is the sample covariance matrix. Here we consider the case where such that this estimate is singular and therefore fundamentally bad. We present a radically new approach to deal with this situation. Let be the data matrix, where the columns are the independent realizations of the random vector with covariance matrix . Without loss of generality, we can assume that the random variables have zero mean. We would like to estimate from . Let be the classical sample covariance matrix. Fix a parameter and consider an ensemble of random unitary matrices, , having Haar probability measure. Pre and post multiply by , and by the conjugate transpose of respectively, to produce a non--singular reduced dimension covariance estimate. A new estimate for , denoted by , is obtained by a) projecting the reduced covariance estimate out (to ) through pre and post multiplication by the conjugate transpose of , and by respectively, and b) taking the expectation over the unitary ensemble. Another new estimate (this time for ), , is obtained by a) inverting the reduced covariance estimate, b) projecting the inverse out (to ) through pre and post multiplication by the conjugate transpose of , and by respectively, and c) taking the expectation over the unitary ensemble. We have a closed analytical expression for and in terms of its eigenvalue decomposition.
View on arXiv