New asymptotic results in principal component analysis

Let be a mean zero Gaussian random vector in a separable Hilbert space with covariance operator Let be the spectral decomposition of with distinct eigenvalues and the corresponding spectral projectors Given a sample of size of i.i.d. copies of the sample covariance operator is defined as The main goal of principal component analysis is to estimate spectral projectors by their empirical counterparts properly defined in terms of spectral decomposition of the sample covariance operator The aim of this paper is to study asymptotic distributions of important statistics related to this problem, in particular, of statistic where is the squared Hilbert--Schmidt norm. This is done in a "high-complexity" asymptotic framework in which the so called effective rank ( being the trace and being the operator norm) of the true covariance is becoming large simultaneously with the sample size but as In this setting, we prove that, in the case of one-dimensional spectral projector the properly centered and normalized statistic with {\it data-dependent} centering and normalization converges in distribution to a Cauchy type limit. The proofs of this and other related results rely on perturbation analysis and Gaussian concentration.
View on arXiv