The stochastic partial differential equation approach to Gaussian processes (GPs) represents Mat\érn GP priors in terms of finite element basis functions and Gaussian coefficients with sparse precision matrix. Such representations enhance the scalability of GP regression and classification to datasets of large size by setting and exploiting sparsity. In this paper we reconsider the standard choice through an analysis of the estimation performance. Our theory implies that, under certain smoothness assumptions, one can reduce the computation and memory cost without hindering the estimation accuracy by setting in the large asymptotics. Numerical experiments illustrate the applicability of our theory and the effect of the prior lengthscale in the pre-asymptotic regime.
View on arXiv