21
0

Further Understanding of a Local Gaussian Process Approximation: Characterising Convergence in the Finite Regime

Abstract

We show that common choices of kernel functions for a highly accurate and massively scalable nearest-neighbour based GP regression model (GPnn: \cite{GPnn}) exhibit gradual convergence to asymptotic behaviour as dataset-size nn increases. For isotropic kernels such as Mat\'{e}rn and squared-exponential, an upper bound on the predictive MSE can be obtained as O(npd)O(n^{-\frac{p}{d}}) for input dimension dd, pp dictated by the kernel (and d>pd>p) and fixed number of nearest-neighbours mm with minimal assumptions on the input distribution. Similar bounds can be found under model misspecification and combined to give overall rates of convergence of both MSE and an important calibration metric. We show that lower bounds on nn can be given in terms of mm, ll, pp, dd, a tolerance ε\varepsilon and a probability δ\delta. When mm is chosen to be O(npp+d)O(n^{\frac{p}{p+d}}) minimax optimal rates of convergence are attained. Finally, we demonstrate empirical performance and show that in many cases convergence occurs faster than the upper bounds given here.

View on arXiv
Comments on this paper