28
75

Scaling Gaussian Process Regression with Derivatives

Abstract

Gaussian processes (GPs) with derivatives are useful in many applications, including Bayesian optimization, implicit surface reconstruction, and terrain reconstruction. Fitting a GP to function values and derivatives at nn points in dd dimensions requires linear solves and log determinants with an n(d+1)×n(d+1){n(d+1) \times n(d+1)} positive definite matrix -- leading to prohibitive O(n3d3)\mathcal{O}(n^3d^3) computations for standard direct methods. We propose iterative solvers using fast O(nd)\mathcal{O}(nd) matrix-vector multiplications (MVMs), together with pivoted Cholesky preconditioning that cuts the iterations to convergence by several orders of magnitude, allowing for fast kernel learning and prediction. Our approaches, together with dimensionality reduction, enables Bayesian optimization with derivatives to scale to high-dimensional problems and large evaluation budgets.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.