31
4

Particle-MALA and Particle-mGRAD: Gradient-based MCMC methods for high-dimensional state-space models

Adrien Corenflos
Axel Finke
Abstract

State-of-the-art methods for Bayesian inference in state-space models are (a) conditional sequential Monte Carlo (CSMC) algorithms; (b) sophisticated 'classical' MCMC algorithms like MALA, or mGRAD from Titsias and Papaspiliopoulos (2018, arXiv:1610.09641v3 [stat.ML]). The former propose NN particles at each time step to exploit the model's 'decorrelation-over-time' property and thus scale favourably with the time horizon, TT , but break down if the dimension of the latent states, DD, is large. The latter leverage gradient-/prior-informed local proposals to scale favourably with DD but exhibit sub-optimal scalability with TT due to a lack of model-structure exploitation. We introduce methods which combine the strengths of both approaches. The first, Particle-MALA, spreads NN particles locally around the current state using gradient information, thus extending MALA to T>1T > 1 time steps and N>1N > 1 proposals. The second, Particle-mGRAD, additionally incorporates (conditionally) Gaussian prior dynamics into the proposal, thus extending the mGRAD algorithm to T>1T > 1 time steps and N>1N > 1 proposals. We prove that Particle-mGRAD interpolates between CSMC and Particle-MALA, resolving the 'tuning problem' of choosing between CSMC (superior for highly informative prior dynamics) and Particle-MALA (superior for weakly informative prior dynamics). We similarly extend other 'classical' MCMC approaches like auxiliary MALA, aGRAD, and preconditioned Crank-Nicolson-Langevin (PCNL) to T>1T > 1 time steps and N>1N > 1 proposals. In experiments, for both highly and weakly informative prior dynamics, our methods substantially improve upon both CSMC and sophisticated 'classical' MCMC approaches.

View on arXiv
Comments on this paper