14
16

Faster Uncertainty Quantification for Inverse Problems with Conditional Normalizing Flows

Abstract

In inverse problems, we often have access to data consisting of paired samples (x,y)pX,Y(x,y)(x,y)\sim p_{X,Y}(x,y) where yy are partial observations of a physical system, and xx represents the unknowns of the problem. Under these circumstances, we can employ supervised training to learn a solution xx and its uncertainty from the observations yy. We refer to this problem as the "supervised" case. However, the data ypY(y)y\sim p_{Y}(y) collected at one point could be distributed differently than observations ypY(y)y'\sim p_{Y}'(y'), relevant for a current set of problems. In the context of Bayesian inference, we propose a two-step scheme, which makes use of normalizing flows and joint data to train a conditional generator qθ(xy)q_{\theta}(x|y) to approximate the target posterior density pXY(xy)p_{X|Y}(x|y). Additionally, this preliminary phase provides a density function qθ(xy)q_{\theta}(x|y), which can be recast as a prior for the "unsupervised" problem, e.g.~when only the observations ypY(y)y'\sim p_{Y}'(y'), a likelihood model yxy'|x, and a prior on xx' are known. We then train another invertible generator with output density qϕ(xy)q'_{\phi}(x|y') specifically for yy', allowing us to sample from the posterior pXY(xy)p_{X|Y}'(x|y'). We present some synthetic results that demonstrate considerable training speedup when reusing the pretrained network qθ(xy)q_{\theta}(x|y') as a warm start or preconditioning for approximating pXY(xy)p_{X|Y}'(x|y'), instead of learning from scratch. This training modality can be interpreted as an instance of transfer learning. This result is particularly relevant for large-scale inverse problems that employ expensive numerical simulations.

View on arXiv
Comments on this paper