An Acceleration Strategy for Randomize-Then-Optimize Sampling via Deep Neural Networks

An Acceleration Strategy for Randomize-Then-Optimize Sampling via Deep Neural Networks

Year:    2021

Author:    Liang Yan, Tao Zhou

Journal of Computational Mathematics, Vol. 39 (2021), Iss. 6 : pp. 848–864

Abstract

Randomize-then-optimize (RTO) is widely used for sampling from posterior distributions in Bayesian inverse problems. However, RTO can be computationally intensive for complexity problems due to repetitive evaluations of the expensive forward model and its gradient. In this work, we present a novel goal-oriented deep neural networks (DNN) surrogate approach to substantially reduce the computation burden of RTO. In particular, we propose to draw the training points for the DNN-surrogate from a local approximated posterior distribution — yielding a flexible and efficient sampling algorithm that converges to the direct RTO approach. We present a Bayesian inverse problem governed by elliptic PDEs to demonstrate the computational accuracy and efficiency of our DNN-RTO approach, which shows that DNN-RTO can significantly outperform the traditional RTO.

You do not have full access to this article.

Already a Subscriber? Sign in as an individual or via your institution

Journal Article Details

Publisher Name:    Global Science Press

Language:    English

DOI:    https://doi.org/10.4208/jcm.2102-m2020-0339

Journal of Computational Mathematics, Vol. 39 (2021), Iss. 6 : pp. 848–864

Published online:    2021-01

AMS Subject Headings:   

Copyright:    COPYRIGHT: © Global Science Press

Pages:    17

Keywords:    Bayesian inverse problems Deep neural network Markov chain Monte Carlo.

Author Details

Liang Yan

Tao Zhou