VAE-KRnet and Its Applications to Variational Bayes

VAE-KRnet and Its Applications to Variational Bayes

Year:    2022

Author:    Xiaoliang Wan, Shuangqing Wei

Communications in Computational Physics, Vol. 31 (2022), Iss. 4 : pp. 1049–1082

Abstract

In this work, we have proposed a generative model, called VAE-KRnet, for density estimation or approximation, which combines the canonical variational autoencoder (VAE) with our recently developed flow-based generative model, called KRnet. VAE is used as a dimension reduction technique to capture the latent space, and KRnet is used to model the distribution of the latent variable. Using a linear model between the data and the latent variable, we show that VAE-KRnet can be more effective and robust than the canonical VAE. VAE-KRnet can be used as a density model to approximate either data distribution or an arbitrary probability density function (PDF) known up to a constant. VAE-KRnet is flexible in terms of dimensionality. When the number of dimensions is relatively small, KRnet can effectively approximate the distribution in terms of the original random variable. For high-dimensional cases, we may use VAE-KRnet to incorporate dimension reduction. One important application of VAE-KRnet is the variational Bayes for the approximation of the posterior distribution. The variational Bayes approaches are usually based on the minimization of the Kullback-Leibler (KL) divergence between the model and the posterior. For high-dimensional distributions, it is very challenging to construct an accurate density model due to the curse of dimensionality, where extra assumptions are often introduced for efficiency. For instance, the classical mean-field approach assumes mutual independence between dimensions, which often yields an underestimated variance due to oversimplification. To alleviate this issue, we include into the loss the maximization of the mutual information between the latent random variable and the original random variable, which helps keep more information from the region of low density such that the estimation of variance is improved. Numerical experiments have been presented to demonstrate the effectiveness of our model.

You do not have full access to this article.

Already a Subscriber? Sign in as an individual or via your institution

Journal Article Details

Publisher Name:    Global Science Press

Language:    English

DOI:    https://doi.org/10.4208/cicp.OA-2021-0087

Communications in Computational Physics, Vol. 31 (2022), Iss. 4 : pp. 1049–1082

Published online:    2022-01

AMS Subject Headings:    Global Science Press

Copyright:    COPYRIGHT: © Global Science Press

Pages:    34

Keywords:    Deep learning variational Bayes uncertainty quantification Bayesian inverse problems generative modeling.

Author Details

Xiaoliang Wan

Shuangqing Wei

  1. On computing the hyperparameter of extreme learning machines: Algorithm and application to computational PDEs, and comparison with classical and high-order finite elements

    Dong, Suchuan | Yang, Jielin

    Journal of Computational Physics, Vol. 463 (2022), Iss. P.111290

    https://doi.org/10.1016/j.jcp.2022.111290 [Citations: 24]
  2. Physics-informed neural networks for approximating dynamic (hyperbolic) PDEs of second order in time: Error analysis and algorithms

    Qian, Yanxia | Zhang, Yongchao | Huang, Yunqing | Dong, Suchuan

    Journal of Computational Physics, Vol. 495 (2023), Iss. P.112527

    https://doi.org/10.1016/j.jcp.2023.112527 [Citations: 4]
  3. Numerical Computation of Partial Differential Equations by Hidden-Layer Concatenated Extreme Learning Machine

    Ni, Naxian | Dong, Suchuan

    Journal of Scientific Computing, Vol. 95 (2023), Iss. 2

    https://doi.org/10.1007/s10915-023-02162-0 [Citations: 8]