Learning variational autoencoders via MCMC speed measures

08/26/2023
by   Marcel Hirt, et al.
0

Variational autoencoders (VAEs) are popular likelihood-based generative models which can be efficiently trained by maximizing an Evidence Lower Bound (ELBO). There has been much progress in improving the expressiveness of the variational distribution to obtain tighter variational bounds and increased generative performance. Whilst previous work has leveraged Markov chain Monte Carlo (MCMC) methods for the construction of variational densities, gradient-based methods for adapting the proposal distributions for deep latent variable models have received less attention. This work suggests an entropy-based adaptation for a short-run Metropolis-adjusted Langevin (MALA) or Hamiltonian Monte Carlo (HMC) chain while optimising a tighter variational bound to the log-evidence. Experiments show that this approach yields higher held-out log-likelihoods as well as improved generative metrics. Our implicit variational density can adapt to complicated posterior geometries of latent hierarchical representations arising in hierarchical VAEs.

READ FULL TEXT

page 8

page 9

research
11/04/2019

Gradient-based Adaptive Markov Chain Monte Carlo

We introduce a gradient-based learning method to automatically adapt Mar...
research
08/04/2017

Learning Model Reparametrizations: Implicit Variational Inference by Fitting MCMC distributions

We introduce a new algorithm for approximate inference that combines rep...
research
09/15/2022

Langevin Autoencoders for Learning Deep Latent Variable Models

Markov chain Monte Carlo (MCMC), such as Langevin dynamics, is valid for...
research
05/20/2018

Conditional Inference in Pre-trained Variational Autoencoders via Cross-coding

Variational Autoencoders (VAEs) are a popular generative model, but one ...
research
09/02/2020

Quasi-symplectic Langevin Variational Autoencoder

Variational autoencoder (VAE) as one of the well investigated generative...
research
05/30/2022

Mitigating Out-of-Distribution Data Density Overestimation in Energy-Based Models

Deep energy-based models (EBMs), which use deep neural networks (DNNs) a...
research
10/27/2021

Entropy-based adaptive Hamiltonian Monte Carlo

Hamiltonian Monte Carlo (HMC) is a popular Markov Chain Monte Carlo (MCM...

Please sign up or login with your details

Forgot password? Click here to reset