Quantitative Understanding of VAE by Interpreting ELBO as Rate Distortion Cost of Transform Coding

07/30/2020
by   Akira Nakagawa, et al.
0

VAE (Variational autoencoder) estimates the posterior parameters (mean and variance) of latent variables corresponding to each input data. While it is used for many tasks, the transparency of the model is still an underlying issue. This paper provides a quantitative understanding of VAE property by interpreting ELBO maximization as Rate-distortion optimization of transform coding. According to the Rate-distortion theory, the optimal transform coding is achieved by using PCA-like orthonormal (orthogonal and unit norm) transform. From this analogy, we show theoretically and experimentally that VAE can be mapped to an implicit orthonormal transform with a scale factor derived from the posterior parameter. As a result, the quantitative importance of each latent variable can be evaluated like the eigenvalue of PCA. We can also estimate the data probabilities in the input space from the prior, loss metrics, and corresponding posterior parameters.

READ FULL TEXT

page 8

page 23

research
12/06/2021

Conditional Deep Hierarchical Variational Autoencoder for Voice Conversion

Variational autoencoder-based voice conversion (VAE-VC) has the advantag...
research
06/08/2023

Unscented Autoencoder

The Variational Autoencoder (VAE) is a seminal approach in deep generati...
research
12/07/2022

Multi-Rate VAE: Train Once, Get the Full Rate-Distortion Curve

Variational autoencoders (VAEs) are powerful tools for learning latent r...
research
09/14/2023

Dataset Size Dependence of Rate-Distortion Curve and Threshold of Posterior Collapse in Linear VAE

In the Variational Autoencoder (VAE), the variational posterior often al...
research
06/29/2020

VAE-KRnet and its applications to variational Bayes

In this work, we have proposed a generative model for density estimation...
research
02/21/2022

Moment Matching Deep Contrastive Latent Variable Models

In the contrastive analysis (CA) setting, machine learning practitioners...
research
01/23/2021

Improved Training of Sparse Coding Variational Autoencoder via Weight Normalization

Learning a generative model of visual information with sparse and compos...

Please sign up or login with your details

Forgot password? Click here to reset