Variational Mixture-of-Experts Autoencoders for Multi-Modal Deep Generative Models

11/08/2019
by   Yuge Shi, et al.
28

Learning generative models that span multiple data modalities, such as vision and language, is often motivated by the desire to learn more useful, generalisable representations that faithfully capture common underlying factors between the modalities. In this work, we characterise successful learning of such models as the fulfillment of four criteria: i) implicit latent decomposition into shared and private subspaces, ii) coherent joint generation over all modalities, iii) coherent cross-generation across individual modalities, and iv) improved model learning for individual modalities through multi-modal integration. Here, we propose a mixture-of-experts multimodal variational autoencoder (MMVAE) to learn generative models on different sets of modalities, including a challenging image-language dataset, and demonstrate its ability to satisfy all four criteria, both qualitatively and quantitatively.

READ FULL TEXT

page 6

page 8

page 13

page 15

page 17

page 18

research
07/05/2022

A survey of multimodal deep generative models

Multimodal learning is a framework for building models that make predict...
research
12/23/2020

Private-Shared Disentangled Multimodal VAE for Learning of Hybrid Latent Representations

Multi-modal generative models represent an important family of deep mode...
research
02/14/2018

Multimodal Generative Models for Scalable Weakly-Supervised Learning

Multiple modalities often co-occur when describing natural phenomena. Le...
research
01/18/2021

Multimodal Variational Autoencoders for Semi-Supervised Learning: In Defense of Product-of-Experts

Multimodal generative models should be able to learn a meaningful latent...
research
09/01/2023

Learning multi-modal generative models with permutation-invariant encoders and tighter variational bounds

Devising deep latent variable models for multi-modal data has been a lon...
research
10/09/2021

Discriminative Multimodal Learning via Conditional Priors in Generative Models

Deep generative models with latent variables have been used lately to le...
research
11/07/2016

Joint Multimodal Learning with Deep Generative Models

We investigate deep generative models that can exchange multiple modalit...

Please sign up or login with your details

Forgot password? Click here to reset