Lifelong Mixture of Variational Autoencoders

07/09/2021
by   Fei Ye, et al.
6

In this paper, we propose an end-to-end lifelong learning mixture of experts. Each expert is implemented by a Variational Autoencoder (VAE). The experts in the mixture system are jointly trained by maximizing a mixture of individual component evidence lower bounds (MELBO) on the log-likelihood of the given training samples. The mixing coefficients in the mixture, control the contributions of each expert in the goal representation. These are sampled from a Dirichlet distribution whose parameters are determined through non-parametric estimation during lifelong learning. The model can learn new tasks fast when these are similar to those previously learnt. The proposed Lifelong mixture of VAE (L-MVAE) expands its architecture with new components when learning a completely new task. After the training, our model can automatically determine the relevant expert to be used when fed with new data samples. This mechanism benefits both the memory efficiency and the required computational cost as only one expert is used during the inference. The L-MVAE inference model is able to perform interpolation in the joint latent space across the data domains associated with different tasks and is shown to be efficient for disentangled learning representation.

READ FULL TEXT

page 1

page 7

page 9

page 10

research
08/31/2020

LaDDer: Latent Data Distribution Modelling with a Generative Prior

In this paper, we show that the performance of a learnt generative model...
research
10/17/2019

Mixture-of-Experts Variational Autoencoder for clustering and generating from similarity-based representations

Clustering high-dimensional data, such as images or biological measureme...
research
02/15/2018

Quantum Variational Autoencoder

Variational autoencoders (VAEs) are powerful generative models with the ...
research
11/18/2016

Expert Gate: Lifelong Learning with a Network of Experts

In this paper we introduce a model of lifelong learning, based on a Netw...
research
10/07/2020

Learning from demonstration using products of experts: applications to manipulation and task prioritization

Probability distributions are key components of many learning from demon...
research
05/28/2020

Variational Autoencoder with Embedded Student-t Mixture Model for Authorship Attribution

Traditional computational authorship attribution describes a classificat...
research
04/22/2019

PLUME: Polyhedral Learning Using Mixture of Experts

In this paper, we propose a novel mixture of expert architecture for lea...

Please sign up or login with your details

Forgot password? Click here to reset