SPMoE: Generate Multiple Pattern-Aware Outputs with Sparse Pattern Mixture of Experts

08/17/2021
by   Shaobo Cui, et al.
0

Many generation tasks follow a one-to-many mapping relationship: each input could be associated with multiple outputs. Existing methods like Conditional Variational AutoEncoder(CVAE) employ a latent variable to model this one-to-many relationship. However, this high-dimensional and dense latent variable lacks explainability and usually leads to poor and uncontrollable generations. In this paper, we innovatively introduce the linguistic concept of pattern to decompose the one-to-many mapping into multiple one-to-one mappings and further propose a model named Sparse Pattern Mixture of Experts(SPMoE). Each one-to-one mapping is associated with a conditional generation pattern and is modeled with an expert in SPMoE. To ensure each language pattern can be exclusively handled with an expert model for better explainability and diversity, a sparse mechanism is employed to coordinate all the expert models in SPMoE. We assess the performance of our SPMoE on the paraphrase generation task and the experiment results prove that SPMoE can achieve a good balance in terms of quality, pattern-level diversity, and corpus-level diversity.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/04/2020

Inferring food intake from multiple biomarkers using a latent variable model

Metabolomic based approaches have gained much attention in recent years ...
research
04/24/2019

Condition-Transforming Variational AutoEncoder for Conversation Response Generation

This paper proposes a new model, called condition-transforming variation...
research
07/07/2020

Benefiting Deep Latent Variable Models via Learning the Prior and Removing Latent Regularization

There exist many forms of deep latent variable models, such as the varia...
research
03/13/2020

MixPoet: Diverse Poetry Generation via Learning Controllable Mixed Latent Space

As an essential step towards computer creativity, automatic poetry gener...
research
02/20/2019

Mixture Models for Diverse Machine Translation: Tricks of the Trade

Mixture models trained via EM are among the simplest, most widely used a...
research
06/16/2021

Latent Mappings: Generating Open-Ended Expressive Mappings Using Variational Autoencoders

In many contexts, creating mappings for gestural interactions can form p...
research
01/20/2022

Lensing Machines: Representing Perspective in Latent Variable Models

Many datasets represent a combination of different ways of looking at th...

Please sign up or login with your details

Forgot password? Click here to reset