Transformer-based Conditional Variational Autoencoder for Controllable Story Generation

01/04/2021
by   Le Fang, et al.
7

We investigate large-scale latent variable models (LVMs) for neural story generation – an under-explored application for open-domain long text – with objectives in two threads: generation effectiveness and controllability. LVMs, especially the variational autoencoder (VAE), have achieved both effective and controllable generation through exploiting flexible distributional latent representations. Recently, Transformers and its variants have achieved remarkable effectiveness without explicit latent representation learning, thus lack satisfying controllability in generation. In this paper, we advocate to revive latent variable modeling, essentially the power of representation learning, in the era of Transformers to enhance controllability without hurting state-of-the-art generation effectiveness. Specifically, we integrate latent representation vectors with a Transformer-based pre-trained architecture to build conditional variational autoencoder (CVAE). Model components such as encoder, decoder and the variational posterior are all built on top of pre-trained language models – GPT2 specifically in this paper. Experiments demonstrate state-of-the-art conditional generation ability of our model, as well as its excellent representation learning capability and controllability.

READ FULL TEXT

page 11

page 12

page 14

page 15

page 16

page 17

research
04/05/2020

Optimus: Organizing Sentences via Pre-trained Modeling of a Latent Space

When trained effectively, the Variational Autoencoder (VAE) can be both ...
research
07/27/2022

A Variational AutoEncoder for Transformers with Nonparametric Variational Information Bottleneck

We propose a VAE for Transformers by developing a variational informatio...
research
11/15/2021

Exploring Story Generation with Multi-task Objectives in Variational Autoencoders

GPT-2 has been frequently adapted in story generation models as it provi...
research
09/14/2021

A Temporal Variational Model for Story Generation

Recent language models can generate interesting and grammatically correc...
research
09/02/2019

A Surprisingly Effective Fix for Deep Latent Variable Modeling of Text

When trained effectively, the Variational Autoencoder (VAE) is both a po...
research
09/06/2023

Self-Supervised Disentanglement of Harmonic and Rhythmic Features in Music Audio Signals

The aim of latent variable disentanglement is to infer the multiple info...
research
07/03/2023

VOLTA: Diverse and Controllable Question-Answer Pair Generation with Variational Mutual Information Maximizing Autoencoder

Previous question-answer pair generation methods aimed to produce fluent...

Please sign up or login with your details

Forgot password? Click here to reset