Discrete Auto-regressive Variational Attention Models for Text Modeling

06/16/2021
by   Xianghong Fang, et al.
0

Variational autoencoders (VAEs) have been widely applied for text modeling. In practice, however, they are troubled by two challenges: information underrepresentation and posterior collapse. The former arises as only the last hidden state of LSTM encoder is transformed into the latent space, which is generally insufficient to summarize the data. The latter is a long-standing problem during the training of VAEs as the optimization is trapped to a disastrous local optimum. In this paper, we propose Discrete Auto-regressive Variational Attention Model (DAVAM) to address the challenges. Specifically, we introduce an auto-regressive variational attention approach to enrich the latent space by effectively capturing the semantic dependency from the input. We further design discrete latent space for the variational attention and mathematically show that our model is free from posterior collapse. Extensive experiments on language modeling tasks demonstrate the superiority of DAVAM against several VAE counterparts.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/21/2020

Discrete Variational Attention Models for Language Generation

Variational autoencoders have been widely applied for natural language g...
research
04/22/2020

Discretized Bottleneck in VAE: Posterior-Collapse-Free Sequence-to-Sequence Learning

Variational autoencoders (VAEs) are important tools in end-to-end repres...
research
02/20/2023

Analyzing the Posterior Collapse in Hierarchical Variational Autoencoders

Hierarchical Variational Autoencoders (VAEs) are among the most popular ...
research
02/01/2021

Hierarchical Variational Autoencoder for Visual Counterfactuals

Conditional Variational Auto Encoders (VAE) are gathering significant at...
research
05/03/2022

Learning Discrete Structured Variational Auto-Encoder using Natural Evolution Strategies

Discrete variational auto-encoders (VAEs) are able to represent semantic...
research
05/26/2023

Evaluating Open-Domain Dialogues in Latent Space with Next Sentence Prediction and Mutual Information

The long-standing one-to-many issue of the open-domain dialogues poses s...
research
08/10/2021

Arbitrage-Free Implied Volatility Surface Generation with Variational Autoencoders

We propose a hybrid method for generating arbitrage-free implied volatil...

Please sign up or login with your details

Forgot password? Click here to reset