A Representation Modeling Based Language GAN with Completely Random Initialization

08/04/2022
by   Da Ren, et al.
0

Text generative models trained via Maximum Likelihood Estimation (MLE) suffer from the notorious exposure bias problem, and Generative Adversarial Networks (GANs) are shown to have potential to tackle it. Existing language GANs adopt estimators like REINFORCE or continuous relaxations to model word distributions. The inherent limitations of such estimators lead current models to rely on pre-training techniques (MLE pre-training or pre-trained embeddings). Representation modeling methods which are free from those limitations, however, are seldom explored because of its poor performance in previous attempts. Our analyses reveal that invalid sampling method and unhealthy gradients are the main contributors to its unsatisfactory performance. In this work, we present two techniques to tackle these problems: dropout sampling and fully normalized LSTM. Based on these two techniques, we propose InitialGAN whose parameters are randomly initialized completely. Besides, we introduce a new evaluation metric, Least Coverage Rate, to better evaluate the quality of generated samples. The experimental results demonstrate that InitialGAN outperforms both MLE and other compared models. To the best of our knowledge, it is the first time a language GAN can outperform MLE without any pre-training techniques.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2020

ColdGANs: Taming Language GANs with Cautious Sampling Strategies

Training regimes based on Maximum Likelihood Estimation (MLE) suffer fro...
research
05/23/2019

Training language GANs from Scratch

Generative Adversarial Networks (GANs) enjoy great success at image gene...
research
04/08/2018

Language Modeling with Generative AdversarialNetworks

Generative Adversarial Networks (GANs) have been promising in the field ...
research
11/06/2018

Language GANs Falling Short

Generating high-quality text with sufficient diversity is essential for ...
research
02/26/2017

Maximum-Likelihood Augmented Discrete Generative Adversarial Networks

Despite the successes in capturing continuous distributions, the applica...
research
10/13/2019

Rethinking Exposure Bias In Language Modeling

Exposure bias describes the phenomenon that a language model trained und...
research
09/01/2021

OptAGAN: Entropy-based finetuning on text VAE-GAN

Transfer learning through large pre-trained models has changed the lands...

Please sign up or login with your details

Forgot password? Click here to reset