Alleviating Exposure Bias via Contrastive Learning for Abstractive Text Summarization

08/26/2021
by   Shichao Sun, et al.
0

Encoder-decoder models have achieved remarkable success in abstractive text summarization, which aims to compress one or more documents into a shorter version without the loss of the essential content. Unfortunately, these models mostly suffer a discrepancy between training and inference, i.e., the exposure bias problem. During the training stage, with teacher forcing these models are optimized to maximize the likelihood of the gold summary given the gold summary tokens as input to the decoder, while at inference the given tokens are replaced by the generated tokens. Consequently, low-quality summaries are very likely to be generated. To remedy this problem, we propose to leverage contrastive learning to decrease the likelihood of these low-quality summaries, and meanwhile increase the likelihood of the gold summary. Since our solution expands the states that the model perceives during training, we expect that the exposure bias problem can be alleviated. We experimentally demonstrate that our method effectively improves the performance of the state-of-the-art model on different datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/03/2022

Latent Prompt Tuning for Text Summarization

Prompts with different control signals (e.g., length, keywords, etc.) ca...
research
09/08/2021

Sequence Level Contrastive Learning for Text Summarization

Contrastive learning models have achieved great success in unsupervised ...
research
05/23/2023

On Learning to Summarize with Large Language Models as References

Recent studies have found that summaries generated by large language mod...
research
09/16/2020

Text Generation by Learning from Off-Policy Demonstrations

Current approaches to text generation largely rely on autoregressive mod...
research
02/16/2023

Learning with Rejection for Abstractive Text Summarization

State-of-the-art abstractive summarization systems frequently hallucinat...
research
05/11/2017

A Deep Reinforced Model for Abstractive Summarization

Attentional, RNN-based encoder-decoder models for abstractive summarizat...
research
10/07/2020

TeaForN: Teacher-Forcing with N-grams

Sequence generation models trained with teacher-forcing suffer from issu...

Please sign up or login with your details

Forgot password? Click here to reset