Question-Based Salient Span Selection for More Controllable Text Summarization

11/15/2021
by   Daniel Deutsch, et al.
0

In this work, we propose a method for incorporating question-answering (QA) signals into a summarization model. Our method identifies salient noun phrases (NPs) in the input document by automatically generating wh-questions that are answered by the NPs and automatically determining whether those questions are answered in the gold summaries. This QA-based signal is incorporated into a two-stage summarization model which first marks salient NPs in the input document using a classification model, then conditionally generates a summary. Our experiments demonstrate that the models trained using QA-based supervision generate higher-quality summaries than baseline methods of identifying salient spans on benchmark summarization datasets. Further, we show that the content of the generated summaries can be controlled based on which NPs are marked in the input document. Finally, we propose a method of augmenting the training data so the gold summaries are more consistent with the marked input spans used during training and show how this results in models which learn to better exclude unmarked document content.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/04/2022

Towards Improving Faithfulness in Abstractive Summarization

Despite the success achieved in neural abstractive summarization based o...
research
09/14/2023

Investigating Gender Bias in News Summarization

Summarization is an important application of large language models (LLMs...
research
05/25/2021

Focus Attention: Promoting Faithfulness and Diversity in Summarization

Professional summaries are written with document-level information, such...
research
08/31/2018

Bottom-Up Abstractive Summarization

Neural network-based methods for abstractive summarization produce outpu...
research
10/01/2020

Towards Question-Answering as an Automatic Metric for Evaluating the Content Quality of a Summary

Recently, there has been growing interest in using question-answering (Q...
research
03/21/2022

HIBRIDS: Attention with Hierarchical Biases for Structure-aware Long Document Summarization

Document structure is critical for efficient information consumption. Ho...
research
10/10/2021

Enhance Long Text Understanding via Distilled Gist Detector from Abstractive Summarization

Long text understanding is important yet challenging in natural language...

Please sign up or login with your details

Forgot password? Click here to reset