ARMAN: Pre-training with Semantically Selecting and Reordering of Sentences for Persian Abstractive Summarization

09/09/2021
by   Alireza Salemi, et al.
0

Abstractive text summarization is one of the areas influenced by the emergence of pre-trained language models. Current pre-training works in abstractive summarization give more points to the summaries with more words in common with the main text and pay less attention to the semantic similarity between generated sentences and the original document. We propose ARMAN, a Transformer-based encoder-decoder model pre-trained with three novel objectives to address this issue. In ARMAN, salient sentences from a document are selected according to a modified semantic score to be masked and form a pseudo summary. To summarize more accurately and similar to human writing patterns, we applied modified sentence reordering. We evaluated our proposed models on six downstream Persian summarization tasks. Experimental results show that our proposed model achieves state-of-the-art performance on all six summarization tasks measured by ROUGE and BERTScore. Our models also outperform prior works in textual entailment, question paraphrasing, and multiple choice question answering. Finally, we established a human evaluation and show that using the semantic score significantly improves summarization results.

READ FULL TEXT
research
12/18/2019

PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization

Recent work pre-training Transformers with self-supervised objectives on...
research
02/18/2020

Learning by Semantic Similarity Makes Abstractive Summarization Better

One of the obstacles of abstractive summarization is the presence of var...
research
10/16/2020

Unsupervised Extractive Summarization by Pre-training Hierarchical Transformers

Unsupervised extractive document summarization aims to select important ...
research
11/16/2020

A Two-Phase Approach for Abstractive Podcast Summarization

Podcast summarization is different from summarization of other data form...
research
10/12/2021

HETFORMER: Heterogeneous Transformer with Sparse Attention for Long-Text Extractive Summarization

To capture the semantic graph structure from raw text, most existing sum...
research
02/22/2022

Learning Cluster Patterns for Abstractive Summarization

Nowadays, pre-trained sequence-to-sequence models such as BERTSUM and BA...
research
05/11/2018

Using Stastical and Semantic Models for Multi-Document Summarization

We report on series of experiments with different semantic models on top...

Please sign up or login with your details

Forgot password? Click here to reset