Felix: Flexible Text Editing Through Tagging and Insertion

03/24/2020
by   Jonathan Mallinson, et al.
0

We present Felix — a flexible text-editing approach for generation, designed to derive the maximum benefit from the ideas of decoding with bi-directional contexts and self-supervised pre-training. In contrast to conventional sequence-to-sequence (seq2seq) models, Felix is efficient in low-resource settings and fast at inference time, while being capable of modeling flexible input-output transformations. We achieve this by decomposing the text-editing task into two sub-tasks: tagging to decide on the subset of input tokens and their order in the output text and insertion to in-fill the missing tokens in the output not present in the input. The tagging model employs a novel Pointer mechanism, while the insertion model is based on a Masked Language Model. Both of these models are chosen to be non-autoregressive to guarantee faster inference. Felix performs favourably when compared to recent text-editing methods and strong seq2seq baselines when evaluated on four NLG tasks: Sentence Fusion, Machine Translation Automatic Post-Editing, Summarization, and Text Simplification.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2022

EdiT5: Semi-Autoregressive Text-Editing with T5 Warm-Start

We present EdiT5 - a novel semi-autoregressive text-editing approach des...
research
03/17/2022

An Imitation Learning Curriculum for Text Editing with Non-Autoregressive Models

We propose a framework for training non-autoregressive sequence-to-seque...
research
06/08/2020

Copy that! Editing Sequences by Copying Spans

Neural sequence-to-sequence models are finding increasing use in editing...
research
03/08/2021

Text Simplification by Tagging

Edit-based approaches have recently shown promising results on multiple ...
research
09/03/2019

Encode, Tag, Realize: High-Precision Text Editing

We propose LaserTagger - a sequence tagging approach that casts text gen...
research
05/27/2019

Levenshtein Transformer

Modern neural sequence generation models are built to either generate to...
research
09/19/2023

RedPenNet for Grammatical Error Correction: Outputs to Tokens, Attentions to Spans

The text editing tasks, including sentence fusion, sentence splitting an...

Please sign up or login with your details

Forgot password? Click here to reset