Neural Language Correction with Character-Based Attention

03/31/2016
by   Ziang Xie, et al.
0

Natural language correction has the potential to help language learners improve their writing skills. While approaches with separate classifiers for different error types have high precision, they do not flexibly handle errors such as redundancy or non-idiomatic phrasing. On the other hand, word and phrase-based machine translation methods are not designed to cope with orthographic errors, and have recently been outpaced by neural models. Motivated by these issues, we present a neural network-based approach to language correction. The core component of our method is an encoder-decoder recurrent neural network with an attention mechanism. By operating at the character level, the network avoids the problem of out-of-vocabulary words. We illustrate the flexibility of our approach on dataset of noisy, user-generated text collected from an English learner forum. When combined with a language model, our method achieves a state-of-the-art F_0.5-score on the CoNLL 2014 Shared Task. We further demonstrate that training the network on additional data with synthesized errors can improve performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/26/2018

A Multilayer Convolutional Encoder-Decoder Neural Network for Grammatical Error Correction

We improve automatic correction of grammatical, orthographic, and colloc...
research
08/16/2016

An Efficient Character-Level Neural Machine Translation

Neural machine translation aims at building a single large neural networ...
research
09/19/2017

Neural Networks for Text Correction and Completion in Keyboard Decoding

Despite the ubiquity of mobile and wearable text messaging applications,...
research
05/21/2017

Spelling Correction as a Foreign Language

In this paper, we reformulated the spell correction problem as a machine...
research
01/08/2019

Choosing the Right Word: Using Bidirectional LSTM Tagger for Writing Support Systems

Scientific writing is difficult. It is even harder for those for whom En...
research
05/28/2018

Graph-based Filtering of Out-of-Vocabulary Words for Encoder-Decoder Models

Encoder-decoder models typically only employ words that are frequently u...
research
01/09/2022

An Ensemble Approach to Acronym Extraction using Transformers

Acronyms are abbreviated units of a phrase constructed by using initial ...

Please sign up or login with your details

Forgot password? Click here to reset