Regularizing Neural Machine Translation by Target-bidirectional Agreement

08/13/2018
by   Zhirui Zhang, et al.
0

Although Neural Machine Translation (NMT) has achieved remarkable progress in the past several years, most NMT systems still suffer from a fundamental shortcoming as in other sequence generation tasks: errors made early in generation process are fed as inputs to the model and can be quickly amplified, harming subsequent sequence generation. To address this issue, we propose a novel model regularization method for NMT training, which aims to improve the agreement between translations generated by left-to-right (L2R) and right-to-left (R2L) NMT decoders. This goal is achieved by introducing two Kullback-Leibler divergence regularization terms into the NMT training objective to reduce the mismatch between output probabilities of L2R and R2L models. In addition, we also employ a joint training strategy to allow L2R and R2L models to improve each other in an interactive update process. Experimental results show that our proposed method significantly outperforms state-of-the-art baselines on Chinese-English and English-German translation tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/13/2019

Synchronous Bidirectional Neural Machine Translation

Existing approaches to neural machine translation (NMT) generate the tar...
research
07/18/2019

Forward-Backward Decoding for Regularizing End-to-End TTS

Neural end-to-end TTS can generate very high-quality synthesized speech,...
research
09/19/2017

Dynamic Oracle for Neural Machine Translation in Decoding Phase

The past several years have witnessed the rapid progress of end-to-end N...
research
09/03/2019

Multi-agent Learning for Neural Machine Translation

Conventional Neural Machine Translation (NMT) models benefit from the tr...
research
08/22/2019

Dual Skew Divergence Loss for Neural Machine Translation

For neural sequence model training, maximum likelihood (ML) has been com...
research
12/16/2021

Learning and Analyzing Generation Order for Undirected Sequence Models

Undirected neural sequence models have achieved performance competitive ...
research
10/24/2022

Finding Memo: Extractive Memorization in Constrained Sequence Generation Tasks

Memorization presents a challenge for several constrained Natural Langua...

Please sign up or login with your details

Forgot password? Click here to reset