Escaping the sentence-level paradigm in machine translation

04/25/2023
by   Matt Post, et al.
0

It is well-known that document context is vital for resolving a range of translation ambiguities, and in fact the document setting is the most natural setting for nearly all translation. It is therefore unfortunate that machine translation – both research and production – largely remains stuck in a decades-old sentence-level translation paradigm. It is also an increasingly glaring problem in light of competitive pressure from large language models, which are natively document-based. Much work in document-context machine translation exists, but for various reasons has been unable to catch hold. This paper suggests a path out of this rut by addressing three impediments at once: what architectures should we use? where do we get document-level information for training them? and how do we know whether they are any good? In contrast to work on specialized architectures, we show that the standard Transformer architecture is sufficient, provided it has enough capacity. Next, we address the training data issue by taking document samples from back-translated data only, where the data is not only more readily available, but is also of higher quality compared to parallel document data, which may contain machine translation output. Finally, we propose generative variants of existing contrastive metrics that are better able to discriminate among document systems. Results in four large-data language pairs (DE→EN, EN→DE, EN→FR, and EN→RU) establish the success of these three pieces together in improving document-level performance.

READ FULL TEXT
research
04/30/2020

Addressing Zero-Resource Domains Using Document-Level Context in Neural Machine Translation

Achieving satisfying performance in machine translation on domains for w...
research
10/31/2019

Document-level Neural Machine Translation with Inter-Sentence Attention

Standard neural machine translation (NMT) is on the assumption of docume...
research
06/08/2023

On Search Strategies for Document-Level Neural Machine Translation

Compared to sentence-level systems, document-level neural machine transl...
research
07/14/2019

Microsoft Translator at WMT 2019: Towards Large-Scale Document-Level Neural Machine Translation

This paper describes the Microsoft Translator submissions to the WMT19 n...
research
04/06/2023

Large language models effectively leverage document-level context for literary translation, but critical errors persist

Large language models (LLMs) are competitive with the state of the art o...
research
10/01/2019

Putting Machine Translation in Context with the Noisy Channel Model

We show that Bayes' rule provides a compelling mechanism for controlling...
research
04/21/2021

On User Interfaces for Large-Scale Document-Level Human Evaluation of Machine Translation Outputs

Recent studies emphasize the need of document context in human evaluatio...

Please sign up or login with your details

Forgot password? Click here to reset