Context-aware Decoder for Neural Machine Translation using a Target-side Document-Level Language Model

7Citations
Citations of this article
65Readers
Mendeley users who have this article in their library.

Abstract

Although many end-to-end context-aware neural machine translation models have been proposed to incorporate inter-sentential contexts in translation, these models can be trained only in domains where parallel documents with sentential alignments exist. We therefore present a simple method to perform context-aware decoding with any pre-trained sentence-level translation model by using a document-level language model. Our context-aware decoder is built upon sentence-level parallel data and target-side document-level monolingual data. From a theoretical viewpoint, our core contribution is the novel representation of contextual information using point-wise mutual information between context and the current sentence. We demonstrate the effectiveness of our method on English to Russian translation, by evaluating with BLEU and contrastive tests for context-aware translation.

Cite

CITATION STYLE

APA

Sugiyama, A., & Yoshinaga, N. (2021). Context-aware Decoder for Neural Machine Translation using a Target-side Document-Level Language Model. In NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 5781–5791). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.naacl-main.461

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free