Breaking corpus bottleneck for context-aware neural machine translation with cross-task pre-training

6Citations
Citations of this article
66Readers
Mendeley users who have this article in their library.

Abstract

Context-aware neural machine translation (NMT) remains challenging due to the lack of large-scale document-level parallel dataset. To break the corpus bottleneck, in this paper we aim to improve context-aware NMT by taking the advantage of the availability of both large-scale sentence-level parallel dataset and source-side monolingual documents. To this end, we propose two pre-training tasks. One learns to translate a sentence from source language to target language on the sentence-level parallel dataset while the other learns to translate a document from deliberately noised to original on the monolingual documents. Importantly, the two pre-training tasks are jointly and simultaneously learned via the same model, thereafter fine-tuned on scale-limited parallel documents from both sentence-level and document-level perspectives. Experimental results on four translation tasks show that our approach significantly improves translation performance. One nice property of our approach is that the fine-tuned model can be used to translate both sentences and documents.

Cite

CITATION STYLE

APA

Chen, L., Li, J., Gong, Z., Chen, B., Luo, W., Zhang, M., & Zhou, G. (2021). Breaking corpus bottleneck for context-aware neural machine translation with cross-task pre-training. In ACL-IJCNLP 2021 - 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Proceedings of the Conference (pp. 2851–2861). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.acl-long.222

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free