A simple and effective unified encoder for document-level machine translation

74Citations
Citations of this article
130Readers
Mendeley users who have this article in their library.

Abstract

Most of the existing models for document-level machine translation adopt dual-encoder structures. The representation of the source sentences and the document-level contexts are modeled with two separate encoders. Although these models can make use of the document-level contexts, they do not fully model the interaction between the contexts and the source sentences, and can not directly adapt to the recent pre-training models (e.g., BERT) which encodes multiple sentences with a single encoder. In this work, we propose a simple and effective unified encoder that can outperform the baseline models of dual-encoder models in terms of BLEU and METEOR scores. Moreover, the pre-training models can further boost the performance of our proposed model.

Cite

CITATION STYLE

APA

Ma, S., Zhang, D., & Zhou, M. (2020). A simple and effective unified encoder for document-level machine translation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 3505–3511). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-main.321

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free