Context-aware neural machine translation with mini-batch embedding

6Citations
Citations of this article
63Readers
Mendeley users who have this article in their library.
Get full text

Abstract

It is crucial to provide an inter-sentence context in Neural Machine Translation (NMT) models for higher-quality translation. With the aim of using a simple approach to incorporate inter-sentence information, we propose mini-batch embedding (MBE) as a way to represent the features of sentences in a mini-batch. We construct a mini-batch by choosing sentences from the same document, and thus the MBE is expected to have contextual information across sentences. Here, we incorporate MBE in an NMT model, and our experiments show that the proposed method consistently outperforms the translation capabilities of strong baselines and improves writing style or terminology to fit the document's context.

Cite

CITATION STYLE

APA

Morishita, M., Suzuki, J., Iwata, T., & Nagata, M. (2021). Context-aware neural machine translation with mini-batch embedding. In EACL 2021 - 16th Conference of the European Chapter of the Association for Computational Linguistics, Proceedings of the Conference (pp. 2513–2521). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.eacl-main.214

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free