Towards Modeling the Style of Translators in Neural Machine Translation

14Citations
Citations of this article
74Readers
Mendeley users who have this article in their library.
Get full text

Abstract

One key ingredient of neural machine translation is the use of large datasets from different domains and resources (e.g. Europarl, TED talks). These datasets contain documents translated by professional translators using different but consistent translation styles. Despite that, the model is usually trained in a way that neither explicitly captures the variety of translation styles present in the data nor translates new data in different and controllable styles. In this work, we investigate methods to augment the state-of-the-art Transformer model with translator information that is available in part of the training data. We show that our style-augmented translation models are able to capture the style variations of translators and to generate translations with different styles on new data. Indeed, the generated variations differ significantly, up to +4.5 BLEU score difference. Despite that, human evaluation confirms that the translations are of the same quality.

Cite

CITATION STYLE

APA

Wang, Y., Hoang, C., & Federico, M. (2021). Towards Modeling the Style of Translators in Neural Machine Translation. In NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 1193–1199). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.naacl-main.94

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free