Findings of the WMT 2019 shared task on automatic post-editing

48Citations
Citations of this article
62Readers
Mendeley users who have this article in their library.

Abstract

We present the results from the 5th round of the WMT task on MT Automatic Post-Editing. The task consists in automatically correcting the output of a “black-box” machine translation system by learning from human corrections. Keeping the same general evaluation setting of the previous four rounds, this year we focused on two language pairs (English-German and English-Russian) and on domain-specific data (Information Technology). For both the language directions, MT outputs were produced by neural systems unknown to participants. Seven teams participated in the English-German task, with a total of 18 submitted runs. The evaluation, which was performed on the same test set used for the 2018 round, shows further progress in APE technology: 4 teams achieved better results than last year’s winning system, with improvements up to -0.78 TER and +1.23 BLEU points over the baseline. Two teams participated in the English-Russian task submitting 2 runs each. On this new language direction, characterized by a higher quality of the original translations, the task proved to be particularly challenging. Indeed, none of the submitted runs improved the very high results of the strong system used to produce the initial translations (16.16 TER, 76.20 BLEU).

Cite

CITATION STYLE

APA

Chatterjee, R., Federmann, C., Negri, M., & Turchi, M. (2019). Findings of the WMT 2019 shared task on automatic post-editing. In WMT 2019 - 4th Conference on Machine Translation, Proceedings of the Conference (Vol. 3, pp. 11–28). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w18-6452

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free