Effort-aware neural automatic post-editing

4Citations
Citations of this article
78Readers
Mendeley users who have this article in their library.
Get full text

Abstract

For this round of the WMT 2019 APE shared task, our submission focuses on addressing the “over-correction” problem in APE. Over-correction occurs when the APE system tends to rephrase an already correct MT output, and the resulting sentence is penalized by a reference-based evaluation against human post-edits. Our intuition is that this problem can be prevented by informing the system about the predicted quality of the MT output or, in other terms, the expected amount of needed corrections. For this purpose, following the common approach in multilingual NMT, we prepend a special token to the beginning of both the source text and the MT output indicating the required amount of post-editing. Following the best submissions to the WMT 2018 APE shared task, our backbone architecture is based on multi-source Transformer to encode both the MT output and the corresponding source text. We participated both in the English-German and English-Russian subtasks. In the first subtask, our best submission improved the original MT output quality up to +0.98 BLEU and -0.47 TER. In the second subtask, where the higher quality of the MT output increases the risk of over-correction, none of our submitted runs was able to improve the MT output.

Cite

CITATION STYLE

APA

Tebbifakhr, A., Negri, M., & Turchi, M. (2019). Effort-aware neural automatic post-editing. In WMT 2019 - 4th Conference on Machine Translation, Proceedings of the Conference (Vol. 3, pp. 139–144). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w19-5416

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free