Improving Machine Translation Formality Control with Weakly-Labelled Data Augmentation and Post Editing Strategies

3Citations
Citations of this article
43Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper describes Amazon Alexa AI’s implementation for the IWSLT 2022 shared task on formality control. We focus on the unconstrained and supervised task for en→hi (Hindi) and en→ja (Japanese) pairs where very limited formality annotated data is available. We propose three simple yet effective post editing strategies namely, T-V conversion, utilizing a verb conjugator and seq2seq models in order to rewrite the translated phrases into formal or informal language. Considering nuances for formality and informality in different languages, our analysis shows that a language-specific post editing strategy achieves the best performance. To address the unique challenge of limited formality annotations, we further develop a formality classifier to perform weakly-labelled data augmentation which automatically generates synthetic formality labels from large parallel corpus. Empirical results on the IWSLT formality testset have shown that proposed system achieved significant improvements in terms of formality accuracy while retaining BLEU score on-par with baseline.

Cite

CITATION STYLE

APA

Zhang, D., Yu, J., Verma, P., Ganesan, A., & Campbell, S. (2022). Improving Machine Translation Formality Control with Weakly-Labelled Data Augmentation and Post Editing Strategies. In IWSLT 2022 - 19th International Conference on Spoken Language Translation, Proceedings of the Conference (pp. 351–360). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.iwslt-1.32

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free