HW-TSC’s Participation in the IWSLT 2022 Isometric Spoken Language Translation

4Citations
Citations of this article
33Readers
Mendeley users who have this article in their library.

Abstract

This paper presents our submissions to the IWSLT 2022 Isometric Spoken Language Translation task. We participate in all three language pairs (English-German, English-French, and English-Spanish) under the constrained setting, and submit an English-German result under the unconstrained setting. We use the standard Transformer model as the baseline and obtain the best performance via one of its variants that shares the decoder input and output embedding. We perform detailed pre-processing and filtering on the provided bilingual data. Several strategies are used to train our models, such as Multilingual Translation, Back Translation, Forward Translation, R-Drop, Average Checkpoint, and Ensemble. We experiment on three methods for biasing the output length: i) conditioning the output to a given target-source length-ratio class; ii) enriching the transformer positional embedding with length information and iii) length control decoding for non-autoregressive translation etc. Our submissions achieve 30.7, 41.6 and 36.7 BLEU respectively on the tst-COMMON test sets for English-German, English-French, English-Spanish tasks and 100% comply with the length requirements.

Cite

CITATION STYLE

APA

Li, Z., Guo, J., Wei, D., Shang, H., Wang, M., Zhu, T., … Qin, Y. (2022). HW-TSC’s Participation in the IWSLT 2022 Isometric Spoken Language Translation. In IWSLT 2022 - 19th International Conference on Spoken Language Translation, Proceedings of the Conference (pp. 361–368). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.iwslt-1.33

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free