Abstract
This paper describes HW-TSC’s submissions to the IWSLT 2023 Offline Speech Translation task, including speech translation of talks from English to German, English to Chinese and English to Japanese. We participated in all three tracks (Constrained training, Constrained with Large Language Models training, Unconstrained training), with using cascaded architectures models. We use data enhancement, pretraining models and other means to improve the quality of ASR, and use a variety of techniques including R-Drop, deep model, domain data selection, etc. to improve the quality of NMT. Compared with last year’s best results, we have improved by 2.1 BLEU in the MuST-C English-German test set.
Cite
CITATION STYLE
Li, Z., Wu, Z., Rao, Z., YuHao, X., JiaXin, G., Wei, D., … Yang, H. (2023). HW-TSC at IWSLT2023: Break the Quality Ceiling of Offline Track via Pre-Training and Domain Adaptation. In 20th International Conference on Spoken Language Translation, IWSLT 2023 - Proceedings of the Conference (pp. 187–193). Association for Computational Linguistics. https://doi.org/10.18653/v1/2023.iwslt-1.14
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.