Recent studies argue that knowledge distillation is promising for speech translation (ST) using end-to-end models. In this work, we investigate the effect of knowledge distillation with a cascade ST using automatic speech recognition (ASR) and machine translation (MT) models. We distill knowledge from a teacher model based on human transcripts to a student model based on erroneous transcriptions. Our experimental results demonstrated that knowledge distillation is beneficial for a cascade ST. Further investigation that combined knowledge distillation and fine-tuning revealed that the combination consistently improved two language pairs: English-Italian and Spanish-English.
CITATION STYLE
Fukuda, R., Sudoh, K., & Nakamura, S. (2021). On Knowledge Distillation for Translating Erroneous Speech Transcriptions. In IWSLT 2021 - 18th International Conference on Spoken Language Translation, Proceedings (pp. 198–205). Association for Computational Linguistics (ACL). https://doi.org/10.5715/jnlp.29.344
Mendeley helps you to discover research relevant for your work.