On Knowledge Distillation for Translating Erroneous Speech Transcriptions

0Citations
Citations of this article
40Readers
Mendeley users who have this article in their library.

Abstract

Recent studies argue that knowledge distillation is promising for speech translation (ST) using end-to-end models. In this work, we investigate the effect of knowledge distillation with a cascade ST using automatic speech recognition (ASR) and machine translation (MT) models. We distill knowledge from a teacher model based on human transcripts to a student model based on erroneous transcriptions. Our experimental results demonstrated that knowledge distillation is beneficial for a cascade ST. Further investigation that combined knowledge distillation and fine-tuning revealed that the combination consistently improved two language pairs: English-Italian and Spanish-English.

Cite

CITATION STYLE

APA

Fukuda, R., Sudoh, K., & Nakamura, S. (2021). On Knowledge Distillation for Translating Erroneous Speech Transcriptions. In IWSLT 2021 - 18th International Conference on Spoken Language Translation, Proceedings (pp. 198–205). Association for Computational Linguistics (ACL). https://doi.org/10.5715/jnlp.29.344

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free