Does Joint Training Really Help Cascaded Speech Translation?

1Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

Currently, in speech translation, the straightforward approach - cascading a recognition system with a translation system - delivers state-of-the-art results. However, fundamental challenges such as error propagation from the automatic speech recognition system still remain. To mitigate these problems, recently, people turn their attention to direct data and propose various joint training methods. In this work, we seek to answer the question of whether joint training really helps cascaded speech translation. We review recent papers on the topic and also investigate a joint training criterion by marginalizing the transcription posterior probabilities. Our findings show that a strong cascaded baseline can diminish any improvements obtained using joint training, and we suggest alternatives to joint training. We hope this work can serve as a refresher of the current speech translation landscape, and motivate research in finding more efficient and creative ways to utilize the direct data for speech translation.

Cite

CITATION STYLE

APA

Tran, V. A. K., Thulke, D., Gao, Y., Herold, C., & Ney, H. (2022). Does Joint Training Really Help Cascaded Speech Translation? In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 4480–4487). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.297

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free