NVIDIA NeMo Offline Speech Translation Systems for IWSLT 2023

3Citations
Citations of this article
18Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper provides an overview of NVIDIA NeMo’s speech translation systems for the IWSLT 2023 Offline Speech Translation Task. This year, we focused on end-to-end system which capitalizes on pre-trained models and synthetic data to mitigate the problem of direct speech translation data scarcity. When trained on IWSLT 2022 constrained data, our best En→De end-to-end model achieves the average score of 31 BLEU on 7 test sets from IWSLT 2010-2020 which improves over our last year cascade (28.4) and end-to-end (25.7) submissions. When trained on IWSLT 2023 constrained data, the average score drops to 29.5 BLEU.

Cite

CITATION STYLE

APA

Hrinchuk, O., Bataev, V., Bakhturina, E., & Ginsburg, B. (2023). NVIDIA NeMo Offline Speech Translation Systems for IWSLT 2023. In 20th International Conference on Spoken Language Translation, IWSLT 2023 - Proceedings of the Conference (pp. 442–448). Association for Computational Linguistics. https://doi.org/10.18653/v1/2023.iwslt-1.42

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free