End-to-end speech translation with adversarial training

0Citations
Citations of this article
59Readers
Mendeley users who have this article in their library.

Abstract

End-to-end speech translation usually leverages audio-to-text parallel data to train an available speech translation model which has shown impressive results on various speech translation tasks. Due to the artificial cost of collecting audio-to-text parallel data, the speech translation is a natural low-resource translation scenario, which greatly hinders its improvement. In this paper, we proposed a new adversarial training method to leverage target monolingual data to relieve the lowresource shortcoming of speech translation. In our method, the existing speech translation model is considered as a Generator to gain a target language output, and another neural Discriminator is used to guide the distinction between outputs of speech translation model and true target monolingual sentences. Experimental results on the CCMT 2019-BSTC dataset speech translation task demonstrate that the proposed methods can significantly improve the performance of the end-to-end speech translation.

Cite

CITATION STYLE

APA

Li, X., Chen, K., Zhao, T., & Yang, M. (2020). End-to-end speech translation with adversarial training. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 10–14). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.autosimtrans-1.2

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free