Cross-lingual transfer learning with data selection for large-scale spoken language understanding

16Citations
Citations of this article
83Readers
Mendeley users who have this article in their library.

Abstract

A typical cross-lingual transfer learning approach boosting model performance on a resource-poor language is to pre-train the model on all available supervised data from another resource-rich language. However, in large-scale systems, this leads to high training times and computational requirements. In addition, characteristic differences between the source and target languages raise a natural question of whether source-language data selection can improve the knowledge transfer. In this paper, we address this question and propose a simple but effective language model based source-language data selection method for cross-lingual transfer learning in large-scale spoken language understanding. The experimental results show that with data selection i) the source data amount and hence training speed is reduced significantly and ii) model performance is improved.

Cite

CITATION STYLE

APA

Do, Q., & Gaspers, J. (2019). Cross-lingual transfer learning with data selection for large-scale spoken language understanding. In EMNLP-IJCNLP 2019 - 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing, Proceedings of the Conference (pp. 1455–1460). Association for Computational Linguistics. https://doi.org/10.18653/v1/d19-1153

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free