Cross-lingual projection for class-based language models

3Citations
Citations of this article
90Readers
Mendeley users who have this article in their library.

Abstract

This paper presents a cross-lingual projection technique for training class-based language models. We borrow from previous success in projecting POS tags and NER mentions to that of a trained classbased language model. We use a CRF to train a model to predict when a sequence of words is a member of a given class and use this to label our language model training data. We show that we can successfully project the contextual cues for these classes across pairs of languages and retain a high quality class model in languages with no supervised class data. We present empirical results that show the quality of the projected models as well as their effect on the down-stream speech recognition objective. We are able to achieve over 70% of the WER reduction when using the projected class models as compared to models trained on human annotations.

Cite

CITATION STYLE

APA

Gfeller, B., Schogol, V., & Hall, K. (2016). Cross-lingual projection for class-based language models. In 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016 - Short Papers (pp. 83–88). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p16-2014

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free