This paper presents a cross-lingual projection technique for training class-based language models. We borrow from previous success in projecting POS tags and NER mentions to that of a trained classbased language model. We use a CRF to train a model to predict when a sequence of words is a member of a given class and use this to label our language model training data. We show that we can successfully project the contextual cues for these classes across pairs of languages and retain a high quality class model in languages with no supervised class data. We present empirical results that show the quality of the projected models as well as their effect on the down-stream speech recognition objective. We are able to achieve over 70% of the WER reduction when using the projected class models as compared to models trained on human annotations.
CITATION STYLE
Gfeller, B., Schogol, V., & Hall, K. (2016). Cross-lingual projection for class-based language models. In 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016 - Short Papers (pp. 83–88). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p16-2014
Mendeley helps you to discover research relevant for your work.