Bag of experts architectures for model reuse in conversational language understanding

15Citations
Citations of this article
92Readers
Mendeley users who have this article in their library.

Abstract

Slot tagging, the task of detecting entities in input user utterances, is a key component of natural language understanding systems for personal digital assistants. Since each new domain requires a different set of slots, the annotation costs for labeling data for training slot tagging models increases rapidly as the number of domains grow. To tackle this, we describe Bag of Experts (BoE) architectures for model reuse for both LSTM and CRF based models. Extensive experimentation over a dataset of 10 domains drawn from data relevant to our commercial personal digital assistant shows that our BoE models outperform the baseline models with a statistically significant average margin of 5.06% in absolute F1- score when training with 2000 instances per domain, and achieve an even higher improvement of 12.16% when only 25% of the training data is used.

Cite

CITATION STYLE

APA

Jha, R., Marin, A., Shivaprasad, S., & Zitouni, I. (2018). Bag of experts architectures for model reuse in conversational language understanding. In NAACL HLT 2018 - 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Conference (Vol. 3, pp. 153–161). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/n18-3019

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free