Accelerating natural language understanding in task-oriented dialog

1Citations
Citations of this article
79Readers
Mendeley users who have this article in their library.

Abstract

Task-oriented dialog models typically leverage complex neural architectures and large-scale, pre-trained Transformers to achieve state-of-the-art performance on popular natural language understanding benchmarks. However, these models frequently have in excess of tens of millions of parameters, making them impossible to deploy on-device where resource-efficiency is a major concern. In this work, we show that a simple convolutional model compressed with structured pruning achieves largely comparable results to BERT (Devlin et al., 2019) on ATIS and Snips, with under 100K parameters. Moreover, we perform acceleration experiments on CPUs, where we observe our multi-task model predicts intents and slots nearly 63× faster than even DistilBERT (Sanh et al., 2019).

Cite

CITATION STYLE

APA

Ahuja, O., & Desai, S. (2020). Accelerating natural language understanding in task-oriented dialog. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 46–53). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.nlp4convai-1.6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free