Accelerating Multiple Intent Detection and Slot Filling via Targeted Knowledge Distillation

18Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

Abstract

Recent non-autoregressive Spoken Language Understanding (SLU) models attracts increasing attention owing to the high inference speed. However, most of them still (1) suffer from the multi-modality problem since the prior knowledge about the reference is relatively poor during inference; (2) fail to achieve a satisfactory inference speed limited by their complex frameworks. To tackle these problems, in this paper, we propose a Targeted Knowledge Distillation Framework (TKDF), which applies knowledge distillation to improve the performance. Specifically, we first train an SLU model as a teacher model, which has higher accuracy while slower inference speed. Then we introduce an evaluator and utilize the curriculum learning strategy to select proper targets for the student model. Experiment results on two public multi-intent SLU datasets demonstrate that our method can realize a flexible trade-off between inference speed and accuracy, achieving comparable performance to the state-of-the-art models while speeding up by over 4.5 times.

Cite

CITATION STYLE

APA

Cheng, X., Zhu, Z., Xu, W., Li, Y., Li, H., & Zou, Y. (2023). Accelerating Multiple Intent Detection and Slot Filling via Targeted Knowledge Distillation. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 8900–8910). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.597

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free