Recent non-autoregressive Spoken Language Understanding (SLU) models attracts increasing attention owing to the high inference speed. However, most of them still (1) suffer from the multi-modality problem since the prior knowledge about the reference is relatively poor during inference; (2) fail to achieve a satisfactory inference speed limited by their complex frameworks. To tackle these problems, in this paper, we propose a Targeted Knowledge Distillation Framework (TKDF), which applies knowledge distillation to improve the performance. Specifically, we first train an SLU model as a teacher model, which has higher accuracy while slower inference speed. Then we introduce an evaluator and utilize the curriculum learning strategy to select proper targets for the student model. Experiment results on two public multi-intent SLU datasets demonstrate that our method can realize a flexible trade-off between inference speed and accuracy, achieving comparable performance to the state-of-the-art models while speeding up by over 4.5 times.
CITATION STYLE
Cheng, X., Zhu, Z., Xu, W., Li, Y., Li, H., & Zou, Y. (2023). Accelerating Multiple Intent Detection and Slot Filling via Targeted Knowledge Distillation. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 8900–8910). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.597
Mendeley helps you to discover research relevant for your work.