Adaptive Global-Local Context Fusion for Multi-Turn Spoken Language Understanding

8Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.

Abstract

Recent years have seen significant advances in multi-turn Spoken Language Understanding (SLU), where dialogue contexts are used to guide intent classification and slot filling. However, how to selectively incorporate dialogue contexts, such as previous utterances and dialogue acts, into multi-turn SLU still remains a substantial challenge. In this work, we propose a novel contextual SLU model for multi-turn intent classification and slot filling tasks. We introduce an adaptive global-local context fusion mechanism to selectively integrate dialogue contexts into our model. The local context fusion aligns each dialogue context using multi-head attention, while the global context fusion measures overall context contribution to intent classification and slot filling tasks. Experiments show that on two benchmark datasets, our model achieves absolute F1 score improvements of 2.73% and 2.57% for the slot filling task on Sim-R and Sim-M datasets, respectively. Additional experiments on a large-scale, de-identified, in-house dataset further verify the measurable accuracy gains of our proposed model.

Cite

CITATION STYLE

APA

Tran, T., Wei, K., Ruan, W., McGowan, R., Susanj, N., & Strimel, G. P. (2022). Adaptive Global-Local Context Fusion for Multi-Turn Spoken Language Understanding. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 12622–12628). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i11.21536

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free