Abstract
Recent years have seen significant advances in multi-turn Spoken Language Understanding (SLU), where dialogue contexts are used to guide intent classification and slot filling. However, how to selectively incorporate dialogue contexts, such as previous utterances and dialogue acts, into multi-turn SLU still remains a substantial challenge. In this work, we propose a novel contextual SLU model for multi-turn intent classification and slot filling tasks. We introduce an adaptive global-local context fusion mechanism to selectively integrate dialogue contexts into our model. The local context fusion aligns each dialogue context using multi-head attention, while the global context fusion measures overall context contribution to intent classification and slot filling tasks. Experiments show that on two benchmark datasets, our model achieves absolute F1 score improvements of 2.73% and 2.57% for the slot filling task on Sim-R and Sim-M datasets, respectively. Additional experiments on a large-scale, de-identified, in-house dataset further verify the measurable accuracy gains of our proposed model.
Cite
CITATION STYLE
Tran, T., Wei, K., Ruan, W., McGowan, R., Susanj, N., & Strimel, G. P. (2022). Adaptive Global-Local Context Fusion for Multi-Turn Spoken Language Understanding. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 12622–12628). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i11.21536
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.