Improving limited labeled dialogue state tracking with self-supervision

15Citations
Citations of this article
73Readers
Mendeley users who have this article in their library.

Abstract

Existing dialogue state tracking (DST) models require plenty of labeled data. However, collecting high-quality labels is costly, especially when the number of domains increases. In this paper, we address a practical DST problem that is rarely discussed, i.e., learning efficiently with limited labeled data. We present and investigate two self-supervised objectives: preserving latent consistency and modeling conversational behavior. We encourage a DST model to have consistent latent distributions given a perturbed input, making it more robust to an unseen scenario. We also add an auxiliary utterance generation task, modeling a potential correlation between conversational behavior and dialogue states. The experimental results show that our proposed self-supervised signals can improve joint goal accuracy by 8.95% when only 1% labeled data is used on the MultiWOZ dataset. We can achieve an additional 1.76% improvement if some unlabeled data is jointly trained as semi-supervised learning. We analyze and visualize how our proposed self-supervised signals help the DST task and hope to stimulate future data-efficient DST research.

Cite

CITATION STYLE

APA

Wu, C. S., Hoi, S., & Xiong, C. (2020). Improving limited labeled dialogue state tracking with self-supervision. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 4462–4472). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.400

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free