Turn-Level Active Learning for Dialogue State Tracking

0Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

Dialogue state tracking (DST) plays an important role in task-oriented dialogue systems. However, collecting a large amount of turn-by-turn annotated dialogue data is costly and inefficient. In this paper, we propose a novel turn-level active learning framework for DST to actively select turns in dialogues to annotate. Given the limited labelling budget, experimental results demonstrate the effectiveness of selective annotation of dialogue turns. Additionally, our approach can effectively achieve comparable DST performance to traditional training approaches with significantly less annotated data, which provides a more efficient way to annotate new dialogue data.

Cite

CITATION STYLE

APA

Zhang, Z., Fang, M., Ye, F., Chen, L., & Namazi-Rad, M. R. (2023). Turn-Level Active Learning for Dialogue State Tracking. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 7705–7719). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-main.478

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free