Deep Reinforcement Learning for On-line Dialogue State Tracking

1Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Dialogue state tracking (DST) is a crucial module in dialogue management. It is usually cast as a supervised training problem, which is not convenient for on-line optimization. In this paper, a novel companion teaching based deep reinforcement learning (DRL) framework for on-line DST optimization is proposed. To the best of our knowledge, this is the first effort to optimize the DST module within DRL framework for on-line task-oriented spoken dialogue systems. In addition, dialogue policy can be further jointly updated. Experiments show that on-line DST optimization can effectively improve the dialogue manager performance while keeping the flexibility of using predefined policy. Joint training of both DST and policy can further improve the performance.

Cite

CITATION STYLE

APA

Chen, Z., Chen, L., Zhou, X., & Yu, K. (2023). Deep Reinforcement Learning for On-line Dialogue State Tracking. In Communications in Computer and Information Science (Vol. 1765 CCIS, pp. 278–292). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-981-99-2401-1_25

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free