Investigating deep reinforcement learning techniques in personalized dialogue generation

20Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper, we propose a personalized dialogue generation system, which combines reinforcement learning techniques with an attention-based hierarchical recurrent encoder-decoder model. Firstly, we incorporate user-specific information into the decoder to capture user's background information and speaking style. Secondly, we employ reinforcement learning techniques to maximize future reward in dialogue, which enables our system to generate topic-coherent, informative and grammatical responses. Moreover, we propose three types of rewards to characterize good conversations. Finally, we compare the performance of the following reinforcement learning methods in dialogue generation: policy gradient, Q-learning, and actor-critic algorithms. We conduct experiments to verify the effectiveness of the proposed model on two dialogue datasets. Experimental results demonstrate that our model can generate better personalized dialogues for different users. Quantitatively, our method achieves better performance than the state-of-the-art dialogue systems in terms of BLEU score, perplexity, and human evaluation.

Cite

CITATION STYLE

APA

Yang, M., Qu, Q., Lei, K., Zhu, J., Zhao, Z., Chen, X., & Huang, J. Z. (2018). Investigating deep reinforcement learning techniques in personalized dialogue generation. In SIAM International Conference on Data Mining, SDM 2018 (pp. 630–638). Society for Industrial and Applied Mathematics Publications. https://doi.org/10.1137/1.9781611975321.71

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free