Two-step training and mixed encoding-decoding for implementing a generative chatbot with a small dialogue corpus

13Citations
Citations of this article
102Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Generative chatbot models based on sequence-to-sequence networks can generate natural conversation interactions if a huge dialogue corpus is used as training data. However, except for a few languages such as English and Chinese, it remains difficult to collect a large dialogue corpus. To address this problem, we propose a chatbot model using a mixture of words and syllables as encoding-decoding units. In addition, we propose a two-step training method, involving pre-training using a large non-dialogue corpus and re-training using a small dialogue corpus. In our experiments, the mixture units were shown to help reduce out-of-vocabulary (OOV) problems. Moreover, the two-step training method was effective in reducing grammatical and semantical errors in responses when the chatbot was trained using a small dialogue corpus (533,997 sentence pairs).

Cite

CITATION STYLE

APA

Kim, J., Lee, H. G., Kim, H., Lee, Y., & Kim, Y. G. (2018). Two-step training and mixed encoding-decoding for implementing a generative chatbot with a small dialogue corpus. In 2IS and NLG 2018 - Workshop on Intelligent Interactive Systems and Language Generation, Proceedings of the Workshop (pp. 31–35). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/W18-6707

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free