Abstract
Generative chatbot models based on sequence-to-sequence networks can generate natural conversation interactions if a huge dialogue corpus is used as training data. However, except for a few languages such as English and Chinese, it remains difficult to collect a large dialogue corpus. To address this problem, we propose a chatbot model using a mixture of words and syllables as encoding-decoding units. In addition, we propose a two-step training method, involving pre-training using a large non-dialogue corpus and re-training using a small dialogue corpus. In our experiments, the mixture units were shown to help reduce out-of-vocabulary (OOV) problems. Moreover, the two-step training method was effective in reducing grammatical and semantical errors in responses when the chatbot was trained using a small dialogue corpus (533,997 sentence pairs).
Cite
CITATION STYLE
Kim, J., Lee, H. G., Kim, H., Lee, Y., & Kim, Y. G. (2018). Two-step training and mixed encoding-decoding for implementing a generative chatbot with a small dialogue corpus. In 2IS and NLG 2018 - Workshop on Intelligent Interactive Systems and Language Generation, Proceedings of the Workshop (pp. 31–35). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/W18-6707
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.