A Model of Cross-Lingual Knowledge-Grounded Response Generation for Open-Domain Dialogue Systems

18Citations
Citations of this article
50Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Research on open-domain dialogue systems that allow free topics is challenging in the field of natural language processing (NLP). The performance of the dialogue system has been improved recently by the method utilizing dialogue-related knowledge; however, nonEnglish dialogue systems suffer from reproducing the performance of English dialogue systems because securing knowledge in the same language with the dialogue system is relatively difficult. Through experiments with a Korean dialogue system, this paper proves that the performance of a non-English dialogue system can be improved by utilizing English knowledge, highlighting the system uses cross-lingual knowledge. For the experiments, we 1) constructed a Korean version of the Wizard of Wikipedia dataset, 2) built KoreanEnglish T5 (KE-T5), a language model pretrained with Korean and English corpus, and 3) developed a knowledge-grounded Korean dialogue model based on KE-T5. We observed the performance improvement in the opendomain Korean dialogue model even only English knowledge was given. The experimental results showed that the knowledge inherent in cross-lingual language models can be helpful for generating responses in open dialogue systems.

Cite

CITATION STYLE

APA

Kim, S., Jang, J. Y., Jung, M., & Shin, S. (2021). A Model of Cross-Lingual Knowledge-Grounded Response Generation for Open-Domain Dialogue Systems. In Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021 (pp. 352–365). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-emnlp.33

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free