DIALOGPT: Large-scale generative pre-training for conversational response generation

685Citations
Citations of this article
704Readers
Mendeley users who have this article in their library.

Abstract

We present a large, tunable neural conversational response generation model, DIALOGPT (dialogue generative pre-trained transformer). Trained on 147M conversation-like exchanges extracted from Reddit comment chains over a period spanning from 2005 through 2017, DialoGPT extends the Hugging Face PyTorch transformer to attain a performance close to human both in terms of automatic and human evaluation in single-turn dialogue settings. We show that conversational systems that leverage DialoGPT generate more relevant, contentful and context-consistent responses than strong baseline systems. The pre-trained model and training pipeline are publicly released to facilitate research into neural response generation and the development of more intelligent open-domain dialogue systems.

Cite

CITATION STYLE

APA

Zhang, Y., Sun, S., Galley, M., Chen, Y. C., Brockett, C., Gao, X., … Dolan, B. (2020). DIALOGPT: Large-scale generative pre-training for conversational response generation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 270–278). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-demos.30

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free