Pretrained language models for dialogue generation with multiple input sources

19Citations
Citations of this article
89Readers
Mendeley users who have this article in their library.

Abstract

Large-scale pretrained language models have achieved outstanding performance on natural language understanding tasks. However, it is still under investigating how to apply them to dialogue generation tasks, especially those with responses conditioned on multiple sources. Previous work simply concatenates all input sources or averages information from different input sources. In this work, we study dialogue models with multiple input sources adapted from the pretrained language model GPT2. We explore various methods to fuse multiple separate attention information corresponding to different sources. Our experimental results show that proper fusion methods deliver higher relevance with dialogue history than simple fusion baselines.

Cite

CITATION STYLE

APA

Cao, Y., Bi, W., Fang, M., & Tao, D. (2020). Pretrained language models for dialogue generation with multiple input sources. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 909–917). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.81

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free