Dialogue response generation requires an agent to generate a response according to the current dialogue history, in terms of which two-party dialogues have been well studied, but leaving a great gap for multi-party dialogues at the same time. Different from two-party dialogues where each response is a direct reply to its previous utterance, the addressee of a response utterance should be specified before it is generated in the multi-party scenario. Thanks to the huge amount of two-party conversational data, various pre-trained language models for two-party dialogue response generation have been proposed. However, due to the lack of annotated addressee labels in multi-party dialogue datasets, it is hard to use them to pre-train a response generation model for multi-party dialogues. To tackle this obstacle, we propose an Expectation-Maximization (EM) approach that iteratively performs the expectation steps to generate addressee labels, and the maximization steps to optimize a response generation model. Theoretical analyses and extensive experiments have justified the feasibility and effectiveness of our proposed method. The official implementation of this paper is available at https://github.com/EricLee8/MPDRG.
Mendeley helps you to discover research relevant for your work.
CITATION STYLE
Li, Y., & Zhao, H. (2023). EM Pre-training for Multi-party Dialogue Response Generation. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 92–103). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-long.7