I already said that! Degenerating redundant questions in open-domain dialogue systems

0Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Neural text generation models have achieved remarkable success in carrying on short open-domain conversations. However, their performance degrades significantly in the long term, especially in their ability to ask coherent questions. A significant issue is the generation of redundant questions where the answer has already been provided by the user. We adapt and evaluate different methods, including negative training, decoding, and classification, to mitigate the redundancy problem. We also propose a simple yet effective method for generating training data without the need for crowdsourcing human-human or human-bot conversations. Experiments with the BlenderBot model show that our combined method significantly reduces the rate of redundant questions from 27.2% to 8.7%, while improving the quality of the original model. The code, dataset, and trained models can be found at our repository.

Cite

CITATION STYLE

APA

Mai, L., & Carson-Berndsen, J. (2023). I already said that! Degenerating redundant questions in open-domain dialogue systems. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 4, pp. 226–236). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-srw.33

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free