Why are Sequence-to-Sequence Models So Dull? Understanding the Low-Diversity Problem of Chatbots

27Citations
Citations of this article
125Readers
Mendeley users who have this article in their library.

Abstract

Diversity is a long-studied topic in information retrieval that usually refers to the requirement that retrieved results should be non-repetitive and cover different aspects. In a conversational setting, an additional dimension of diversity matters: an engaging response generation system should be able to output responses that are diverse and interesting. Sequence-to-sequence (Seq2Seq) models have been shown to be very effective for response generation. However, dialogue responses generated by Seq2Seq models tend to have low diversity. In this paper, we review known sources and existing approaches to this low-diversity problem. We also identify a source of low diversity that has been little studied so far, namely model over-confidence. We sketch several directions for tackling model over-confidence and, hence, the low-diversity problem, including confidence penalties and label smoothing.

Cite

CITATION STYLE

APA

Jiang, S., & de Rijke, M. (2018). Why are Sequence-to-Sequence Models So Dull? Understanding the Low-Diversity Problem of Chatbots. In Proceedings of the 2018 EMNLP Workshop SCAI 2018: The 2nd International Workshop on Search-Oriented Conversational AI (pp. 81–86). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w18-5712

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free