Abstract
Neural generative dialogue agents have shown an increasing ability to hold short chitchat conversations, when evaluated by crowdworkers in controlled settings. However, their performance in real-life deployment - talking to intrinsically-motivated users in noisy environments - is less well-explored. In this paper, we perform a detailed case study of a neural generative model deployed as part of Chirpy Cardinal, an Alexa Prize socialbot. We find that unclear user utterances are a major source of generative errors such as ignoring, hallucination, unclearness and repetition. However, even in unambiguous contexts the model frequently makes reasoning errors. Though users express dissatisfaction in correlation with these errors, certain dissatisfaction types (such as offensiveness and privacy objections) depend on additional factors - such as the user's personal attitudes, and prior unaddressed dissatisfaction in the conversation. Finally, we show that dissatisfied user utterances can be used as a semisupervised learning signal to improve the dialogue system. We train a model to predict nextturn dissatisfaction, and show through human evaluation that as a ranking function, it selects higher-quality neural-generated utterances. 021 Association for Computational Linguistics.
Cite
CITATION STYLE
See, A., & Manning, C. D. (2021). Understanding and predicting user dissatisfaction in a neural generative chatbot. In SIGDIAL 2021 - 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue, Proceedings of the Conference (pp. 1–12). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.sigdial-1.1
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.