Dialogue in the Wild: Learning from a Deployed Role-Playing Game with Humans and Bots

5Citations
Citations of this article
68Readers
Mendeley users who have this article in their library.

Abstract

Much of NLP research has focused on crowdsourced static datasets and the supervised learning paradigm of training once and then evaluating test performance. As argued in de Vries et al. (2020), crowdsourced data has the issues of lack of naturalness and relevance to real-world use cases, while the static dataset paradigm does not allow for a model to learn from its experiences of using language (Silver et al., 2013). In contrast, one might hope for machine learning systems that become more useful as they interact with people. In this work, we build and deploy a role-playing game, whereby human players converse with learning agents situated in an open-domain fantasy world. We show that by training models on the conversations they have with humans in the game the models progressively improve, as measured by automatic metrics and online engagement scores. This learning is shown to be more efficient than crowdsourced data when applied to conversations with real users, as well as being far cheaper to collect.

Cite

CITATION STYLE

APA

Shuster, K., Urbanek, J., Dinan, E., Szlam, A., & Weston, J. (2021). Dialogue in the Wild: Learning from a Deployed Role-Playing Game with Humans and Bots. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 (pp. 611–624). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-acl.54

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free