Image-Chat: Engaging grounded conversations

60Citations
Citations of this article
161Readers
Mendeley users who have this article in their library.
Get full text

Abstract

To achieve the long-term goal of machines being able to engage humans in conversation, our models should captivate the interest of their speaking partners. Communication grounded in images, whereby a dialogue is conducted based on a given photo, is a setup naturally appealing to humans (Hu et al., 2014). In this work we study large-scale architectures and datasets for this goal. We test a set of neural architectures using state-of-the-art image and text representations, considering various ways to fuse the components. To test such models, we collect a dataset of grounded human-human conversations, where speakers are asked to play roles given a provided emotional mood or style, as the use of such traits is also a key factor in engagingness (Guo et al., 2019). Our dataset, Image-Chat, consists of 202k dialogues over 202k images using 215 possible style traits. Automatic metrics and human evaluations of engagingness show the efficacy of our approach; in particular, we obtain state-of-the-art performance on the existing IGC task, and our best performing model is almost on par with humans on the Image-Chat test set (preferred 47.7% of the time).

Cite

CITATION STYLE

APA

Shuster, K., Humeau, S., Bordes, A., & Weston, J. (2020). Image-Chat: Engaging grounded conversations. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 2414–2429). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-main.219

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free