History for visual dialog: Do we really need it?

49Citations
Citations of this article
146Readers
Mendeley users who have this article in their library.

Abstract

Visual Dialog involves “understanding” the dialog history (what has been discussed previously) and the current question (what is asked), in addition to grounding information in the image, to generate the correct response. In this paper, we show that co-attention models which explicitly encode dialog history outperform models that don't, achieving state-of-the-art performance (72 % NDCG on val set). However, we also expose shortcomings of the crowd-sourcing dataset collection procedure by showing that history is indeed only required for a small amount of the data and that the current evaluation metric encourages generic replies. To that end, we propose a challenging subset (VisDialConv) of the VisDial val set and provide a benchmark of 63% NDCG.

Cite

CITATION STYLE

APA

Agarwal, S., Bui, T., Lee, J. Y., Konstas, I., & Rieser, V. (2020). History for visual dialog: Do we really need it? In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 8182–8197). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-main.728

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free