Co-VQA: Answering by Interactive Sub Question Sequence

9Citations
Citations of this article
37Readers
Mendeley users who have this article in their library.

Abstract

Most existing approaches to Visual Question Answering (VQA) answer questions directly, however, people usually decompose a complex question into a sequence of simple sub questions and finally obtain the answer to the original question after answering the sub question sequence(SQS). By simulating the process, this paper proposes a conversation-based VQA (Co-VQA) framework, which consists of three components: Questioner, Oracle, and Answerer. Questioner raises the sub questions using an extending HRED model, and Oracle answers them one-by-one. An Adaptive Chain Visual Reasoning Model (ACVRM) for Answerer is also proposed, where the question-answer pair is used to update the visual representation sequentially. To perform supervised learning for each model, we introduce a well-designed method to build a SQS for each question on VQA 2.0 and VQA-CP v2 datasets. Experimental results show that our method achieves state-of-the-art on VQA-CP v2. Further analyses show that SQSs help build direct semantic connections between questions and images, provide question-adaptive variable-length reasoning chains, and with explicit interpretability as well as error traceability.

Cite

CITATION STYLE

APA

Wang, R., Qian, Y., Feng, F., Wang, X., & Jiang, H. (2022). Co-VQA: Answering by Interactive Sub Question Sequence. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 2396–2408). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-acl.188

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free