Most existing approaches to Visual Question Answering (VQA) answer questions directly, however, people usually decompose a complex question into a sequence of simple sub questions and finally obtain the answer to the original question after answering the sub question sequence(SQS). By simulating the process, this paper proposes a conversation-based VQA (Co-VQA) framework, which consists of three components: Questioner, Oracle, and Answerer. Questioner raises the sub questions using an extending HRED model, and Oracle answers them one-by-one. An Adaptive Chain Visual Reasoning Model (ACVRM) for Answerer is also proposed, where the question-answer pair is used to update the visual representation sequentially. To perform supervised learning for each model, we introduce a well-designed method to build a SQS for each question on VQA 2.0 and VQA-CP v2 datasets. Experimental results show that our method achieves state-of-the-art on VQA-CP v2. Further analyses show that SQSs help build direct semantic connections between questions and images, provide question-adaptive variable-length reasoning chains, and with explicit interpretability as well as error traceability.
CITATION STYLE
Wang, R., Qian, Y., Feng, F., Wang, X., & Jiang, H. (2022). Co-VQA: Answering by Interactive Sub Question Sequence. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 2396–2408). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-acl.188
Mendeley helps you to discover research relevant for your work.