Solving Visual Madlibs with Multiple Cues

9Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.

Abstract

This paper focuses on answering fill-in-the-blank style multiple choice questions from the Visual Madlibs dataset. Previous approaches to Visual Question Answering (VQA) have mainly used generic image features from networks trained on the ImageNet dataset, despite the wide scope of questions. In contrast, our approach employs features derived from networks trained for specialized tasks of scene classification, person activity prediction, and person and object attribute prediction. We also present a method for selecting sub-regions of an image that are relevant for evaluating the appropriateness of a putative answer. Visual features are computed both from the whole image and from local regions, while sentences are mapped to a common space using a simple normalized canonical correlation analysis (CCA) model. Our results show a significant improvement over the previous state of the art, and indicate that answering different question types benefits from examining a variety of image cues and carefully choosing informative image sub-regions.

Cite

CITATION STYLE

APA

Tommasi, T., Mallya, A., Plummer, B., Lazebnik, S., Berg, A. C., & Berg, T. L. (2016). Solving Visual Madlibs with Multiple Cues. In British Machine Vision Conference 2016, BMVC 2016 (Vol. 2016-September, pp. 77.1-77.13). British Machine Vision Conference, BMVC. https://doi.org/10.5244/C.30.77

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free