Object-less Vision-language Model on Visual Question Classification for Blind People

4Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Despite the long-standing appearance of question types in the Visual Question Answering dataset, Visual Question Classification does not received enough public interest in research. Different from general text classification, a visual question requires an understanding of visual and textual features simultaneously. Together with the enthusiasm and novelty of Visual Question Classification, the most important and practical goal we concentrate on is to deal with the weakness of Object Detection on object-less images. We thus propose an Object-less Visual Question Classification model, OL–LXMERT, to generate virtual objects replacing the dependence of Object Detection in previous Vision-Language systems. Our architecture is effective and powerful enough to digest local and global features of images in understanding the relationship between multiple modalities. Through our experiments in our modified VizWiz-VQC 2020 dataset of blind people, our Object-less LXMERT achieves promising results in the brand-new multi-modal task. Furthermore, the detailed ablation studies show the strength and potential of our model in comparison to competitive approaches.

Cite

CITATION STYLE

APA

Le, T., Pho, K., Bui, T., Nguyen, H. T., & Le Nguyen, M. (2022). Object-less Vision-language Model on Visual Question Classification for Blind People. In International Conference on Agents and Artificial Intelligence (Vol. 3, pp. 180–187). Science and Technology Publications, Lda. https://doi.org/10.5220/0010797400003116

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free