Visual question generation from radiology images

15Citations
Citations of this article
79Readers
Mendeley users who have this article in their library.

Abstract

Visual Question Generation (VQG), the task of generating a question based on image contents, is an increasingly important area that combines natural language processing and computer vision. Although there are some recent works that have attempted to generate questions from images in the open domain, the task of VQG in the medical domain has not been explored so far. In this paper, we introduce an approach to generation of visual questions about radiology images called VQGR, i.e. an algorithm that is able to ask a question when shown an image. VQGR first generates new training data from the existing examples, based on contextual word embeddings and image augmentation techniques. It then uses the variational auto-encoders model to encode images into a latent space and decode natural language questions. Experimental automatic evaluations performed on the VQA-RAD dataset of clinical visual questions show that VQGR achieves good performances compared with the baseline system. The source code is available at https://github.com/ sarrouti/vqgr.

Cite

CITATION STYLE

APA

Sarrouti, M., Abacha, A. B., & Demner-Fushmen, D. (2020). Visual question generation from radiology images. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 12–18). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.alvr-1.3

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free