Towards visual dialog for radiology

26Citations
Citations of this article
77Readers
Mendeley users who have this article in their library.

Abstract

Current research in machine learning for radiology is focused mostly on images. There exists limited work in investigating intelligent interactive systems for radiology. To address this limitation, we introduce a realistic and information-rich task of Visual Dialog in radiology, specific to chest X-ray images. Using MIMIC-CXR, an openly available database of chest X-ray images, we construct both a synthetic and a real-world dataset and provide baseline scores achieved by state-of-theart models. We show that incorporating medical history of the patient leads to better performance in answering questions as opposed to conventional visual question answering model which looks only at the image. While our experiments show promising results, they indicate that the task is extremely challenging with significant scope for improvement. We make both the datasets (synthetic and gold standard) and the associated code publicly available to the research community

Cite

CITATION STYLE

APA

Kovaleva, O., Shivade, C., Kashyap, S., Kanjaria, K., Coy, A., Ballah, D., … Mukherjee, V. (2020). Towards visual dialog for radiology. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 60–69). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.bionlp-1.6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free