Deep multimodal classification of image types in biomedical journal figures

13Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper presents a robust method for the classification of medical image types in figures of the biomedical literature using the fusion of visual and textual information. A deep convolutional network is trained to discriminate among 31 image classes including compound figures, diagnostic image types and generic illustrations, while another shallow convolutional network is used for the analysis of the captions paired with the images. Various fusion methods are analyzed as well as data augmentation approaches. The proposed system is validated on the ImageCLEF 2013 and 2016 figure and subfigure classification tasks, largely improving the currently best performance from 83.5% to 93.7% accuracy and 88.4% to 89.0% respectively.

Cite

CITATION STYLE

APA

Andrearczyk, V., & Müller, H. (2018). Deep multimodal classification of image types in biomedical journal figures. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11018 LNCS, pp. 3–14). Springer Verlag. https://doi.org/10.1007/978-3-319-98932-7_1

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free