Bag of Tricks for Improving Deep Learning Performance on Multimodal Image Classification

3Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.

Abstract

A comprehensive medical image-based diagnosis is usually performed across various image modalities before passing a final decision; hence, designing a deep learning model that can use any medical image modality to diagnose a particular disease is of great interest. The available methods are multi-staged, with many computational bottlenecks in between. This paper presents an improved end-to-end method of multimodal image classification using deep learning models. We present top research methods developed over the years to improve models trained from scratch and transfer learning approaches. We show that when fully trained, a model can first implicitly discriminate the imaging modality and then diagnose the relevant disease. Our developed models were applied to COVID-19 classification from chest X-ray, CT scan, and lung ultrasound image modalities. The model that achieved the highest accuracy correctly maps all input images to their respective modality, then classifies the disease achieving overall 91.07% accuracy.

Cite

CITATION STYLE

APA

Adeshina, S. A., & Adedigba, A. P. (2022). Bag of Tricks for Improving Deep Learning Performance on Multimodal Image Classification. Bioengineering, 9(7). https://doi.org/10.3390/bioengineering9070312

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free