Learning visual-semantic embeddings for reporting abnormal findings on chest x-rays

33Citations
Citations of this article
86Readers
Mendeley users who have this article in their library.

Abstract

Automatic medical image report generation has drawn growing attention due to its potential to alleviate radiologists’ workload. Existing work on report generation often trains encoder-decoder networks to generate complete reports. However, such models are affected by data bias (e.g. label imbalance) and face common issues inherent in text generation models (e.g. repetition). In this work, we focus on reporting abnormal findings on radiology images; instead of training on complete radiology reports, we propose a method to identify abnormal findings from the reports in addition to grouping them with unsupervised clustering and minimal rules. We formulate the task as cross-modal retrieval and propose Conditional Visual-Semantic Embeddings to align images and fine-grained abnormal findings in a joint embedding space. We demonstrate that our method is able to retrieve abnormal findings and outperforms existing generation models on both clinical correctness and text generation metrics.

Cite

CITATION STYLE

APA

Ni, J., Hsu, C. N., Gentili, A., & McAuley, J. (2020). Learning visual-semantic embeddings for reporting abnormal findings on chest x-rays. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 1954–1960). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.176

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free