Learning to Ground Medical Text in a 3D Human Atlas

4Citations
Citations of this article
62Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we develop a method for grounding medical text into a physically meaningful and interpretable space corresponding to a human atlas. We build on text embedding architectures such as BERT and introduce a loss function that allows us to reason about the semantic and spatial relatedness of medical texts by learning a projection of the embedding into a 3D space representing the human body. We quantitatively and qualitatively demonstrate that our proposed method learns a context sensitive and spatially aware mapping, in both the inter-organ and intra-organ sense, using a large scale medical text dataset from the “Large-scale online biomedical semantic indexing” track of the 2020 BioASQ challenge. We extend our approach to a self-supervised setting, and find it to be competitive with a classification based method, and a fully supervised variant of approach.

Cite

CITATION STYLE

APA

Grujicic, D., Radevski, G., Tuytelaars, T., & Blaschko, M. B. (2020). Learning to Ground Medical Text in a 3D Human Atlas. In CoNLL 2020 - 24th Conference on Computational Natural Language Learning, Proceedings of the Conference (pp. 302–312). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.conll-1.23

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free