Hybrid attention mechanism of feature fusion for medical image segmentation

8Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.

Abstract

Traditional convolution neural networks (CNN) have achieved good performance in multi-organ segmentation of medical images. Due to the lack of ability to model long-range dependencies and correlations between image pixels, CNN usually ignores the information of channel dimension. To further improve the performance of multi-organ segmentation, a hybrid attention mechanism model is proposed. First, a CNN was used to extract multi-scale feature maps and fed into the Channel Attention Enhancement Module (CAEM) to selectively pay attention to target organs in medical images, and the Transformer encoded tokenized image patches from CNN feature maps as the input sequence to model long-range dependencies. Second, the decoder upsampled the output from Transformer and fused with the CAEM features in multi-scale through skip connections. Finally, we introduced a Refinement Module (RM) after the decoder to improve feature correlations of the same organ and the feature discriminability between different organs. The model outperformed on dice coefficient (%) and hd95 on both the synapse multi-organ segmentation and cardiac diagnosis challenge datasets. The hybrid attention mechanisms exhibited high efficiency and high segmentation accuracy in medical images.

Cite

CITATION STYLE

APA

Tong, S., Zuo, Z., Liu, Z., Sun, D., & Zhou, T. (2024). Hybrid attention mechanism of feature fusion for medical image segmentation. IET Image Processing, 18(1), 77–87. https://doi.org/10.1049/ipr2.12934

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free