Multi-focus image fusion with Siamese self-attention network

20Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

Recently, convolutional neural networks (CNNs) have achieved impressive progress in multi-focus image fusion (MFF). However, it always fails to capture sufficient discrimination features due to the local receptive field limitations of the convolutional operator, restricting most current CNN-based methods’ performance. To address this issue, by leveraging self-attention (SA) mechanism, the authors propose Siamese SA network (SSAN) for MFF. Specifically, two kinds of SA modules, position SA (PSA) and channel SA (CSA) are utilised to model the long-range dependencies across focused and defocused regions in the multi-focus image, alleviating the local receptive field limitations of convolution operators in CNN. To search a better feature representation of the input image for MFF, the captured features obtained by PSA and CSA are further merged through a learnable 1 × 1 convolution operator. The whole pipeline is in a Siamese network fashion to reduce the complexity. After training, the authors SSAN can accomplish well the fusion task with no post-processing. Experiments demonstrate that their approach outperforms other current state-of-the-art methods, not only in subjective visual perception but also in the quantitative assessment.

Cite

CITATION STYLE

APA

Guo, X., Meng, L., Mei, L., Weng, Y., & Tong, H. (2020). Multi-focus image fusion with Siamese self-attention network. IET Image Processing, 14(7), 1303–1315. https://doi.org/10.1049/iet-ipr.2019.0883

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free