Multimodal remote sensing image matching is a challenging task because of the existence of significant radiometric differences. To address the problem, we develop a novel multimodal remote sensing image matching method based on self-similarity features. The offset mean filtering method is proposed first to calculate the self-similarity features fast based on the symmetry of the self-similarity. The self-similarity features are presented through a multichannel self-similarity map (SSM) and a corresponding multichannel symmetric SSM. On this basis, we develop the image matching method, including a feature detector named improved maximal self-dissimilarities (IMSD) and a feature descriptor named oriented self-similarity (OSS). The IMSD detector is designed by introducing the two multichannel SSMs into the maximal self-dissimilarities (MSD) detector for feature point detection. The OSS descriptor is proposed based on the orientations of the self-similarities extracted from the multichannel SSMs. We conduct experiments with a variety of optical, synthetic aperture radar, and light detection, and ranging data. Our results demonstrate the advantages of our proposed IMSD detector and OSS descriptor in comparison with state-of-the-art detectors and descriptors, respectively. The image registration results further confirm the effectiveness of the proposed method.
CITATION STYLE
Xiong, X., Jin, G., Xu, Q., & Zhang, H. (2021). Self-Similarity Features for Multimodal Remote Sensing Image Matching. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 14, 12440–12454. https://doi.org/10.1109/JSTARS.2021.3131489
Mendeley helps you to discover research relevant for your work.