Semantic Segmentation of Remote Sensing Images with Self-Supervised Multitask Representation Learning

48Citations
Citations of this article
45Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Existing deep learning-based remote sensing images semantic segmentation methods require large-scale labeled datasets. However, the annotation of segmentation datasets is often too time-consuming and expensive. To ease the burden of data annotation, self-supervised representation learning methods have emerged recently. However, the semantic segmentation methods need to learn both high-level and low-level features, but most of the existing self-supervised representation learning methods usually focus on one level, which affects the performance of semantic segmentation for remote sensing images. In order to solve this problem, we propose a self-supervised multitask representation learning method to capture effective visual representations of remote sensing images. We design three different pretext tasks and a triplet Siamese network to learn the high-level and low-level image features at the same time. The network can be trained without any labeled data, and the trained model can be fine-tuned with the annotated segmentation dataset. We conduct experiments on Potsdam, Vaihingen dataset, and cloud/snow detection dataset Levir_CS to verify the effectiveness of our methods. Experimental results show that our proposed method can effectively reduce the demand of labeled datasets and improve the performance of remote sensing semantic segmentation. Compared with the recent state-of-the-art self-supervised representation learning methods and the mostly used initialization methods (such as random initialization and ImageNet pretraining), our proposed method has achieved the best results in most experiments, especially in the case of few training data. With only 10% to 50% labeled data, our method can achieve the comparable performance compared with random initialization. Codes are available at https://github.com/flyakon/SSLRemoteSensing.

References Powered by Scopus

Deep residual learning for image recognition

176401Citations
N/AReaders
Get full text

U-net: Convolutional networks for biomedical image segmentation

65876Citations
N/AReaders
Get full text

ImageNet: A Large-Scale Hierarchical Image Database

51965Citations
N/AReaders
Get full text

Cited by Powered by Scopus

RingMo: A Remote Sensing Foundation Model With Masked Image Modeling

172Citations
N/AReaders
Get full text

Artificial intelligence for geoscience: Progress, challenges, and perspectives

160Citations
N/AReaders
Get full text

Self-Supervised Learning in Remote Sensing: A review

133Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Li, W., Chen, H., & Shi, Z. (2021). Semantic Segmentation of Remote Sensing Images with Self-Supervised Multitask Representation Learning. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 14, 6438–6450. https://doi.org/10.1109/JSTARS.2021.3090418

Readers over time

‘21‘22‘23‘24‘2506121824

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 13

68%

Researcher 4

21%

Professor / Associate Prof. 2

11%

Readers' Discipline

Tooltip

Computer Science 9

45%

Engineering 6

30%

Earth and Planetary Sciences 4

20%

Chemical Engineering 1

5%

Save time finding and organizing research with Mendeley

Sign up for free
0