4.7 Article

MSNet: multispectral semantic segmentation network for remote sensing images

Journal

GISCIENCE & REMOTE SENSING
Volume 59, Issue 1, Pages 1177-1198

Publisher

TAYLOR & FRANCIS LTD
DOI: 10.1080/15481603.2022.2101728

Keywords

Multispectral remote sensing images; spectral feature; feature fusion; semantic segmentation

Funding

  1. National Key Research and Development Program of China [2021YFB3900603]

Ask authors/readers for more resources

In the study of automatic interpretation of remote sensing images, semantic segmentation based on deep convolutional neural networks has been widely developed and applied. However, most current network designs focus on the visible RGB bands, neglecting the spectral information in the invisible light bands such as NIR. To address this issue, this paper proposes a novel deep neural network structure called the multispectral semantic segmentation network (MSNet), which achieves competitive performance for semantic segmentation of multi-classified feature scenes by leveraging the advantages of multispectral data and incorporating visible and invisible bands.
In the research of automatic interpretation of remote sensing images, semantic segmentation based on deep convolutional neural networks has been rapidly developed and applied, and the feature segmentation accuracy and network model generalization ability have been gradually improved. However, most of the network designs are mainly oriented to the three visible RGB bands of remote sensing images, aiming to be able to directly borrow the mature natural image semantic segmentation networks and pre-trained models, but simultaneously causing the waste and loss of spectral information in the invisible light bands such as near-infrared (NIR) of remote sensing images. Combining the advantages of multispectral data in distinguishing typical features such as water and vegetation, we propose a novel deep neural network structure called the multispectral semantic segmentation network (MSNet) for semantic segmentation of multi-classified feature scenes. The multispectral remote sensing image bands are split into two groups, visible and invisible, and ResNet-50 is used for feature extraction in both coding stages, and cascaded upsampling is used to recover feature map resolution in the decoding stage, and the multi-scale image features and spectral features from the upsampling process are fused layer by layer using the feature pyramid structure to finally obtain semantic segmentation results. The training and validation results on two publicly available datasets show that MSNet has competitive performance. The code is available: https://github.com/taochx/MSNet.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available