4.7 Article

A Gather-to-Guide Network for Remote Sensing Semantic Segmentation of RGB and Auxiliary Image

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TGRS.2021.3103517

关键词

Semantics; Image segmentation; Remote sensing; Feature extraction; Convolutional neural networks; Calibration; Task analysis; Deep learning; remote sensing; semantic segmentation

资金

  1. National Key Research and Development Program of China [2018Y-FB0505401]
  2. National Natural Science Foundation of China Project [42071370, 41871361]

向作者/读者索取更多资源

The proposed unified gather-to-guide network (G2GNet) for remote sensing semantic segmentation utilizes a gather-to-guide module (G2GM) to calibrate RGB features and improve segmentation performance. By generating cross-modal descriptors and using channel-wise guide weights, the G2GM preserves informative features while suppressing redundant and noisy information. Extensive experiments demonstrate the robustness of G2GNet to data uncertainties and its ability to enhance the semantic segmentation of RGB and auxiliary remote sensing data.
Convolutional neural network (CNN)-based feature fusion of RGB and auxiliary remote sensing data is known to enable improved semantic segmentation. However, such fusion is challengeable because of the substantial variance in data characteristics and quality (e.g., data uncertainties and misalignment) between two modality data. In this article, we propose a unified gather-to-guide network (G2GNet) for remote sensing semantic segmentation of RGB and auxiliary data. The key aspect of the proposed architecture is a novel gather-to-guide module (G2GM) that consists of a feature gatherer and a feature guider. The feature gatherer generates a set of cross-modal descriptors by absorbing the complementary merits of RGB and auxiliary modality data. The feature guider calibrates the RGB feature response by using the channel-wise guide weights extracted from the cross-modal descriptors. In this way, the G2GM can perform RGB feature calibration with different modality data in a gather-to-guide fashion, thus preserving the informative features while suppressing redundant and noisy information. Extensive experiments conducted on two benchmark datasets show that the proposed G2GNet is robust to data uncertainties while also improving the semantic segmentation performance of RGB and auxiliary remote sensing data.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据