期刊
REMOTE SENSING
卷 15, 期 20, 页码 -出版社
MDPI
DOI: 10.3390/rs15205048
关键词
building extraction; remote sensing; recurrent residual convolution; U-Net; multi-head self-attention
In this paper, a novel method for building extraction based on U-Net is proposed. By incorporating a recurrent residual deformable convolution unit (RDCU) module and augmented multi-head self-attention (AMSA), the method enhances the ability to learn building shape details and improves feature expression and positions. Experimental results on three datasets demonstrate the effectiveness of the proposed method.
Considering the challenges associated with accurately identifying building shape features and distinguishing between building and non-building features during the extraction of buildings from remote sensing images using deep learning, we propose a novel method for building extraction based on U-Net, incorporating a recurrent residual deformable convolution unit (RDCU) module and augmented multi-head self-attention (AMSA). By replacing conventional convolution modules with an RDCU, which adopts a deformable convolutional neural network within a residual network structure, the proposed method enhances the module's capacity to learn intricate details such as building shapes. Furthermore, AMSA is introduced into the skip connection function to enhance feature expression and positions through content-position enhancement operations and content-content enhancement operations. Moreover, AMSA integrates an additional fusion channel attention mechanism to aid in identifying cross-channel feature expression Intersection over Union (IoU) score differences. For the Massachusetts dataset, the proposed method achieves an Intersection over Union (IoU) score of 89.99%, PA (Pixel Accuracy) score of 93.62%, and Recall score of 89.22%. For the WHU Satellite dataset I, the proposed method achieves an IoU score of 86.47%, PA score of 92.45%, and Recall score of 91.62%, For the INRIA dataset, the proposed method achieves an IoU score of 80.47%, PA score of 90.15%, and Recall score of 85.42%.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据