4.7 Article

Hyper-ES2T: Efficient Spatial-Spectral Transformer for the classification of hyperspectral remote sensing images

出版社

ELSEVIER
DOI: 10.1016/j.jag.2022.103005

关键词

Deep learning; Hyperspectral image classification; Remote sensing; Spatial-spectral Transformer; Feature enhancement; Bilateral classification network

资金

  1. Fundamental Research Funds for the China Central Universities of USTB, China
  2. Sci-entific and Technological Innovation Foundation of Shunde Graduate School, USTB, China
  3. [FRF-DF-19-002]
  4. [BK20BE014]

向作者/读者索取更多资源

Convolutional neural networks have been dominating the downstream tasks on hyperspectral remote sensing images with their strong local feature extraction capability. However, they fail to effectively capture long-range dependencies, which the Transformer architecture can handle. This paper introduces a dual-branch Transformer architecture called Hyper-ES2T, which effectively utilizes spatial information and spectral correlations in hyperspectral images. The design also includes an efficient multi-head self-attention block to balance model accuracy and efficiency.
In recent years, convolutional neural networks have continuously dominated the downstream tasks on hyperspectral remote sensing images with its strong local feature extraction capability. However, convolution operations cannot effectively capture the long-range dependencies and repeatedly stacking convolutional layers to pursue a hierarchical structure can only make this problem alleviated but not completely solved. Meantime, the appearance of Transformer happens to cope with this problem and provides an opportunity to capture long-distance dependencies between tokens. Although Transformer has been introduced into HSI classification field recently, most of these related works only focus on exploiting a single kind of spatial or spectral information and neglect to explore the optimal fusion method for these two different-level features. Therefore, to fully exploit the abundant spatial information and spectral correlations in HSIs in a highly effective and efficient way, we present the initial attempt to explore the Transformer architecture in a dual-branch manner and propose a novel bilateral classification network named Hyper-ES2T. Besides, the Aggregated Feature Enhancement Module is proposed for effective feature aggregation and further spatial- spectral feature enhancement. Furthermore, to tackle the problem of high computational costs brought by vanilla self-attention block in Transformer, we design the Efficient Multi-Head Self-Attention block, pursuing the trade-off between model accuracy and efficiency. The proposed Hyper-ES2T reaches new state-of-the-art performance and outperforms previous methods by a significant margin on four benchmark datasets for HSI classification, which demonstrates the powerful generalization ability and superior feature representation capability of our Hyper-ES2T. It can be anticipated that this work provides a novel insight to design network architecture based on Transformer with superior performance and great model efficiency, which may inspire more following research in this direction of HSI processing field. The source codes will be available at https://github.com/Wenxuan-1119/Hyper-ES2T.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据