4.6 Article

KAConv: Kernel attention convolutions

期刊

NEUROCOMPUTING
卷 514, 期 -, 页码 477-485

出版社

ELSEVIER
DOI: 10.1016/j.neucom.2022.10.017

关键词

Convolutional neural networks; Attention mechanism; Convolutional kernel; Deep learning

资金

  1. National Key Research and Development Program of China [2018AAA0100500]
  2. National Nature Science Foundation of China [62273150]
  3. Shanghai Natural Science Foundation [22ZR1421000]
  4. Science and Technology Commission of Shanghai Municipality [21XD1430600]
  5. Fundamental Research Funds for the Central Universities
  6. Shanghai Outstanding Academic Leaders Plan [22DZ2229004]

向作者/读者索取更多资源

The study introduces a novel attention convolution method called Kernel Attention Convolution (KAConv) to enhance the flexibility of convolution. By embedding attention into the convolution kernel, KAConv generates different attention weights to dynamically adjust the parameters of convolution kernels, improving the flexibility of convolution. Experiment results demonstrate that KAConv outperforms existing attention mechanism-based methods in the ImageNet-1K benchmark.
Most of the previous network optimization works applied attention mechanism to feature maps, but neglected to embed attention into convolution kernel of the end-to-end network that is convenient to deploy. To address this issue, we present a novel attention convolution method named Kernel Attention Convolution (KAConv) to enhance the flexibility of convolution. The proposed KAConv gener-ates different attention weights for different spatial positions of convolution kernels based on the input features, so as to dynamically adjust the parameters of convolution kernels during the forward propaga-tion to enhance the flexibility of convolution. We decompose the convolution kernels into subkernels spatially, and generate the corresponding feature maps through which attention weights are obtained. The final refined feature maps are aggregated by the attention weighted feature maps corresponding to each subkernel. KAConv is a computationally lightweight convolution method, which not only incor-porates attention into kernels but also enhances informative representations. By replacing the standard convolution with the proposed KAConv in convolutional neural networks (CNNs), the networks yield sig-nificant performance improvement. Extensive experiments on the ImageNet-1K benchmark demonstrate that KAConv outperforms existing attention mechanism-based methods. We also carry out experiments on the MS COCO and PASCAL VOC datasets to show the generalization ability of our method.(c) 2022 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据