4.6 Article

UPANets: Learning from the Universal Pixel Attention Neworks

期刊

ENTROPY
卷 24, 期 9, 页码 -

出版社

MDPI
DOI: 10.3390/e24091243

关键词

computer vision; image classification; CNN; attention

资金

  1. Taiwan Ministry of Science and Technology [MOST 111-2410-H-A49-019]

向作者/读者索取更多资源

This work proposes an efficient and robust backbone, UPANets, which utilizes channel and spatial direction attentions to expand the receptive fields in shallow convolutional layers. Experimental results show that UPANets achieve better performance with fewer resources on CIFAR-{10, 100} than existing state-of-the-art methods.
With the successful development in computer vision, building a deep convolutional neural network (CNNs) has been mainstream, considering the character of shared parameters in a convolutional layer. Stacking convolutional layers into a deep structure improves performance, but over-stacking also ramps up the needed resources for GPUs. Seeing another surge of Transformers in computer vision, the issue has aroused severely. A resource-hungry model is hardly implemented for limited hardware or single-customers-based GPU. Therefore, this work focuses on these concerns and proposes an efficient but robust backbone, which equips with channel and spatial direction attentions, so the attentions help to expand receptive fields in shallow convolutional layers and pass the information to every layer. An attention-boosted network based on already efficient CNNs, Universal Pixel Attention Networks (UPANets), is proposed. Through a series of experiments, UPANets fulfil the purposes of learning global information with less needed resources and outshine many existing SOTAs in CIFAR-{10, 100}.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据