4.3 Article

An efficient pruning scheme of deep neural networks for Internet of Things applications

出版社

SPRINGER
DOI: 10.1186/s13634-021-00744-4

关键词

Deep neural networks; Deep learning; Internet of Things; Resource-limited edge computing; Pruning; Efficiency

资金

  1. National Key RAMP
  2. D Program of China [2020YFB1804804]
  3. National Natural Science Foundation of China [62071425, 61731002]
  4. Zhejiang Key Research and Development Plan [2019C01002, 2019C03131]
  5. Huawei Cooperation Project
  6. Zhejiang Lab [2019LC0AB01]
  7. Zhejiang Provincial Natural Science Foundation of China [LY20F010016]

向作者/读者索取更多资源

The paper proposes a novel pruning-based paradigm to reduce the computational cost of DNNs while maintaining their expressive capability. The approach is evaluated on various benchmark datasets and compared with typical advanced CNN architectures, demonstrating superior performance and effectiveness.
Nowadays, deep neural networks (DNNs) have been rapidly deployed to realize a number of functionalities like sensing, imaging, classification, recognition, etc. However, the computational-intensive requirement of DNNs makes it difficult to be applicable for resource-limited Internet of Things (IoT) devices. In this paper, we propose a novel pruning-based paradigm that aims to reduce the computational cost of DNNs, by uncovering a more compact structure and learning the effective weights therein, on the basis of not compromising the expressive capability of DNNs. In particular, our algorithm can achieve efficient end-to-end training that transfers a redundant neural network to a compact one with a specifically targeted compression rate directly. We comprehensively evaluate our approach on various representative benchmark datasets and compared with typical advanced convolutional neural network (CNN) architectures. The experimental results verify the superior performance and robust effectiveness of our scheme. For example, when pruning VGG on CIFAR-10, our proposed scheme is able to significantly reduce its FLOPs (floating-point operations) and number of parameters with a proportion of 76.2% and 94.1%, respectively, while still maintaining a satisfactory accuracy. To sum up, our scheme could facilitate the integration of DNNs into the common machine-learning-based IoT framework and establish distributed training of neural networks in both cloud and edge.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.3
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据