4.3 Article

An Overhead-Free Max-Pooling Method for SNN

期刊

IEEE EMBEDDED SYSTEMS LETTERS
卷 12, 期 1, 页码 21-24

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/LES.2019.2919244

关键词

Neurons; Microsoft Windows; Mathematical model; Training; Computational modeling; Biological neural networks; Task analysis; Approximate; max pooling; overhead; spiking neural network (SNN)

资金

  1. National S&T Major Project of China [2017ZX01028103]
  2. National Natural Science Foundation of China [61802427]

向作者/读者索取更多资源

Spiking neural networks (SNNs) have been shown to be accurate, fast, and efficient in classical machine vision tasks, such as object recognition or detection. It is typical to convert a pretrained deep neural network into an SNN since training SNN is not easy. The max-pooling (MP) function is widely adopted in most state-of-the-art deep neural networks. To maintain the accuracy of the SNN obtained through conversion, this function is an important element to be implemented. However, it is difficult due to the dynamic characteristics of spikes. As far as we know, existing solutions adopt additional technologies except the spiking neuron model to implement MP or approximate MP, which introduce overhead of memory storage and computation. In this letter, we propose a novel method that utilizes only the spiking neuron model to approximate MP. Our method does not incur any overhead. We validate our method with three datasets and six networks including three oxford visual geometry group-like networks. And the experimental results show that the performance (accuracy and convergence rate) of our method is as good as or even better than that of the existing method.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.3
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据