3.8 Proceedings Paper

Channel-wise Mixed-precision Assignment for DNN Inference on Constrained Edge Nodes

出版社

IEEE

关键词

Deep Learning; NAS; Quantization; TinyML

资金

  1. ECSEL Joint Undertaking (JU) [101007321]
  2. European Union, France
  3. European Union, Belgium
  4. European Union, Czech Republic
  5. European Union, Germany
  6. European Union, Italy
  7. European Union, Sweden
  8. European Union, Switzerland
  9. European Union, Turkey

向作者/读者索取更多资源

Quantization is a widely used method to reduce memory occupation, latency, and energy consumption of deep neural networks in cloud and edge systems. This study proposes a novel neural architecture search method that allows higher precision assignment based on important features, resulting in reduced memory and energy consumption while maintaining the same accuracy.
Quantization is widely employed in both cloud and edge systems to reduce the memory occupation, latency, and energy consumption of deep neural networks. In particular, mixed-precision quantization, i.e., the use of different bit-widths for different portions of the network, has been shown to provide excellent efficiency gains with limited accuracy drops, especially with optimized bit-width assignments determined by automated Neural Architecture Search (NAS) tools. State-of-the-art mixed-precision works layer-wise, i.e., it uses different bit-widths for the weights and activations tensors of each network layer. In this work, we widen the search space, proposing a novel NAS that selects the bit-width of each weight tensor channel independently. This gives the tool the additional flexibility of assigning a higher precision only to the weights associated with the most informative features. Testing on the MLPerf Tiny benchmark suite, we obtain a rich collection of Pareto-optimal models in the accuracy vs model size and accuracy vs energy spaces. When deployed on the MPIC RISC-V edge processor, our networks reduce the memory and energy for inference by up to 63% and 27% respectively compared to a layer-wise approach, for the same accuracy.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据