4.5 Article

LUTNet: Learning FPGA Configurations for Highly Efficient Neural Network Inference

期刊

IEEE TRANSACTIONS ON COMPUTERS
卷 69, 期 12, 页码 1795-1808

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TC.2020.2978817

关键词

Deep neural network; hardware architecture; field-programmable gate array; lookup table

资金

  1. United Kingdom EPSRC [EP/P010040/1]
  2. Imagination Technologies
  3. Royal Academy of Engineering
  4. EPSRC [EP/P010040/1, EP/S030069/1] Funding Source: UKRI

向作者/读者索取更多资源

Research has shown that deep neural networks contain significant redundancy, and thus that high classification accuracy can be achieved even when weights and activations are quantized down to binary values. Network binarization on FPGAs greatly increases area efficiency by replacing resource-hungry multipliers with lightweight XNOR gates. However, an FPGA's fundamental building block, the K-LUT, is capable of implementing far more than an XNOR: it can perform any K-input Boolean operation. Inspired by this observation, we propose LUTNet, an end-to-end hardware-software framework for the construction of area-efficient FPGA-based neural network accelerators using the native LUTs as inference operators. We describe the realization of both unrolled and tiled LUTNet architectures, with the latter facilitating smaller, less power-hungry deployment over the former while sacrificing area and energy efficiency along with throughput. For both varieties, we demonstrate that the exploitation of LUT flexibility allows for far heavier pruning than possible in prior works, resulting in significant area savings while achieving comparable accuracy. Against the state-of-the-art binarized neural network implementation, we achieve up to twice the area efficiency for several standard network models when inferencing popular datasets. We also demonstrate that even greater energy efficiency improvements are obtainable.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据