4.7 Article

Bimodal-Distributed Binarized Neural Networks

Journal

MATHEMATICS
Volume 10, Issue 21, Pages -

Publisher

MDPI
DOI: 10.3390/math10214107

Keywords

convolutional neural networks; binarization; quantization; efficient inference deployment

Categories

Funding

  1. HIROSHI FUJIWARA Cyber Security Research Center, Technion

Ask authors/readers for more resources

This paper proposes a new binarization method that achieves a bimodal distribution of network weights through kurtosis regularization. Preserving the weight distribution during binarization training significantly reduces the generalization error of the BNN. Experimental results demonstrate that this new method outperforms current state-of-the-art schemes on CIFAR-10 and ImageNet.
Binary neural networks (BNNs) are an extremely promising method for reducing deep neural networks' complexity and power consumption significantly. Binarization techniques, however, suffer from ineligible performance degradation compared to their full-precision counterparts. Prior work mainly focused on strategies for sign function approximation during the forward and backward phases to reduce the quantization error during the binarization process. In this work, we propose a bimodal-distributed binarization method (BD-BNN). The newly proposed technique aims to impose a bimodal distribution of the network weights by kurtosis regularization. The proposed method consists of a teacher-trainer training scheme termed weight distribution mimicking (WDM), which efficiently imitates the full-precision network weight distribution to their binary counterpart. Preserving this distribution during binarization-aware training creates robust and informative binary feature maps and thus it can significantly reduce the generalization error of the BNN. Extensive evaluations on CIFAR-10 and ImageNet demonstrate that our newly proposed BD-BNN outperforms current state-of-the-art schemes.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available