4.5 Article

AppCiP: Energy-Efficient Approximate Convolution-in-Pixel Scheme for Neural Network Acceleration

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JETCAS.2023.3242167

Keywords

Convolution-in-pixel; approximate computing; convolutional neural network; CMOS image sensor

Ask authors/readers for more resources

This paper introduces an architecture called AppCiP that efficiently enables artificial intelligence on resource-limited sensing devices. AppCiP features instant and reconfigurable RGB to grayscale conversion, highly parallel analog convolution-in-pixel, and low-precision quinary weight neural networks. The circuit-to-application co-simulation results show that AppCiP achieves significantly higher power consumption efficiency compared to existing designs, with a frame rate of 3000 and an efficiency of 4.12 TOp/s/W. The performance accuracy of AppCiP on different datasets is evaluated and compared with state-of-the-art designs, showing the best results among other processing in/near pixel architectures with less than 1% degradation in accuracy.
Nowadays, always-on intelligent and self-powered visual perception systems have gained considerable attention and are widely used. However, capturing data and analyzing it via a backend/cloud processor are energy-intensive and long-latency, resulting in a memory bottleneck and low-speed feature extraction at the edge. This paper presents AppCiP architecture as a sensing and computing integration design to efficiently enable Artificial Intelligence (AI) on resource-limited sensing devices. AppCiP provides a number of unique capabilities, including instant and reconfigurable RGB to grayscale conversion, highly parallel analog convolution-in-pixel, and realizing low-precision quinary weight neural networks. These features significantly mitigate the overhead of analog-to-digital converters and analog buffers, leading to a considerable reduction in power consumption and area overhead. Our circuit-to-application co-simulation results demonstrate that AppCiP achieves similar to 3 orders of magnitude higher efficiency on power consumption compared with the fastest existing designs considering different CNN workloads. It reaches a frame rate of 3000 and an efficiency of similar to 4.12 TOp/s/W. The performance accuracy of the AppCiP architecture on different datasets such as SVHN, Pest, CIFAR-10, MHIST, and CBL Face detection is evaluated and compared with the state-of-the-art design. The obtained results exhibit the best results among other processing in/near pixel architectures, while AppCip only degrades the accuracy by less than 1% on average compared to the floating-point baseline.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available