期刊
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING
卷 11, 期 2, 页码 358-372出版社
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TETC.2023.3237914
关键词
Edge artificial intelligence; in-memory computing; hardware/software co-design; convolutional neural networks; low-power software optimization
Bit-line Computing (BC) architectures enable parallel execution of bit-wise operations in-memory, resulting in efficient arithmetic operations at the array periphery. This paradigm offers new opportunities for edge AI with its inherent parallelism and energy efficiency. This manuscript proposes a framework that leverages BC optimizations to enable high parallelism and aggressive compression of AI models, resulting in significant energy savings compared to state-of-the-art BC computing approaches.
By supporting the access of multiple memory words at the same time, Bit-line Computing (BC) architectures allow the parallel execution of bit-wise operations in-memory. At the array periphery, arithmetic operations are then derived with little additional overhead. Such a paradigm opens novel opportunities for Artificial Intelligence (AI) at the edge, thanks to the massive parallelism inherent in memory arrays and the extreme energy efficiency of computing in-situ, hence avoiding data transfers. Previous works have shown that BC brings disruptive efficiency gains when targeting AI workloads, a key metric in the context of emerging edge AI scenarios. This manuscript builds on these findings by proposing an end-to-end framework that leverages BC-specific optimizations to enable high parallelism and aggressive compression of AI models. Our approach is supported by a novel hardware module performing real-time decoding, as well as new algorithms to enable BC-friendly model compression. Our hardware/software approach results in a 91% energy savings (for a 1% accuracy degradation constraint) regarding state-of-the-art BC computing approaches.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据