4.8 Article

A compute-in-memory chip based on resistive random-access memory

期刊

NATURE
卷 608, 期 7923, 页码 504-+

出版社

NATURE PORTFOLIO
DOI: 10.1038/s41586-022-04992-8

关键词

-

资金

  1. NSF Expeditions in Computing (Penn State) [1317470]
  2. Office of Naval Research (Science of AI program)
  3. National Natural Science Foundation of China [61851404]
  4. Western Digital Corporation
  5. SRC JUMP ASCENT Center
  6. Stanford SystemX Alliance
  7. Stanford NMTRI
  8. Beijing Innovation Center for Future Chips
  9. Division of Computing and Communication Foundations
  10. Direct For Computer & Info Scie & Enginr [1317470] Funding Source: National Science Foundation

向作者/读者索取更多资源

This study presents NeuRRAM, a RRAM-based CIM chip that offers versatility, high energy efficiency, and accuracy. By co-optimizing algorithms, architecture, circuits, and devices, the chip can be reconfigured for different model architectures and provides twice the energy efficiency of previous state-of-the-art RRAM-CIM chips across various computational bit-precisions. It achieves inference accuracy comparable to software models quantized to four-bit weights across various AI tasks.
Realizing increasingly complex artificial intelligence (AI) functionalities directly on edge devices calls for unprecedented energy efficiency of edge hardware. Compute-in-memory (CIM) based on resistive random-access memory (RRAM)(1) promises to meet such demand by storing AI model weights in dense, analogue and non-volatile RRAM devices, and by performing AI computation directly within RRAM, thus eliminating power-hungry data movement between separate compute and memory(2-5). Although recent studies have demonstrated in-memory matrix-vector multiplication on fully integrated RRAM-CIM hardware(6-17), it remains a goal for a RRAM-CIM chip to simultaneously deliver high energy efficiency, versatility to support diverse models and software-comparable accuracy. Although efficiency, versatility and accuracy are all indispensable for broad adoption of the technology, the inter-related trade-offs among them cannot be addressed by isolated improvements on any single abstraction level of the design. Here, by co-optimizing across all hierarchies of the design from algorithms and architecture to circuits and devices, we present NeuRRAM-a RRAM-based CIM chip that simultaneously delivers versatility in reconfiguring CIM cores for diverse model architectures, energy efficiency that is two-times better than previous state-of-the-art RRAM-CIM chips across various computational bit-precisions, and inference accuracy comparable to software models quantized to four-bit weights across various AI tasks, including accuracy of 99.0 percent on MNIST18 and 85.7 percent on CIFAR-10(19) image classification, 84.7-percent accuracy on Google speech command recognition(20), and a 70-percent reduction in image-reconstruction error on a Bayesian image-recovery task.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据