Journal
ACM COMPUTING SURVEYS
Volume 55, Issue 4, Pages -Publisher
ASSOC COMPUTING MACHINERY
DOI: 10.1145/3524500
Keywords
Deep Neural Networks; Convolutional Neural Networks; Neural Architecture Search; Hardware-Aware Neural Architecture Search; general purpose hardware; domain specific accelerators; quantization; accelerator network co-search; literature review; survey; CPU; GPU; ASIC; FPGA
Categories
Ask authors/readers for more resources
In this article, we address the issue of automating the hardware-aware architectural design process of Deep Neural Networks (DNNs). The advancements in Convolutional Neural Network (CNN) algorithm design have had a significant impact in fields such as computer vision, virtual reality, and autonomous driving. However, the end-to-end design process of a CNN is a complex and time-consuming task that requires expertise in various areas. The HW-NAS approach automates the architectural design process of DNNs to improve efficiency and achieve acceptable accuracy-performance tradeoffs, considering different hardware platforms.
We review the problem of automating hardware-aware architectural design process of Deep Neural Networks (DNNs). The field of Convolutional Neural Network (CNN) algorithm design has led to advancements in many fields, such as computer vision, virtual reality, and autonomous driving. The end-to-end design process of a CNN is a challenging and time-consuming task, as it requires expertise in multiple areas such as signal and image processing, neural networks, and optimization. At the same time, several hardware platforms, general-and special-purpose, have equally contributed to the training and deployment of these complex networks in a different setting. Hardware-Aware Neural Architecture Search (HW-NAS) automates the architectural design process of DNNs to alleviate human effort and generate efficient models accomplishing acceptable accuracy-performance tradeoffs. The goal of this article is to provide insights and understanding of HW-NAS techniques for various hardware platforms (MCU, CPU, GPU, ASIC, FPGA, ReRAM, DSP, and VPU), followed by the co-search methodologies of neural algorithm and hardware accelerator specifications.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available