Related references
Note: Only part of the references are listed.The Decline of Computers as a General Purpose Technology
Neil C. Thompson et al.
COMMUNICATIONS OF THE ACM (2021)
Kunlun: A 14nm High-Performance AI Processor for Diversified Workloads
Jian Ouyang et al.
2021 IEEE INTERNATIONAL SOLID-STATE CIRCUITS CONFERENCE (ISSCC) (2021)
A Survey of Accelerator Architectures for Deep Neural Networks
Yiran Chen et al.
ENGINEERING (2020)
Compute Solution for Tesla's Full Self-Driving Computer
Emil Talpes et al.
IEEE MICRO (2020)
Habana Labs Purpose-Built AI Inference and Training Processor Architectures: Scaling AI Training Systems Using Standard Ethernet With Gaudi Processor
Eitan Medina et al.
IEEE MICRO (2020)
A Domain-Specific Supercomputer for Training Deep Neural Networks
Norman P. Jouppi et al.
COMMUNICATIONS OF THE ACM (2020)
Domain-Specific Hardware Accelerators
William J. Dally et al.
COMMUNICATIONS OF THE ACM (2020)
There's plenty of room at the Top: What will drive computer performance after Moore's law?
Charles E. Leiserson et al.
SCIENCE (2020)
High-Performance Deep-Learning Coprocessor Integrated into x86 SoC with Server-Class CPUs
Glenn Henry et al.
2020 ACM/IEEE 47TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA 2020) (2020)
A New Golden Age for Computer Architecture
John L. Hennessy et al.
COMMUNICATIONS OF THE ACM (2019)
Deep Neural Network Approximation for Custom Hardware: Where We've Been, Where We're Going
Erwei Wang et al.
ACM COMPUTING SURVEYS (2019)
Towards artificial general intelligence with hybrid Tianjic chip architecture
Jing Pei et al.
NATURE (2019)
Digital Neuromorphic Chips for Deep Learning Inference: A Comprehensive Study
Hamed F. Langroudi et al.
APPLICATIONS OF MACHINE LEARNING (2019)
A Domain-Specific Architecture for Deep Neural Networks
Norman P. Jouppi et al.
COMMUNICATIONS OF THE ACM (2018)
DLA: Compiler and FPGA Overlay for Neural Network Inference Acceleration
Mohamed S. Abdelfattah et al.
2018 28TH INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS (FPL) (2018)
Eyeriss: An Energy-Efficient Reconfigurable Accelerator for Deep Convolutional Neural Networks
Yu-Hsin Chen et al.
IEEE JOURNAL OF SOLID-STATE CIRCUITS (2017)
Efficient Processing of Deep Neural Networks: A Tutorial and Survey
Vivienne Sze et al.
PROCEEDINGS OF THE IEEE (2017)
The End of Moore's Law: A New Beginning for Information Technology
Thomas N. Theis et al.
COMPUTING IN SCIENCE & ENGINEERING (2017)
DianNao Family: Energy-Efficient Hardware Accelerators for Machine Learning
Yunji Chen et al.
COMMUNICATIONS OF THE ACM (2016)
Convolutional networks for fast, energy-efficient neuromorphic computing
Steven K. Esser et al.
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA (2016)
True North: Design and Tool Flow of a 65 mW 1 Million Neuron Programmable Neurosynaptic Chip
Filipp Akopyan et al.
IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS (2015)
Artificial neural networks in hardware A survey of two decades of progress
Janardari Misra et al.
NEUROCOMPUTING (2010)
Roofline: An Insightful Visual Performance Model for Multicore Architectures
Samuel Williams et al.
COMMUNICATIONS OF THE ACM (2009)