相关参考文献
注意:仅列出部分参考文献,下载原文获取全部文献信息。Recurrent Neural Networks With Column-Wise Matrix-Vector Multiplication on FPGAs
Zhiqiang Que et al.
IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS (2022)
A low-latency LSTM accelerator using balanced sparsity based on FPGA
Jingfei Jiang et al.
MICROPROCESSORS AND MICROSYSTEMS (2022)
When Massive GPU Parallelism Ain't Enough: A Novel Hardware Architecture of 2D-LSTM Neural Network
Vladimir Rybalkin et al.
ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS (2022)
GBC: An Energy-Efficient LSTM Accelerator With Gating Units Level Balanced Compression Strategy
Bi Wu et al.
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS (2022)
PermLSTM: A High Energy-Efficiency LSTM Accelerator Architecture
Yong Zheng et al.
ELECTRONICS (2021)
Specializing FGPU for Persistent Deep Learning
Rui Ma et al.
ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS (2021)
AERO: A 1.28 MOP/s/LUT Reconfigurable Inference Processor for Recurrent Neural Networks in a Resource-Limited FPGA
Jinwon Kim et al.
ELECTRONICS (2021)
POLAR: A Pipelined/Overlapped FPGA-Based LSTM Accelerator
Erfan Bank-Tavakoli et al.
IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS (2020)
An 8.93 TOPS/W LSTM Recurrent Neural Network Accelerator Featuring Hierarchical Coarse-Grain Sparsity for On-Device Speech Recognition
Deepak Kadetotad et al.
IEEE JOURNAL OF SOLID-STATE CIRCUITS (2020)
Efficient Hardware Architectures for 1D-and MD-LSTM Networks
Vladimir Rybalkin et al.
JOURNAL OF SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY (2020)
Mapping Large LSTMs to FPGAs with Weight Reuse
Zhiqiang Que et al.
JOURNAL OF SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY (2020)
EdgeDRNN: Recurrent Neural Network Accelerator for Edge Inference
Chang Gao et al.
IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS (2020)
Approximate LSTM Computing for Energy-Efficient Speech Recognition
Junseo Jo et al.
ELECTRONICS (2020)
Beyond Peak Performance: Comparing the Real Performance of AI-Optimized FPGAs and GPUs
Andrew Boutros et al.
2020 INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE TECHNOLOGY (ICFPT 2020) (2020)
Achieving Full Parallelism in LSTM via a Unified Accelerator Design
Xinyi Zhang et al.
2020 IEEE 38TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2020) (2020)
E-LSTM: An Efficient Hardware Architecture for Long Short-Term Memory
Meiqi Wang et al.
IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS (2019)
Efficient and Effective Sparse LSTM on FPGA with Bank-Balanced Sparsity
Shijie Cao et al.
PROCEEDINGS OF THE 2019 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS (FPGA'19) (2019)
Why Compete When You Can Work Together: FPGA-ASIC Integration for Persistent RNNs
Eriko Nurvitadhi et al.
2019 27TH IEEE ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM) (2019)
Sequence classification for credit-card fraud detection
Johannes Jurgovsky et al.
EXPERT SYSTEMS WITH APPLICATIONS (2018)
LSTM: A Search Space Odyssey
Klaus Greff et al.
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS (2017)
DNPU: An 8.1TOPS/W Reconfigurable CNN-RNN Processor for General-Purpose Deep Neural Networks
Dongjoo Shin et al.
2017 IEEE INTERNATIONAL SOLID-STATE CIRCUITS CONFERENCE (ISSCC) (2017)
Channel Pruning for Accelerating Very Deep Neural Networks
Yihui He et al.
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV) (2017)
ESE: Efficient Speech Recognition Engine with Sparse LSTM on FPGA
Song Han et al.
FPGA'17: PROCEEDINGS OF THE 2017 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS (2017)
A Novel Connectionist System for Unconstrained Handwriting Recognition
Alex Graves et al.
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE (2009)