Energy Efficient Boosting of GEMM Accelerators for DNN via Reuse

6Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.

Abstract

Reuse-centric convolutional neural networks (CNN) acceleration speeds up CNN inference by reusing computations for similar neuron vectors in CNN's input layer or activation maps. This new paradigm of optimizations is, however, largely limited by the overheads in neuron vector similarity detection, an important step in reuse-centric CNN. This article presents an in-depth exploration of architectural support for reuse-centric CNN. It addresses some major limitations of the state-of-the-art design and proposes a novel hardware accelerator that improves neuron vector similarity detection and reduces the energy consumption of reuse-centric CNN inference. The accelerator is implemented to support a wide variety of neural network settings with a banked memory subsystem. Design exploration is performed through RTL simulation and synthesis on an FPGA platform. When integrated into Eyeriss, the accelerator can potentially provide improvements up to 7.75 in performance. Furthermore, it can reduce the energy used for similarity detection up to 95.46%, and it can accelerate the convolutional layer up to 3.63 compared to the software-based implementation running on the CPU.

Cite

CITATION STYLE

APA

Cicek, N. M., Shen, X., & Ozturk, O. (2022). Energy Efficient Boosting of GEMM Accelerators for DNN via Reuse. ACM Transactions on Design Automation of Electronic Systems, 27(5). https://doi.org/10.1145/3503469

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free