DSCU: Accelerating CNN Inference in FPGAs with Dual Sizes of Compute Unit

3Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.

Abstract

FPGA-based accelerators have shown great potential in improving the performance of CNN inference. However, the existing FPGA-based approaches suffer from a low compute unit (CU) efficiency due to their large number of redundant computations, thus leading to high levels of performance degradation. In this paper, we show that no single CU can perform best across all the convolutional layers (CONV-layers). To this end, we propose the use of dual sizes of compute unit (DSCU), an approach that aims to accelerate CNN inference in FPGAs. The key idea of DSCU is to select the best combination of CUs via dynamic programming scheduling for each CONV-layer and then assemble each CONV-layer combination into a computing solution for the given CNN to deploy in FPGAs. The experimental results show that DSCU can achieve a performance density of 3.36 × 10−3 GOPs/slice on a Xilinx Zynq ZU3EG, which is 4.29 times higher than that achieved by other approaches.

Cite

CITATION STYLE

APA

Bao, Z., Guo, J., Zhang, W., & Dang, H. (2022). DSCU: Accelerating CNN Inference in FPGAs with Dual Sizes of Compute Unit. Journal of Low Power Electronics and Applications, 12(1). https://doi.org/10.3390/jlpea12010011

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free