Cross-layer CNN Approximations for Hardware Implementation

1Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Convolution Neural Networks (CNNs) are widely used for image classification and object detection applications. The deployment of these architectures in embedded applications is a great challenge. This challenge arises from CNNs’ high computation complexity that is required to be implemented on platforms with limited hardware resources like FPGA. Since these applications are inherently error-resilient, approximate computing (AC) offers an interesting trade-off between resource utilization and accuracy. In this paper, we study the impact on CNN performances when several approximation techniques are applied simultaneously. We focus on two of the widely used approximation techniques, namely quantization and pruning. Our experimental results showed that for CNN networks of different parameter sizes and 3% loss in accuracy, we can obtain up to 27.9%–47.2% reduction in computation complexity in terms of FLOPs for CIFAR-10 and MNIST datasets.

Author supplied keywords

Cite

CITATION STYLE

APA

Ali, K. M. A., Alouani, I., El Cadi, A. A., Ouarnoughi, H., & Niar, S. (2020). Cross-layer CNN Approximations for Hardware Implementation. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12083 LNCS, pp. 151–165). Springer. https://doi.org/10.1007/978-3-030-44534-8_12

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free