Exploring deep reuse in winograd CNN inference

9Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Convolutional neural networks (CNNs), as representatives of deep learning, are one of the most commonly used neural networks in applications such as graphic image analysis. However, CNN has heavy computation patterns; network training processes could take several hours even with modern processors. Different from the training process, the inference process is more often executed on devices with low computing power, such as CPUs. Fortunately, a minimal filtering algorithm, Winograd, can reduce the convolution computations by reducing the number of multiplication operations. We find that the Winograd convolution can be further accelerated by reusing the similar data and computation patterns, which is called deep reuse.

Author supplied keywords

Cite

CITATION STYLE

APA

Wu, R., Zhang, F., Zheng, Z., Du, X., & Shen, X. (2021). Exploring deep reuse in winograd CNN inference. In Proceedings of the ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming, PPOPP (pp. 483–484). Association for Computing Machinery. https://doi.org/10.1145/3437801.3441588

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free