DWM: A decomposablewinograd method for convolution acceleration

27Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

Winograd's minimal filtering algorithm has been widely used in Convolutional Neural Networks (CNNs) to reduce the number of multiplications for faster processing. However, it is only effective on convolutions with kernel size as 3x3 and stride as 1, because it suffers from significantly increased FLOPs and numerical accuracy problem for kernel size larger than 3x3 and fails on convolution with stride larger than 1. In this paper, we propose a novel Decomposable Winograd Method (DWM), which breaks through the limitation of original Winograd's minimal filtering algorithm to a wide and general convolutions. DWM decomposes kernels with large size or large stride to several small kernels with stride as 1 for further applying Winograd method, so that DWM can reduce the number of multiplications while keeping the numerical accuracy. It enables the fast exploring of larger kernel size and larger stride value in CNNs for high performance and accuracy and even the potential for new CNNs. Comparing against the original Winograd, the proposed DWM is able to support all kinds of convolutions with a speedup of ∼2, without affecting the numerical accuracy.

Cite

CITATION STYLE

APA

Huang, D., Zhang, X., Zhang, R., Zhi, T., He, D., Guo, J., … Chen, Y. (2020). DWM: A decomposablewinograd method for convolution acceleration. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 4174–4181). AAAI press. https://doi.org/10.1609/aaai.v34i04.5838

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free