Accelerating training of deep neural networks via sparse edge processing

10Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We propose a reconfigurable hardware architecture for deep neural networks (DNNs) capable of online training and inference, which uses algorithmically pre-determined, structured sparsity to significantly lower memory and computational requirements. This novel architecture introduces the notion of edge-processing to provide flexibility and combines junction pipelining and operational parallelization to speed up training. The overall effect is to reduce network complexity by factors up to 30x and training time by up to 35x relative to GPUs, while maintaining high fidelity of inference results. This has the potential to enable extensive parameter searches and development of the largely unexplored theoretical foundation of DNNs. The architecture automatically adapts itself to different network sizes given available hardware resources. As proof of concept, we show results obtained for different bit widths.

Cite

CITATION STYLE

APA

Dey, S., Shao, Y., Chugg, K. M., & Beerel, P. A. (2017). Accelerating training of deep neural networks via sparse edge processing. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 10613 LNCS, pp. 273–280). Springer Verlag. https://doi.org/10.1007/978-3-319-68600-4_32

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free