Performance issues of parallel, scalable convolutional neural networks in deep learning

0Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this work, we investigate the performance issues in the parallel and scalable of Convolutional Neural Networks (CNNs). This will accelerate the training performance of CNN. In this paper we propose the parallel recognition using Compute Unified Device Architecture (CUDA) Technology and Message Passing Interface (MPI). We demonstrate scalability and performance that can be achieved on the GPU using CUDA framework where the computation-intensive tasks shifted on GPU. It compares result on GPU hardware architecture with the serial algorithm on CPU. The main novelty of our method is a new scalable CNN architecture that integrates a category hierarchy with deep CNN.

Cite

CITATION STYLE

APA

Chavan, U., & Kulkarni, D. (2018). Performance issues of parallel, scalable convolutional neural networks in deep learning. In Advances in Intelligent Systems and Computing (Vol. 810, pp. 333–343). Springer Verlag. https://doi.org/10.1007/978-981-13-1513-8_35

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free