In this work, we investigate the performance issues in the parallel and scalable of Convolutional Neural Networks (CNNs). This will accelerate the training performance of CNN. In this paper we propose the parallel recognition using Compute Unified Device Architecture (CUDA) Technology and Message Passing Interface (MPI). We demonstrate scalability and performance that can be achieved on the GPU using CUDA framework where the computation-intensive tasks shifted on GPU. It compares result on GPU hardware architecture with the serial algorithm on CPU. The main novelty of our method is a new scalable CNN architecture that integrates a category hierarchy with deep CNN.
CITATION STYLE
Chavan, U., & Kulkarni, D. (2018). Performance issues of parallel, scalable convolutional neural networks in deep learning. In Advances in Intelligent Systems and Computing (Vol. 810, pp. 333–343). Springer Verlag. https://doi.org/10.1007/978-981-13-1513-8_35
Mendeley helps you to discover research relevant for your work.