Benchmarking Deep Learning Infrastructures by Means of TensorFlow and Containers

5Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Ever growing interest and usage of deep learning rises a question on the performance of various infrastructures suitable for training of neural networks. We present here our approach and first results of tests performed with TensorFlow Benchmarks which use best practices for multi-GPU and distributed training. We pack the Benchmarks in Docker containers and execute them by means of uDocker and Singularity container tools on a single machine and in the HPC environment. The Benchmarks comprise a number of convolutional neural network models run across synthetic data and e.g. the ImageNet dataset. For the same Nvidia K80 GPU card we achieve the same performance in terms of processed images per second and similar scalability between 1-2-4 GPUs as presented by the TensorFlow developers. We therefore do not obtain statistically significant overhead due to the usage of containers in the multi-GPU case, and the approach of using TF Benchmarks in a Docker container can be applied across various systems.

Cite

CITATION STYLE

APA

Grupp, A., Kozlov, V., Campos, I., David, M., Gomes, J., & López García, Á. (2019). Benchmarking Deep Learning Infrastructures by Means of TensorFlow and Containers. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11887 LNCS, pp. 478–489). Springer. https://doi.org/10.1007/978-3-030-34356-9_36

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free