Understanding and optimizing packed neural network training for hyper-parameter tuning

9Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

As neural networks are increasingly employed in machine learning practice, how to efficiently share limited training resources among a diverse set of model training tasks becomes a crucial issue. To achieve better utilization of the shared resources, we explore the idea of jointly training multiple neural network models on a single GPU in this paper. We realize this idea by proposing a primitive, called pack. We further present a comprehensive empirical study of pack and end-to-end experiments that suggest significant improvements for hyperparameter tuning. The results suggest: (1) packing two models can bring up to 40% performance improvement over unpacked setups for a single training step and the improvement increases when packing more models; (2) the benefit of the pack primitive largely depends on a number of factors including memory capacity, chip architecture, neural network structure, and batch size; (3) there exists a trade-off between packing and unpacking when training multiple neural network models on limited resources; (4) a pack-aware Hyperband is up to 2.7X faster than the original Hyperband, with this improvement growing as memory size increases and subsequently the density of models packed.

Cite

CITATION STYLE

APA

Liu, R., Krishnan, S., Elmore, A. J., & Franklin, M. J. (2021). Understanding and optimizing packed neural network training for hyper-parameter tuning. In Proceedings of the 5th Workshop on Data Management for End-To-End Machine Learning, DEEM 2021 - In conjunction with the 2021 ACM SIGMOD/PODS Conference. Association for Computing Machinery, Inc. https://doi.org/10.1145/3462462.3468880

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free