Enabling energy-efficient DNN training on hybrid GPU-FPGA accelerators

15Citations
Citations of this article
18Readers
Mendeley users who have this article in their library.
Get full text

Abstract

DNN training consumes orders of magnitude more energy than inference and requires innovative use of accelerators to improve energy-efficiency. However, despite having complementary features, GPUs and FPGAs have been mostly used independently for the entire training process, thus neglecting the opportunity in assigning individual but distinct operations to the most suitable hardware. In this paper, we take the initiative to explore new opportunities and viable solutions in enabling energy-efficient DNN training on hybrid accelerators. To overcome fundamental challenges including avoiding training throughput loss, enabling fast design space exploration, and efficient scheduling, we propose a comprehensive framework, Hype-training, that utilizes a combination of offline characterization, performance modeling, and online scheduling of individual operations. Experimental tests using NVIDIA V100 GPUs and Intel Stratix 10 FPGAs show that, Hype-training is able to exploit a mixture of GPUs and FPGAs at a fine granularity to achieve significant energy reduction, by 44.3% on average and up to 59.7%, without any loss in training throughput. Hype-training can also enforce power caps more effectively than state-of-the-art power management mechanisms on GPUs.

Cite

CITATION STYLE

APA

He, X., Liu, J., Xie, Z., Chen, H., Chen, G., Zhang, W., & Li, D. (2021). Enabling energy-efficient DNN training on hybrid GPU-FPGA accelerators. In Proceedings of the International Conference on Supercomputing (pp. 227–241). Association for Computing Machinery. https://doi.org/10.1145/3447818.3460371

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free