Power Efficiency Analysis of a Deep Learning Workload on an IBM “Minsky” Platform

1Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The rise of Deep Learning techniques has attracted special attention to GPUs usage for better performance of model computation. Most frameworks for Cognitive Computing include support to offload model training and inferencing to graphics hardware, and this is so common that GPU designers are reserving die area for special function units tailored to accelerating Deep Learning computation. Measuring the capability of a hardware platform to run these workloads is a major concern for vendors and consumers of this exponentially growing market. In a previous work [9] we analyzed the execution times of the Fathom AI workloads [2] in CPUs and CPUs+GPUs. In this work we measure the Fathom workloads in the POWER8-based “Minsky” [15] platform, profiling power consumption and energy efficiency in GPUs. We explore alternative forms of execution via GPU power and frequency capping with the aim of reducing Energy-to-Solution (ETS) and Energy-Delay-Product (EDP). We show important ETS savings of up to 27% with half of the workloads decreasing the EDP. We also expose the advantages of frequency capping with respect to power capping in NVIDIA GPUs.

Cite

CITATION STYLE

APA

Mazuecos Pérez, M. D., Seiler, N. G., Bederián, C. S., Wolovick, N., & Vega, A. J. (2019). Power Efficiency Analysis of a Deep Learning Workload on an IBM “Minsky” Platform. In Communications in Computer and Information Science (Vol. 979, pp. 255–262). Springer Verlag. https://doi.org/10.1007/978-3-030-16205-4_19

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free