Tuning DNN Model Compression to Resource and Data Availability in Cooperative Training

0Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Model compression is a fundamental tool to execute machine learning (ML) tasks on the diverse set of devices populating current- and next-generation networks, thereby exploiting their resources and data. At the same time, how much and when to compress ML models are very complex decisions, as they have to jointly account for such aspects as the model being used, the resources (e.g., computational) and local datasets available at each node, as well as network latencies. In this work, we address the multi-dimensional problem of adapting the model compression, data selection, and node allocation decisions to each other: our objective is to perform the DNN training at the minimum energy cost, subject to learning quality and time constraints. To this end, we propose an algorithmic framework called PACT, combining a time-expanded graph representation of the training process, a dynamic programming solution strategy, and a data-driven approach to the estimation of the loss evolution. We prove that PACT's complexity is polynomial, and its decisions can get arbitrarily close to the optimum. Through our numerical evaluation, we further show how PACT can consistently outperform state-of-the-art alternatives and closely matches the optimal energy consumption.

Cite

CITATION STYLE

APA

Malandrino, F., Giacomo, G. D., Karamzade, A., Levorato, M., & Chiasserini, C. F. (2024). Tuning DNN Model Compression to Resource and Data Availability in Cooperative Training. IEEE/ACM Transactions on Networking, 32(2), 1600–1615. https://doi.org/10.1109/TNET.2023.3323023

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free