The art of getting deep neural networks in shape

9Citations
Citations of this article
38Readers
Mendeley users who have this article in their library.

Abstract

Training a deep neural network (DNN) involves selecting a set of hyperparameters that define the network topology and influence the accuracy of the resulting network. Often, the goal is to maximize prediction accuracy on a given dataset. However, non-functional requirements of the trained network - such as inference speed, size, and energy consumption - can be very important as well. In this article, we aim to automate the process of selecting an appropriate DNN topology that fulfills both functional and non-functional requirements of the application. Specifically, we focus on tuning two important hyperparameters, depth and width, which together define the shape of the resulting network and directly affect its accuracy, speed, size, and energy consumption. To reduce the time needed to search the design space, we train a fraction of DNNs and build a model to predict the performances of the remaining ones.We are able to produce tuned ResNets, which are up to 4.22 times faster than original depth-scaled ResNets on a batch of 128 images while matching their accuracy.

Cite

CITATION STYLE

APA

Mammadli, R., Wolf, F., & Jannesari, A. (2019). The art of getting deep neural networks in shape. ACM Transactions on Architecture and Code Optimization, 15(4). https://doi.org/10.1145/3291053

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free