Abstract
Training a deep neural network (DNN) involves selecting a set of hyperparameters that define the network topology and influence the accuracy of the resulting network. Often, the goal is to maximize prediction accuracy on a given dataset. However, non-functional requirements of the trained network - such as inference speed, size, and energy consumption - can be very important as well. In this article, we aim to automate the process of selecting an appropriate DNN topology that fulfills both functional and non-functional requirements of the application. Specifically, we focus on tuning two important hyperparameters, depth and width, which together define the shape of the resulting network and directly affect its accuracy, speed, size, and energy consumption. To reduce the time needed to search the design space, we train a fraction of DNNs and build a model to predict the performances of the remaining ones.We are able to produce tuned ResNets, which are up to 4.22 times faster than original depth-scaled ResNets on a batch of 128 images while matching their accuracy.
Author supplied keywords
Cite
CITATION STYLE
Mammadli, R., Wolf, F., & Jannesari, A. (2019). The art of getting deep neural networks in shape. ACM Transactions on Architecture and Code Optimization, 15(4). https://doi.org/10.1145/3291053
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.