Real-Time Neural Network Inference on ExtremelyWeak Devices: Agile Offloading with Explainable AI

58Citations
Citations of this article
33Readers
Mendeley users who have this article in their library.
Get full text

Abstract

With the wide adoption of AI applications, there is a pressing need of enabling real-Time neural network (NN) inference on small embedded devices, but deploying NNs and achieving high performance of NN inference on these small devices is challenging due to their extremely weak capabilities. Although NN partitioning and offloading can contribute to such deployment, they are incapable of minimizing the local costs at embedded devices. Instead, we suggest to address this challenge via agile NN offloading, which migrates the required computations in NN offloading from online inference to offline learning. In this paper, we present AgileNN, a new NN offloading technique that achieves real-Time NN inference on weak embedded devices by leveraging eXplainable AI techniques, so as to explicitly enforce feature sparsity during the training phase and minimize the online computation and communication costs. Experiment results show that AgileNN's inference latency is >6X lower than the existing schemes, ensuring that sensory data on embedded devices can be timely consumed. It also reduces the local device's resource consumption by >8X, without impairing the inference accuracy.

Cite

CITATION STYLE

APA

Huang, K., & Gao, W. (2022). Real-Time Neural Network Inference on ExtremelyWeak Devices: Agile Offloading with Explainable AI. In Proceedings of the Annual International Conference on Mobile Computing and Networking, MOBICOM (pp. 200–213). Association for Computing Machinery. https://doi.org/10.1145/3495243.3560551

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free