Towards real-time DNN inference on mobile platforms with model pruning and compiler optimization

5Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

Abstract

High-end mobile platforms rapidly serve as primary computing devices for a wide range of Deep Neural Network (DNN) applications. However, the constrained computation and storage resources on these devices still pose significant challenges for real-time DNN inference executions. To address this problem, we propose a set of hardware-friendly structured model pruning and compiler optimization techniques to accelerate DNN executions on mobile devices. This demo shows that these optimizations can enable real-time mobile execution of multiple DNN applications, including style transfer, DNN coloring and super resolution.

Cite

CITATION STYLE

APA

Niu, W., Zhao, P., Zhan, Z., Lin, X., Wang, Y., & Ren, B. (2020). Towards real-time DNN inference on mobile platforms with model pruning and compiler optimization. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2021-January, pp. 5306–5308). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2020/778

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free