High-end mobile platforms rapidly serve as primary computing devices for a wide range of Deep Neural Network (DNN) applications. However, the constrained computation and storage resources on these devices still pose significant challenges for real-time DNN inference executions. To address this problem, we propose a set of hardware-friendly structured model pruning and compiler optimization techniques to accelerate DNN executions on mobile devices. This demo shows that these optimizations can enable real-time mobile execution of multiple DNN applications, including style transfer, DNN coloring and super resolution.
CITATION STYLE
Niu, W., Zhao, P., Zhan, Z., Lin, X., Wang, Y., & Ren, B. (2020). Towards real-time DNN inference on mobile platforms with model pruning and compiler optimization. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2021-January, pp. 5306–5308). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2020/778
Mendeley helps you to discover research relevant for your work.