PROFIT: A Novel Training Method for sub-4-bit MobileNet Models

32Citations
Citations of this article
74Readers
Mendeley users who have this article in their library.
Get full text

Abstract

4-bit and lower precision mobile models are required due to the ever-increasing demand for better energy efficiency in mobile devices. In this work, we report that the activation instability induced by weight quantization (AIWQ) is the key obstacle to sub-4-bit quantization of mobile networks. To alleviate the AIWQ problem, we propose a novel training method called PROgressive-Freezing Iterative Training (PROFIT), which attempts to freeze layers whose weights are affected by the instability problem stronger than the other layers. We also propose a differentiable and unified quantization method (DuQ) and a negative padding idea to support asymmetric activation functions such as h-swish. We evaluate the proposed methods by quantizing MobileNet-v1, v2, and v3 on ImageNet and report that 4-bit quantization offers comparable (within 1.48% top-1 accuracy) accuracy to full precision baseline. In the ablation study of the 3-bit quantization of MobileNet-v3, our proposed method outperforms the state-of-the-art method by a large margin, 12.86% of top-1 accuracy. The quantized model and source code is available at https://github.com/EunhyeokPark/PROFIT.

Cite

CITATION STYLE

APA

Park, E., & Yoo, S. (2020). PROFIT: A Novel Training Method for sub-4-bit MobileNet Models. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12351 LNCS, pp. 430–446). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-58539-6_26

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free