HLHLp: Quantized neural networks training for reaching flat minima in loss surface

4Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.

Abstract

Quantization of deep neural networks is extremely essential for efficient implementations. Low-precision networks are typically designed to represent original floating-point counterparts with high fidelity, and several elaborate quantization algorithms have been developed. We propose a novel training scheme for quantized neural networks to reach flat minima in the loss surface with the aid of quantization noise. The proposed training scheme employs high-low-high-low precision in an alternating manner for network training. The learning rate is also abruptly changed at each stage for coarse- or fine-tuning. With the proposed training technique, we show quite good performance improvements for convolutional neural networks when compared to the previous fine-tuning based quantization scheme. We achieve the state-of-the-art results for recurrent neural network based language modeling with 2-bit weight and activation.

Cite

CITATION STYLE

APA

Shin, S., Park, J., Boo, Y., & Sung, W. (2020). HLHLp: Quantized neural networks training for reaching flat minima in loss surface. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 5784–5791). AAAI press. https://doi.org/10.1609/aaai.v34i04.6035

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free