Towards convolutional neural network compression via k-means cluster

0Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.

Abstract

With the rapid development of computer field, the computing power is no longer the bottleneck of the machine learning research. At present, deep learning technologies such as DNN (Deep Neural Networks) have been widely used in various fields such as speech recognition, smart driving, image recognition and so on. However, as the structure of DNN becomes more complex and the requirements of accuracy become higher, the network size is getting larger. The storage costs of such networks are high, which prohibits their usage in resource-constrained devices (embedded devices or mobile devices). In this paper, we propose a new compression method, compressing the DNN models without losing the accuracy by clustering the trained weights. Specifically, in our work, K-means is used to cluster the weights of fully connected layers. Then, we encode the clustered value as an index label. When we store the network model, we only store the index labels of weights. Compared with the original 32-bits weight, the index label is usually only three or four bits (based on the number of clusters), so as to achieve the purpose of compressing. Our compressing method achieves 11.4x compression rate on AlexNet network.

Cite

CITATION STYLE

APA

Chen, G., Ma, S., & Guo, Y. (2018). Towards convolutional neural network compression via k-means cluster. In Proceedings of 2018 the 8th International Workshop on Computer Science and Engineering, WCSE 2018 (pp. 282–287). International Workshop on Computer Science and Engineering (WCSE). https://doi.org/10.18178/wcse.2018.06.050

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free