Convolutional neural networks (CNNs) were created for image classification tasks. Shortly after their creation, they were applied to other domains, including natural language processing (NLP). Nowadays, solutions based on artificial intelligence appear on mobile devices and embedded systems, which places constraints on memory and power consumption, among others. Due to CNN mem-ory and computing requirements, it is necessary to compress them in order to be mapped to the hardware. This paper presents the results of the compression of ecient CNNs for sentiment analysis. The main steps involve pruning and quantization. The process of mapping the compressed network to an FPGA and the results of this implementation are described. The conducted simulations showed that the 5-bit width is enough to ensure no drop in accuracy when compared to the oating-point version of the network. Additionally, the memory footprint was significantly reduced (between 85 and 93% as compared to the original model).
CITATION STYLE
Wróbel, K., Karwatowski, M., Wielgosz, M., Pietroń, M., & Wiatr, K. (2020). Compressing sentiment analysis cnn models for efficient hardware processing. Computer Science, 21(1), 25–41. https://doi.org/10.7494/csci.2020.21.1.3375
Mendeley helps you to discover research relevant for your work.