Reborn filters: Pruning convolutional neural networks with limited data

22Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

Abstract

Channel pruning is effective in compressing the pretrained CNNs for their deployment on low-end edge devices. Most existing methods independently prune some of the original channels and need the complete original dataset to fix the performance drop after pruning. However, due to commercial protection or data privacy, users may only have access to a tiny portion of training examples, which could be insufficient for the performance recovery. In this paper, for pruning with limited data, we propose to use all original filters to directly develop new compact filters, named reborn filters, so that all useful structure priors in the original filters can be well preserved into the pruned networks, alleviating the performance drop accordingly. During training, reborn filters can be easily implemented via 1 × 1 convolutional layers and then be fused in the inference stage for acceleration. Based on reborn filters, the proposed channel pruning algorithm shows its effectiveness and superiority on extensive experiments.

Cite

CITATION STYLE

APA

Tang, Y., You, S., Xu, C., Han, J., Qian, C., Shi, B., … Zhang, C. (2020). Reborn filters: Pruning convolutional neural networks with limited data. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 5972–5980). AAAI press. https://doi.org/10.1609/aaai.v34i04.6058

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free