ThumbNet: One Thumbnail Image Contains All You Need for Recognition

10Citations
Citations of this article
29Readers
Mendeley users who have this article in their library.

Abstract

Although deep convolutional neural networks (CNNs) have achieved great success in computer vision tasks, its real-world application is still impeded by its voracious demand of computational resources. Current works mostly seek to compress the network by reducing its parameters or parameter-incurred computation, neglecting the influence of the input image on the system complexity. Based on the fact that input images of a CNN contain substantial redundancy, in this paper, we propose a unified framework, dubbed as ThumbNet, to simultaneously accelerate and compress CNN models by enabling them to infer on one thumbnail image. We provide three effective strategies to train ThumbNet. In doing so, ThumbNet learns an inference network that performs equally well on small images as the original-input network on large images. With ThumbNet, not only do we obtain the thumbnail-input inference network that can drastically reduce computation and memory requirements, but also we obtain an image downscaler that can generate thumbnail images for generic classification tasks. Extensive experiments show the effectiveness of ThumbNet, and demonstrate that the thumbnail-input inference network learned by ThumbNet can adequately retain the accuracy of the original-input network even when the input images are downscaled 16 times.

Cite

CITATION STYLE

APA

Zhao, C., & Ghanem, B. (2020). ThumbNet: One Thumbnail Image Contains All You Need for Recognition. In MM 2020 - Proceedings of the 28th ACM International Conference on Multimedia (pp. 1506–1514). Association for Computing Machinery, Inc. https://doi.org/10.1145/3394171.3413937

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free