This paper extends R-CNN, a state-of-the-art object detection method, to larger scales. To apply R-CNN to a large database storing thousands to millions of images, the SVM classification of millions to billions of DCNN features extracted from object proposals is indispensable, which imposes unrealistic computational and memory costs. Our method dramatically narrows down the number of object proposals by using an inverted index and efficiently searches by using residual vector quantization (RVQ). Instead of k-means that has been used in inverted indices, we present a novel quantization method designed for linear classification wherein the quantization error is re-defined for linear classification. It approximates the error as the empirical error with pre-defined multiple exemplar classifiers and captures the variance and common attributes of object category classifiers effectively. Experimental results show that our method achieves comparable performance to that of applying R-CNN to all images while achieving a 250 times speed-up and 180 times memory reduction. Moreover, our approach significantly outperforms the stateof- the-art large-scale category detection method, with about a 40∼58% increase in top-K precision. Scalability is also validated, and we demonstrate that our method can process 100K images in 0.13 s while retaining precision.
CITATION STYLE
Hinami, R., & Satoh, S. (2016). Large-scale R-CNN with classifier adaptive quantization. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9907 LNCS, pp. 403–419). Springer Verlag. https://doi.org/10.1007/978-3-319-46487-9_25
Mendeley helps you to discover research relevant for your work.