On trivial solution and high correlation problems in deep supervised hashing

12Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.

Abstract

Deep supervised hashing (DSH), which combines binary learning and convolutional neural network, has attracted considerable research interests and achieved promising performance for highly efficient image retrieval. In this paper, we show that the widely used loss functions, pair-wise loss and triplet loss, suffer from the trivial solution problem and usually lead to highly correlated bits in practice, limiting the performance of DSH. One important reason is that it is difficult to incorporate proper constraints into the loss functions under the mini-batch based optimization algorithm. To tackle these problems, we propose to adopt ensemble learning strategy for deep model training. We found out that this simple strategy is capable of effectively decorrelating different bits, making the hashcodes more informative. Moreover, it is very easy to parallelize the training and support incremental model learning, which are very useful for real-world applications but usually ignored by existing DSH approaches. Experiments on benchmarks demonstrate the proposed ensemble based DSH can improve the performance of DSH approaches significant.

Cite

CITATION STYLE

APA

Guo, Y., Zhao, X., Ding, G., & Han, J. (2018). On trivial solution and high correlation problems in deep supervised hashing. In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018 (pp. 2240–2247). AAAI press. https://doi.org/10.1609/aaai.v32i1.11855

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free