Unsupervised deep cross-modal hashing by knowledge distillation for large-scale cross-modal retrieval

26Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Cross-modal hashing (CMH) maps heterogeneous multiple modality data into compact binary code to achieve fast and flexible retrieval across different modalities, especially in large-scale retrieval. As the data don't need a lot of manual annotation, unsupervised cross-modal hashing has a wider application prospect than supervised method. However, the existing unsupervised methods are difficult to achieve satisfactory performance due to the lack of credible supervisory information. To solve this problem, inspired by knowledge distillation, we propose a novel unsupervised Knowledge Distillation Cross-Modal Hashing method (KDCMH), which can use similarity information distilled from unsupervised method to guide supervised method. Specifically, firstly, the teacher model adopted an unsupervised distribution-based similarity hashing method, which can construct a modal fusion similarity matrix.Secondly, under the supervision of teacher model distillation information, student model can generate more discriminative hash codes. In two public datasets NUS-WIDE and MIRFLICKR-25K, extensive experiments have proved the significant improvement of KDCMH on several representative unsupervised cross-modal hashing methods.

Cite

CITATION STYLE

APA

Li, M., & Wang, H. (2021). Unsupervised deep cross-modal hashing by knowledge distillation for large-scale cross-modal retrieval. In ICMR 2021 - Proceedings of the 2021 International Conference on Multimedia Retrieval (pp. 183–191). Association for Computing Machinery, Inc. https://doi.org/10.1145/3460426.3463626

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free