Similarity-Preserving Hash for Content-based Audio Retrieval using Unsupervised deep Neural Networks

6Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.

Abstract

Due to its efficiency in storage and search speed, binary hashing has become an attractive approach for a large audio database search. However, most existing hashing-based methods focus on data-independent scheme where random linear projections or some arithmetic expression are used to construct hash functions. Hence, the binary codes do not preserve the similarity and may degrade the search performance. In this paper, an unsupervised similarity-preserving hashing method for content-based audio retrieval is proposed. Different from data-independent hashing methods, we develop a deep network to learn compact binary codes from multiple hierarchical layers of nonlinear and linear transformations such that the similarity between samples is preserved. The independence and balance properties are included and optimized in the objective function to improve the codes. Experimental results on the Extended Ballroom dataset with 8 genres of 3,000 musical excerpts show that our proposed method significantly outperforms state-of-the-art data-independent method in both effectiveness and efficiency.

Cite

CITATION STYLE

APA

Panyapanuwat, P., Kamonsantiroj, S., & Pipanmaekaporn, L. (2021). Similarity-Preserving Hash for Content-based Audio Retrieval using Unsupervised deep Neural Networks. International Journal of Electrical and Computer Engineering, 11(1), 879–891. https://doi.org/10.11591/ijece.v11i1.pp879-891

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free