Encoding document semantic into binary codes space

0Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We develop a deep neural network model to encode document semantic into compact binary codes with the elegant property that semantically similar documents have similar embedding codes. The deep learning model is constructed with three stacked auto-encoders. The input of the lowest auto-encoder is the representation of word-count vector of a document, while the learned hidden features of the deepest auto-encoder are thresholded to be binary codes to represent the document semantic. Retrieving similar document is very efficient by simply returning the documents whose codes have small Hamming distances to that of the query document. We illustrate the effectiveness of our model on two public real datasets - 20NewsGroup and Wikipedia, and the experiments demonstrate that the compact binary codes sufficiently embed the semantic of documents and bring improvement in retrieval accuracy. © 2014 Springer International Publishing Switzerland.

Cite

CITATION STYLE

APA

Yu, Z., Zhao, X., & Wang, L. (2014). Encoding document semantic into binary codes space. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8485 LNCS, pp. 535–539). Springer Verlag. https://doi.org/10.1007/978-3-319-08010-9_59

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free