A multimodal information collector for content-based image retrieval system

0Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Explicit relevance feedback requires the user to explicitly refine the search queries for content-based image retrieval. This may become laborious or even impossible due to the ever-increasing volume of digital databases. We present a multimodal information collector that can unobtrusively record and asynchronously transmit the user's implicit relevance feedback on a displayed image to the remote CBIR server for assisting in retrieving relevant images. The modalities of user interaction include eye movements, pointer tracks and clicks, keyboard strokes, and audio including speech. The client-side information collector has been implemented as a browser extension using the JavaScript programming language and has been integrated with an existing CBIR server. We verify its functionality by evaluating the performance of the gaze-enhanced CBIR system in on-line image tagging tasks. © 2011 Springer-Verlag.

Cite

CITATION STYLE

APA

Zhang, H., Sjöberg, M., Laaksonen, J., & Oja, E. (2011). A multimodal information collector for content-based image retrieval system. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7064 LNCS, pp. 737–746). https://doi.org/10.1007/978-3-642-24965-5_83

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free