Explicit relevance feedback requires the user to explicitly refine the search queries for content-based image retrieval. This may become laborious or even impossible due to the ever-increasing volume of digital databases. We present a multimodal information collector that can unobtrusively record and asynchronously transmit the user's implicit relevance feedback on a displayed image to the remote CBIR server for assisting in retrieving relevant images. The modalities of user interaction include eye movements, pointer tracks and clicks, keyboard strokes, and audio including speech. The client-side information collector has been implemented as a browser extension using the JavaScript programming language and has been integrated with an existing CBIR server. We verify its functionality by evaluating the performance of the gaze-enhanced CBIR system in on-line image tagging tasks. © 2011 Springer-Verlag.
CITATION STYLE
Zhang, H., Sjöberg, M., Laaksonen, J., & Oja, E. (2011). A multimodal information collector for content-based image retrieval system. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7064 LNCS, pp. 737–746). https://doi.org/10.1007/978-3-642-24965-5_83
Mendeley helps you to discover research relevant for your work.