In this paper, we study how to automatically exploit visual concepts in a text-based image retrieval task. First, we use Forest of Fuzzy Decision Trees (FFDTs) to automatically annotate images with visual concepts. Second, using optionally WordNet, we match visual concepts and textual query. Finally, we filter the text-based image retrieval result list using the FFDTs. This study is performed in the context of two tasks of the CLEF2008 international campaign: the Visual Concept Detection Task (VCDT) (17 visual concepts) and the photographic retrieval task (ImageCLEFphoto) (39 queries and 20k images). Our best VCDT run is the 4th best of the 53 submitted runs. The ImageCLEFphoto results show that there is a clear improvement, in terms of precision at 20, when using the visual concepts explicitly appearing in the query. © Springer-Verlag Berlin Heidelberg 2009.
CITATION STYLE
Tollari, S., Detyniecki, M., Marsala, C., Fakeri-Tabrizi, A., Amini, M. R., & Gallinari, P. (2009). Exploiting visual concepts to improve text-based image retrieval. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 5478 LNCS, pp. 701–705). https://doi.org/10.1007/978-3-642-00958-7_70
Mendeley helps you to discover research relevant for your work.