Harvesting training images for fine-grained object categories using visual descriptions

0Citations
Citations of this article
4Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We harvest training images for visual object recognition by casting it as an IR task. In contrast to previous work, we concentrate on fine-grained object categories, such as the large number of particular animal subspecies, for which manual annotation is expensive. We use ‘visual descriptions’ from nature guides as a novel augmentation to the well-known use of category names. We use these descriptions in both the query process to find potential category images as well as in image reranking where an image is more highly ranked if web page text surrounding it is similar to the visual descriptions. We show the potential of this method when harvesting images for 10 butterfly categories: when compared to a method that relies on the category name only, using visual descriptions improves precision for many categories.

Cite

CITATION STYLE

APA

Wang, J., Markert, K., & Everingham, M. (2016). Harvesting training images for fine-grained object categories using visual descriptions. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9626, pp. 549–560). Springer Verlag. https://doi.org/10.1007/978-3-319-30671-1_40

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free