Abstract
For the task of assigning labels to an image to summarize its contents, many early attempts use segment-level information and try to determine which parts of the images correspond to which labels. Best performing methods use global image similarity and nearest neighbor techniques to transfer labels from training images to test images. However, global methods cannot localize the labels in the images, unlike segment-level methods. Also, they cannot take advantage of training images that are only locally similar to a test image. We propose several ways to combine recent image-level and segment-level techniques to predict both image and segment labels jointly. We cast our experimental study in an unified framework for both image-level and segment-level annotation tasks. On three challenging datasets, our joint prediction of image and segment labels outperforms either prediction alone on both tasks. This confirms that the two levels offer complementary information. © 2012 Springer-Verlag.
Author supplied keywords
Cite
CITATION STYLE
Kuettel, D., Guillaumin, M., & Ferrari, V. (2012). Combining image-level and segment-level models for automatic annotation. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7131 LNCS, pp. 16–28). https://doi.org/10.1007/978-3-642-27355-1_5
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.