LSTM for image annotation with relative visual importance

1Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We consider the problem of image annotations that takes into account of the relative visual importance of tags. Previous works usually consider the tags associated with an image as an unordered set of object names. In contrast, we exploit the implicit cues about the relative importance of objects mentioned by the tags. For example, important objects tend to be mentioned first in a list of tags. We propose a recurrent neural network with long-short term memory to model this. Given an image, our model can produce a ranked list of tags, where tags for objects of higher visual importance appear earlier in the list. Experimental results demonstrate that our model achieves better performance on several benchmark datasets.

Cite

CITATION STYLE

APA

Yan, G., Wang, Y., & Liao, Z. (2016). LSTM for image annotation with relative visual importance. In British Machine Vision Conference 2016, BMVC 2016 (Vol. 2016-September, pp. 78.1-78.11). British Machine Vision Conference, BMVC. https://doi.org/10.5244/C.30.78

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free