Visual Word Ambiguity

  • Gemert J
  • Veenman C
  • Geusebroek J
N/ACitations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

This paper studies automatic image classification by modeling soft-assignment in the popular codebook model. The codebook model describes an image as a bag of discrete visual words selected from a vocabulary, where the frequency distributions of visual words in an image allow classification. One inherent component of the codebook model is the assign- ment of discrete visual words to continuous image features. Despite the clear mismatch of this hard assignment with the nature of continuous features, the approach has been applied successfully for some years. In this paper we investigate four types of soft-assignment of visual words to image features. We demonstrate that explicitly modeling visual word assignment ambiguity improves classification performance compared to the hard-assignment of the traditional codebook model. The tradi- tional codebook model is compared against our method for five well-known datasets: 15 natural scenes, Caltech-101, Caltech- 256, and Pascal VOC 2007/2008. We demonstrate that large codebook vocabulary sizes completely deteriorate the perfor- mance of the traditional model, whereas the proposed model performs consistently. Moreover, we show that our method profits in high-dimensional feature spaces and reaps higher benefits when increasing the number of image categories.

Cite

CITATION STYLE

APA

Gemert, J. C. V., Veenman, C. J., & Geusebroek, J. M. (2008). Visual Word Ambiguity. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1–13.

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free