Semantic video annotation by mining association patterns from visual and speech features

2Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper, we propose a novel approach for semantic video annotation through integrating visual features and speech features. By employing statistics and association patterns, the relations between video shots and human concept can be discovered effectively to conceptualize videos. In other words, the utilization of high-level rules can effectively complement the insufficiency of statistics-based methods in dealing with broad and complex keyword identification in video annotation. Empirical evaluations on NIST TRECVID video datasets reveal that our proposed approach can enhance the annotation accuracy substantially. © 2008 Springer-Verlag Berlin Heidelberg.

Cite

CITATION STYLE

APA

Tseng, V. S., Su, J. H., Huang, J. H., & Chen, C. J. (2008). Semantic video annotation by mining association patterns from visual and speech features. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 5012 LNAI, pp. 1035–1041). https://doi.org/10.1007/978-3-540-68125-0_110

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free