Evaluating multimedia features and fusion for example-based event detection

4Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Multimedia event detection (MED) is a challenging problem because of the heterogeneous content and variable quality found in large collections of Internet videos. To study the value of multimedia features and fusion for representing and learning events from a set of example video clips, we created SESAME, a system for video SEarch with Speed and Accuracy for Multimedia Events. SESAME includes multiple bag-of-words event classifiers based on single data types: low-level visual, motion, and audio features; high-level semantic visual concepts; and automatic speech recognition (ASR). Event detection performance was evaluated for each event classifier. The performance of low-level visual and motion features was improved by the use of difference coding. The accuracy of the visual concepts was nearly as strong as that of the low-level visual features. Experiments with a number of fusion methods for combining the event detection scores from these classifiers revealed that simple fusion methods, such as arithmetic mean, perform as well as or better than other, more complex fusion methods.

Cite

CITATION STYLE

APA

Myers, G. K., Snoek, C. G. M., Nevatia, R., Nallapati, R., van Hout, J., Pancoast, S., … Smeulders, A. W. M. (2014). Evaluating multimedia features and fusion for example-based event detection. In Advances in Computer Vision and Pattern Recognition (Vol. 64, pp. 109–133). Springer London. https://doi.org/10.1007/978-3-319-05696-8_5

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free