Multimedia event detection (MED) is a challenging problem because of the heterogeneous content and variable quality found in large collections of Internet videos. To study the value of multimedia features and fusion for representing and learning events from a set of example video clips, we created SESAME, a system for video SEarch with Speed and Accuracy for Multimedia Events. SESAME includes multiple bag-of-words event classifiers based on single data types: low-level visual, motion, and audio features; high-level semantic visual concepts; and automatic speech recognition (ASR). Event detection performance was evaluated for each event classifier. The performance of low-level visual and motion features was improved by the use of difference coding. The accuracy of the visual concepts was nearly as strong as that of the low-level visual features. Experiments with a number of fusion methods for combining the event detection scores from these classifiers revealed that simple fusion methods, such as arithmetic mean, perform as well as or better than other, more complex fusion methods.
CITATION STYLE
Myers, G. K., Snoek, C. G. M., Nevatia, R., Nallapati, R., van Hout, J., Pancoast, S., … Smeulders, A. W. M. (2014). Evaluating multimedia features and fusion for example-based event detection. In Advances in Computer Vision and Pattern Recognition (Vol. 64, pp. 109–133). Springer London. https://doi.org/10.1007/978-3-319-05696-8_5
Mendeley helps you to discover research relevant for your work.