Bag of spatio-temporal synonym sets for human action recognition

1Citations
Citations of this article
4Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recently, bag of spatio-temporal local features based methods have received significant attention in human action recognition. However, it remains a big challenge to overcome intra-class variations in cases of viewpoint, geometric and illumination variance. In this paper we present Bag of Spatio-temporal Synonym Sets (ST-SynSets) to represent human actions, which can partially bridge the semantic gap between visual appearances and category semantics. Firstly, it re-clusters the original visual words into a higher level ST-SynSet based on the distribution consistency among different action categories using Information Bottleneck clustering method. Secondly, it adaptively learns a distance metric with both the visual and semantic constraints for ST-SynSets projection. Experiments and comparison with state-of-art methods show the effectiveness and robustness of the proposed method for human action recognition, especially in multiple viewpoints and illumination conditions. © 2010 Springer-Verlag Berlin Heidelberg.

Cite

CITATION STYLE

APA

Pang, L., Cao, J., Guo, J., Lin, S., & Song, Y. (2009). Bag of spatio-temporal synonym sets for human action recognition. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 5916 LNCS, pp. 422–432). https://doi.org/10.1007/978-3-642-11301-7_43

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free