Appearance features are good at discriminating activities in a fixed view, but behave poorly when aspect is changed. We describe a method to build features that are highly stable under change of aspect. It is not necessary to have multiple views to extract our features. Our features make it possible to learn a discriminative model of activity in one view, and spot that activity in another view, for which one might poses no labeled examples at all. Our construction uses labeled examples to build activity models, and unlabeled, but corresponding, examples to build an implicit model of how appearance changes with aspect. We demonstrate our method with challenging sequences of real human motion, where discriminative methods built on appearance alone fail badly. © 2008 Springer Berlin Heidelberg.
CITATION STYLE
Farhadi, A., & Tabrizi, M. K. (2008). Learning to recognize activities from the wrong view point. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 5302 LNCS, pp. 154–166). Springer Verlag. https://doi.org/10.1007/978-3-540-88682-2_13
Mendeley helps you to discover research relevant for your work.