Human action recognition based on temporal pose CNN and multi-dimensional fusion

8Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

To take advantage of recent advances in human pose estimation from images, we develop a deep neural network model for action recognition from videos by computing temporal human pose features with a 3D CNN model. The proposed temporal pose features can provide more discriminative human action information than previous video features, such as appearance and short-term motion. In addition, we propose a novel fusion network that combines temporal pose, spatial and motion feature maps for the classification by bridging the gap between the dimension difference between 3D and 2D CNN feature maps. We show that the proposed action recognition system provides superior accuracy compared to the previous methods through experiments on Sub-JHMDB and PennAction datasets.

Cite

CITATION STYLE

APA

Huang, Y., Lai, S. H., & Tai, S. H. (2019). Human action recognition based on temporal pose CNN and multi-dimensional fusion. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11130 LNCS, pp. 426–440). Springer Verlag. https://doi.org/10.1007/978-3-030-11012-3_33

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free