Exploiting motion information from unlabeled videos for static image action recognition

5Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

Static image action recognition, which aims to recognize action based on a single image, usually relies on expensive human labeling effort such as adequate labeled action images and large-scale labeled image dataset. In contrast, abundant unlabeled videos can be economically obtained. Therefore, several works have explored using unlabeled videos to facilitate image action recognition, which can be categorized into the following two groups: (a) enhance visual representations of action images with a designed proxy task on unlabeled videos, which falls into the scope of self-supervised learning; (b) generate auxiliary representations for action images with the generator learned from unlabeled videos. In this paper, we integrate the above two strategies in a unified framework, which consists of Visual Representation Enhancement (VRE) module and Motion Representation Augmentation (MRA) module. Specifically, the VRE module includes a proxy task which imposes pseudo motion label constraint and temporal coherence constraint on unlabeled videos, while the MRA module could predict the motion information of a static action image by exploiting unlabeled videos. We demonstrate the superiority of our framework based on four benchmark human action datasets with limited labeled data.

Cite

CITATION STYLE

APA

Zhang, Y., Niu, L., Pan, Z., Luo, M., Zhang, J., Cheng, D., & Zhang, L. (2020). Exploiting motion information from unlabeled videos for static image action recognition. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 12918–12925). AAAI press. https://doi.org/10.1609/aaai.v34i07.6990

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free