Towards Efficient Coarse-to-Fine Networks for Action and Gesture Recognition

10Citations
Citations of this article
62Readers
Mendeley users who have this article in their library.
Get full text

Abstract

State-of-the-art approaches to video-based action and gesture recognition often employ two key concepts: First, they employ multistream processing; second, they use an ensemble of convolutional networks. We improve and extend both aspects. First, we systematically yield enhanced receptive fields for complementary feature extraction via coarse-to-fine decomposition of input imagery along the spatial and temporal dimensions, and adaptively focus on training important feature pathways using a reparameterized fully connected layer. Second, we develop a ‘use when needed’ scheme with a ‘coarse-exit’ strategy that allows selective use of expensive high-resolution processing in a data-dependent fashion to retain accuracy while reducing computation cost. Our C2F learning approach builds ensemble networks that outperform most competing methods in terms of both reduced computation cost and improved accuracy on the Something-Something V1, V2, and Jester datasets, while also remaining competitive on the Kinetics-400 dataset. Uniquely, our C2F ensemble networks can operate at varying computation budget constraints.

Cite

CITATION STYLE

APA

Quader, N., Lu, J., Dai, P., & Li, W. (2020). Towards Efficient Coarse-to-Fine Networks for Action and Gesture Recognition. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12375 LNCS, pp. 35–51). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-58577-8_3

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free