Frame segmentation networks for temporal action localization

1Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Temporal action localization is an important task of computer vision. Though many methods have been proposed, it still remains an open question how to predict the temporal location of action segments precisely. Most state-of-the-art works train action classifiers on video segments pre-determined by action proposal. However, recent work found that a desirable model should move beyond segment-level and make dense predictions at a fine granularity in time to determine precise temporal boundaries. In this paper, we propose a Frame Segmentation Network (FSN) that places a temporal CNN on top of the 2D spatial CNNs. Spatial CNNs are responsible for abstracting semantics in spatial dimension while temporal CNN is responsible for introducing temporal context information and performing dense predictions. The proposed FSN can make dense predictions at frame-level for a video clip using both spatial and temporal context information. FSN is trained in an end-to-end manner, so the model can be optimized in spatial and temporal domain jointly. Experiment results on public dataset show that FSN achieves superior performance in both frame-level action localization and temporal action localization.

Cite

CITATION STYLE

APA

Yang, K., Qiao, P., Wang, Q., Li, S., Niu, X., Li, D., & Dou, Y. (2018). Frame segmentation networks for temporal action localization. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11165 LNCS, pp. 242–252). Springer Verlag. https://doi.org/10.1007/978-3-030-00767-6_23

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free