Temporal localization of actions with actoms

171Citations
Citations of this article
104Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We address the problem of localizing actions, such as opening a door, in hours of challenging video data. We propose a model based on a sequence of atomic action units, termed 'actoms,' that are semantically meaningful and characteristic for the action. Our actom sequence model (ASM) represents an action as a sequence of histograms of actom-anchored visual features, which can be seen as a temporally structured extension of the bag-of-features. Training requires the annotation of actoms for action examples. At test time, actoms are localized automatically based on a nonparametric model of the distribution of actoms, which also acts as a prior on an action's temporal structure. We present experimental results on two recent benchmarks for action localization 'Coffee and Cigarettes' and the 'DLSBP' dataset. We also adapt our approach to a classification-by-localization set-up and demonstrate its applicability on the challenging 'Hollywood 2' dataset. We show that our ASM method outperforms the current state of the art in temporal action localization, as well as baselines that localize actions with a sliding window method. © 1979-2012 IEEE.

Cite

CITATION STYLE

APA

Gaidon, A., Harchaoui, Z., & Schmid, C. (2013). Temporal localization of actions with actoms. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(11), 2782–2795. https://doi.org/10.1109/TPAMI.2013.65

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free