Appearances can be deceiving: Learning visual tracking from few trajectory annotations

5Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Visual tracking is the task of estimating the trajectory of an object in a video given its initial location. This is usually done by combining at each step an appearance and a motion model. In this work, we learn from a small set of training trajectory annotations how the objects in the scene typically move. We learn the relative weight between the appearance and the motion model. We call this weight: visual deceptiveness. At test time, we transfer the deceptiveness and the displacement from the closest trajectory annotation to infer the next location of the object. Further, we condition the transference on an event model. On a set of 161 manually annotated test trajectories, we show in our experiments that learning from just 10 trajectory annotations halves the center location error and improves the success rate by about 10%. © 2014 Springer International Publishing.

Cite

CITATION STYLE

APA

Manen, S., Kwon, J., Guillaumin, M., & Van Gool, L. (2014). Appearances can be deceiving: Learning visual tracking from few trajectory annotations. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8693 LNCS, pp. 157–172). Springer Verlag. https://doi.org/10.1007/978-3-319-10602-1_11

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free