Recently our research has focused on tools for video annotation — the generation of symbolic descriptions of dynamic scenes. In this paper I describe two specific examples of developing capabilities necessary for the understanding of action in video sequences. The first is a novel tracking technique that builds context-specific templates that are used only locally in time and space. The second example is some work on gesture recognition where time is implicitly represented in a probabilistic manner. We believe the future of computer vision lies in the processing of video and that while much work has been devoted to the representation and manipulation of static images, we are far behind in developing tools for considering action.
CITATION STYLE
Bobick, A. F. (1996). Video annotation: Computers watching video. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 1035, pp. 23–31). Springer Verlag. https://doi.org/10.1007/3-540-60793-5_59
Mendeley helps you to discover research relevant for your work.