Abstract
This paper introduces the video mesh, a data structure for representing video as 2.5D "paper cutouts." The video mesh allows interactive editing of moving objects and modeling of depth, which enables 3D effects and post-exposure camera control. The video mesh sparsely encodes optical flow as well as depth, and handles occlusion using local layering and alpha mattes. Motion is described by a sparse set of points tracked over time. Each point also stores a depth value. The video mesh is a triangulation over this point set and per-pixel information is obtained by interpolation. The user rotoscopes occluding contours and we introduce an algorithm to cut the video mesh along them. Object boundaries are refined with per-pixel alpha values. The video mesh is at its core a set of texture mapped triangles, we leverage graphics hardware to enable interactive editing and rendering of a variety of effects. We demonstrate the effectiveness of our representation with special effects such as 3D viewpoint changes, object insertion, depth-of-field manipulation, and 2D to 3D video conversion. © 2011 IEEE.
Cite
CITATION STYLE
Chen, J., Paris, S., Wang, J., Matusik, W., Cohen, M., & Durand, F. (2011). The video mesh: A data structure for image-based three-dimensional video editing. In 2011 IEEE International Conference on Computational Photography, ICCP 2011. https://doi.org/10.1109/ICCPHOT.2011.5753118
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.