We propose a light-weight yet highly robust method for real- time human performance capture based on a single depth camera and sparse inertial measurement units (IMUs). Our method combines non- rigid surface tracking and volumetric fusion to simultaneously recon- struct challenging motions, detailed geometries and the inner human body of a clothed subject. The proposed hybrid motion tracking algo- rithm and efficient per-frame sensor calibration technique enable non- rigid surface reconstruction for fast motions and challenging poses with severe occlusions. Significant fusion artifacts are reduced using a new confidence measurement for our adaptive TSDF-based fusion. The above contributions are mutually beneficial in our reconstruction system, which enable practical human performance capture that is real-time, robust, low-cost and easy to deploy. Experiments show that extremely challeng- ing performances and loop closure problems can be handled successfully. Keywords:
CITATION STYLE
Ren, L., Yuan, X., B, J. L., Yang, M., & Zhou, J. (2018). Computer Vision – ECCV 2018 (Vol. 11213, pp. 697–713). Retrieved from http://dx.doi.org/10.1007/978-3-030-01240-3_42%0Ahttp://link.springer.com/10.1007/978-3-030-01240-3
Mendeley helps you to discover research relevant for your work.