Real-Time human pose recognition in parts from single depth images

1.3kCitations
Citations of this article
14.4kReaders
Mendeley users who have this article in their library.
Get full text

Abstract

We propose a new method to quickly and accurately predict human pose-the 3D positions of body joints-from a single depth image, without depending on information from preceding frames. Our approach is strongly rooted in current object recognition strategies. By designing an intermediate representation in terms of body parts, the difficult pose estimation problem is transformed into a simpler per-pixel classification problem, for which efficient machine learning techniques exist. By using computer graphics to synthesize a very large dataset of training image pairs, one can train a classifier that estimates body part labels from test images invariant to pose, body shape, clothing, and other irrelevances. Finally, we generate confidence-scored 3D proposals of several body joints by reprojecting the classification result and finding local modes. The system runs in under 5ms on the Xbox 360. Our evaluation shows high accuracy on both synthetic and real test sets, and investigates the effect of several training parameters. We achieve state-of-the-art accuracy in our comparison with related work and demonstrate improved generalization over exact whole-skeleton nearest neighbor matching. © 2013 ACM 0001-0782/13/01.

Cite

CITATION STYLE

APA

Shotton, J., Sharp, T., Fitzgibbon, A., Blake, A., Cook, M., Kipman, A., … Moore, R. (2013). Real-Time human pose recognition in parts from single depth images. Communications of the ACM, 56(1), 116–124. https://doi.org/10.1145/2398356.2398381

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free