Hand tracking is a challenging problem due to the complexity of searching in a 20+ degrees of freedom space for an optimal estimate. This paper develops a statistical method for robust visual hand tracking, in which graphical model decoupling different hand joints is performed to represent the hand constraints. Each node of the graphical model represents the position and the orientation of each hand joint in world coordinate. Then, the problem of hand tracking is transformed into an inference of graphical model. We extend Nonparametric Belief Propagation to a sequential process to track hand motion. The Experiment results show that this approach is robust for 3D hand motion tracking. © Springer-Verlag Berlin Heidelberg 2005.
CITATION STYLE
Liang, W., Jia, Y., & Ge, C. (2005). Visual hand tracking using nonparametric sequential belief propagation. In Lecture Notes in Computer Science (Vol. 3644, pp. 679–687). Springer Verlag. https://doi.org/10.1007/11538059_71
Mendeley helps you to discover research relevant for your work.