Track facial points in unconstrained videos

2Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.

Abstract

Tracking Facial Points in unconstrained videos is challenging due to the non-rigid deformation that changes over time. In this paper, we propose to exploit incremental learning for person-specific alignment in wild conditions. Our approach takes advantage of part-based representation and cascade regression for robust and efficient alignment on each frame. Unlike existing methods that usually rely on models trained offline, we incrementally update the representation subspace and the cascade of regressors in a unified framework to achieve personalized modeling on the fly. To alleviate the drifting issue, the fitting results are evaluated using a deep neural network, where well-aligned faces are picked out to incrementally update the representation and fitting models. Both image and video datasets are employed to valid the proposed method. The results demonstrate the superior performance of our approach compared with existing approaches in terms of fitting accuracy and efficiency.

Cite

CITATION STYLE

APA

Peng, X., Hu, Q., Huang, J., & Metaxas, D. N. (2016). Track facial points in unconstrained videos. In British Machine Vision Conference 2016, BMVC 2016 (Vol. 2016-September). British Machine Vision Conference, BMVC. https://doi.org/10.5244/C.30.129

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free