GrabCut-based human segmentation in video sequences

33Citations
Citations of this article
27Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we present a fully-automatic Spatio-Temporal GrabCut human segmentation methodology that combines tracking and segmentation. GrabCut initialization is performed by a HOG-based subject detection, face detection, and skin color model. Spatial information is included by Mean Shift clustering whereas temporal coherence is considered by the historical of Gaussian Mixture Models. Moreover, full face and pose recovery is obtained by combining human segmentation with Active Appearance Models and Conditional Random Fields. Results over public datasets and in a new Human Limb dataset show a robust segmentation and recovery of both face and pose using the presented methodology. © 2012 by the authors; licensee MDPI, Basel, Switzerland.

Cite

CITATION STYLE

APA

Hernández-Vela, A., Reyes, M., Ponce, V., & Escalera, S. (2012). GrabCut-based human segmentation in video sequences. Sensors (Switzerland), 12(11), 15376–15393. https://doi.org/10.3390/s121115376

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free