Abstract
It is common to view people in real applications walking in arbitrary directions, holding items, or wearing heavy coats. These factors are challenges in gait-based application methods, because they significantly change a person‘s appearance. This paper proposes a novel method for classifying human gender in real time using gait information. The use of an average gait image, rather than a gait energy image, allows this method to be computationally efficient and robust against view changes. A viewpoint model is created for automatically determining the viewing angle during the testing phase. A distance signal model is constructed to remove any areas with an attachment (carried items, worn coats) from a silhouette to reduce the interference in the resulting classification. Finally, the human gender is classified using multiple-view-dependent classifiers trained using a support vector machine. Experiment results confirm that the proposed method achieves a high accuracy of 98.8% on the CASIA Dataset B and outperforms the recent state-of-the-art methods.
Author supplied keywords
Cite
CITATION STYLE
Do, T. D., Nguyen, V. H., & Kim, H. (2020). Real-time and robust multiple-view gender classification using gait features in video surveillance. Pattern Analysis and Applications, 23(1), 399–413. https://doi.org/10.1007/s10044-019-00802-6
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.