Representation learning, scene understanding, and feature fusion for drowsiness detection

29Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We propose a novel drowsiness detection method based on 3D-Deep Convolutional Neural Network (3D-DCNN). We design a learning architecture for the drowsiness detection, which consists of three building blocks for representation learning, scene understanding, and feature fusion. In this framework, the model generates a spatio-temporal representation from multiple consecutive frames and analyze the scene conditions which are defined as head, eye, and mouth movements. The result of analysis from the scene condition understanding model is used to auxiliary information for the drowsiness detection. Then the method subsequently generates fusion features using the spatio-temporal representation and the results of the classification of scene conditions. By using the fusion features, we show that the proposed method can boost the performance of drowsiness detection. The proposed method demonstrates with the NTHU Drowsy Driver Detection (NTHU-DDD) video dataset.

Cite

CITATION STYLE

APA

Yu, J., Park, S., Lee, S., & Jeon, M. (2017). Representation learning, scene understanding, and feature fusion for drowsiness detection. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 10118 LNCS, pp. 165–177). Springer Verlag. https://doi.org/10.1007/978-3-319-54526-4_13

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free