We tested whether head-movements under automated driving can be used to classify a vehicle occupant as either situation-aware or unaware. While manually cornering, an active driver´s head tilt correlates with the road angle which serves as a visual reference, whereas an inactive passenger´s head follows the g-forces. Transferred to partial/conditional automation, the question arises whether aware occupant´s head-movements are comparable to drivers and if this can be used for classification. In a driving-simulator-study (n=43, within-subject design), four scenarios were used to generate or deteriorate situation awareness (manipulation checked). Recurrent neural networks were trained with the resulting head-movements. Inference statistics were used to extract the discriminating feature, ensuring explain ability. A very accurate classification was achieved and the mean side rotation-rate was identified as the most differentiating factor. Aware occupants behave more like drivers. Therefore, head-movements can be used to classify situation awareness in experimental settings but also in real driving.
CITATION STYLE
Schewe, F., Cheng, H., Hafner, A., Sester, M., & Vollrath, M. (2019). Occupant Monitoring in Automated Vehicles: Classification of Situation Awareness Based on Head Movements While Cornering. In Proceedings of the Human Factors and Ergonomics Society (Vol. 63). SAGE Publications Inc. https://doi.org/10.1177/1071181319631048
Mendeley helps you to discover research relevant for your work.