Abstract
Purpose: Humans are required to respond to a vehicle’s request to take-over anytime even when they are not responsible for monitoring driving environments in automated driving, e.g., a SAE level-3 vehicle. Thus, a safe and effective delivery of a take-over request from an automated vehicle to a human is critical for the successful commercialization of automated vehicles. Methods: In the current study, a set of human-in-the-loop experiments was conducted to compare diverse warning combinations by applying visual, auditory, and haptic modalities under systematically classified take-over request scenarios in conditionally automated driving. Forty-one volunteers consisting of 16 females and 25 males participated in the study. Vehicle and human data on response to take-over request were collected in two take-over scenarios, i.e., a disabled vehicle on the road ahead and a highway exit. Results: Visual-auditory-haptic modal combination showed the best performance in both human behavioral and physiological data and visual-auditory warning in vehicle data. Visual-auditory-haptic warning combination showed the best performance when considering all performance indices. Meanwhile, visual-only warning, which is considered as a basic modality in manual driving, performed the worst in the conditionally automated driving situation. Conclusions: These findings imply that the warning design in automated vehicles must be clearly differentiated from that of conventional manual driving vehicles. Future work shall include a follow-up experiment to verify the study results and compare more diverse multimodal combinations.
Author supplied keywords
Cite
CITATION STYLE
Yun, H., & Yang, J. H. (2020). Multimodal warning design for take-over request in conditionally automated driving. European Transport Research Review, 12(1). https://doi.org/10.1186/s12544-020-00427-5
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.