Abstract
High-Stress interactions include cases in which decisions must be made, communicated, and agreed upon in a short amount of time to avoid dire consequences. Such interactions can be a source of different multimodal signals indicating participant cognitive and emotional states, which can vary with factors such as the difficulty of the interaction. By utilizing behavioral cues, a multimodal deep neural network (with audio, video, and text modalities) was developed to predict the performance of users in these interactions. An ablation study was conducted to compare impact of different modalities. Our best model can predict the user performance with 73% accuracy in a 3-class classification task.
Author supplied keywords
Cite
CITATION STYLE
Nasihati Gilani, S., Pollard, K., & Traum, D. (2023). Multimodal Prediction of User’s Performance in High-Stress Dialogue Interactions. In ACM International Conference Proceeding Series (pp. 71–75). Association for Computing Machinery. https://doi.org/10.1145/3610661.3617166
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.