Multimodal Prediction of User's Performance in High-Stress Dialogue Interactions

2Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.
Get full text

Abstract

High-Stress interactions include cases in which decisions must be made, communicated, and agreed upon in a short amount of time to avoid dire consequences. Such interactions can be a source of different multimodal signals indicating participant cognitive and emotional states, which can vary with factors such as the difficulty of the interaction. By utilizing behavioral cues, a multimodal deep neural network (with audio, video, and text modalities) was developed to predict the performance of users in these interactions. An ablation study was conducted to compare impact of different modalities. Our best model can predict the user performance with 73% accuracy in a 3-class classification task.

Cite

CITATION STYLE

APA

Nasihati Gilani, S., Pollard, K., & Traum, D. (2023). Multimodal Prediction of User’s Performance in High-Stress Dialogue Interactions. In ACM International Conference Proceeding Series (pp. 71–75). Association for Computing Machinery. https://doi.org/10.1145/3610661.3617166

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free