Multimodal conversation scene analysis for understanding people's communicative behaviors in face-to-face meetings

N/ACitations
Citations of this article
13Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

This presentation overviews our recent progress in multimodal conversation scene analysis, and discusses its future in terms of designing better human-to-human communication systems. Conversation scene analysis aims to provide the automatic description of conversation scenes from the multimodal nonverbal behaviors of participants as captured by cameras and microphones. So far, the author's group has proposed a research framework based on the probabilistic modeling of conversation phenomena for solving several basic problems including speaker diarization, i.e. "who is speaking when", addressee identification, i.e. "who is talking to whom", interaction structure, i.e. "who is responding to whom", the estimation of visual focus of attention (VFOA), i.e. "who is looking at whom", and the inference of interpersonal emotion such as "who has empathy/antipathy with whom", from observed multimodal behaviors including utterances, head pose, head gestures, eye-gaze, and facial expressions. This paper overviews our approach and discusses how conversation scene analysis can be extended to enhance the design process of computer-mediated communication systems. © 2011 Springer-Verlag.

Cite

CITATION STYLE

APA

Otsuka, K. (2011). Multimodal conversation scene analysis for understanding people’s communicative behaviors in face-to-face meetings. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6772 LNCS, pp. 171–179). https://doi.org/10.1007/978-3-642-21669-5_21

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free