A Multimodal Dataset for Robot Learning to Imitate Social Human-Human Interaction

10Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Humans tend to use various nonverbal signals to communicate their messages to their interaction partners. Previous studies utilised this channel as an essential clue to develop automatic approaches for understanding, modelling and synthesizing individual behaviours in human-human interaction and human-robot interaction settings. On the other hand, in small-group interactions, an essential aspect of communication is the dynamic exchange of social signals among interlocutors. This paper introduces LISI-HHI - Learning to Imitate Social Human-Human Interaction, a dataset of dyadic human interactions recorded in a wide range of communication scenarios. The dataset contains multiple modalities simultaneously captured by high-accuracy sensors, including motion capture, RGB-D cameras, eye trackers, and microphones. LISI-HHI is designed to be a benchmark for HRI and multimodal learning research for modelling intraand interpersonal nonverbal signals in social interaction contexts and investigating how to transfer such models to social robots.

Cite

CITATION STYLE

APA

Tuyen, N. T. V., Georgescu, A. L., Di Giulio, I., & Celiktutan, O. (2023). A Multimodal Dataset for Robot Learning to Imitate Social Human-Human Interaction. In ACM/IEEE International Conference on Human-Robot Interaction (pp. 238–242). IEEE Computer Society. https://doi.org/10.1145/3568294.3580080

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free