A multi-modal approach for natural human-robot interaction

9Citations
Citations of this article
28Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present a robot that is able to interact with people in a natural, multi-modal way by using both speech and gesture. The robot is able to track people, process speech and understand language. To track people and recognize gestures, the robot uses an RGB-D sensor (e.g., a Microsoft Kinect). To recognize speech, the robot uses a cloud-based service. To understand language, the robot uses a probabilistic graphical model to infer the meaning of a natural language query. We have evaluated our system in two domains. The first domain is a robot receptionist (roboceptionist); we show that the roboceptionist is able to interact successfully with people 77% of the time when people are primed with the capabilities of the robot compared to 57% when people are not primed with its capabilities. The second domain is a mobile service robot, which is able to interact with people via natural language. © 2012 Springer-Verlag.

Cite

CITATION STYLE

APA

Kollar, T., Vedantham, A., Sobel, C., Chang, C., Perera, V., & Veloso, M. (2012). A multi-modal approach for natural human-robot interaction. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7621 LNAI, pp. 458–467). https://doi.org/10.1007/978-3-642-34103-8_46

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free