Recognition of gestural object reference with auditory feedback

1Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present a cognitively motivated vision architecture for the evaluation of pointing gestures. The system views a scene of several structured objects and a pointing human hand. A neural classifier gives an estimation of the pointing direction, then the object correspondence is established using a sub-symbolic representation of both the scene and the pointing direction. The system achieves high robustness because the result (the indicated location) does not primarily depend on the accuracy of the pointing direction classification. Instead, the scene is analysed for low level saliency features to restrict the set of all possible pointing locations to a subset of highly likely locations. This transformation of the "continuous" to a "discrete" pointing problem simultaneously facilitates an auditory feedback whenever the object reference changes, which leads to a significantly improved human-machine interaction. © Springer-Verlag Berlin Heidelberg 2003.

Cite

CITATION STYLE

APA

Bax, I., Bekel, H., & Heidemann, G. (2003). Recognition of gestural object reference with auditory feedback. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2714, 425–432. https://doi.org/10.1007/3-540-44989-2_51

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free