Gesture-based system for next generation natural and intuitive interfaces

17Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.

Abstract

We present a novel and trainable gesture-based system for next-generation intelligent interfaces. The system requires a non-contact depth sensing device such as an RGB-D (color and depth) camera for user input. The camera records the user's static hand pose and palm center dynamic motion trajectory. Both static pose and dynamic trajectory are used independently to provide commands to the interface. The sketches/symbols formed by palm center trajectory is recognized by the Support Vector Machine classifier. Sketch/symbol recognition process is based on a set of geometrical and statistical features. Static hand pose recognizer is incorporated to expand the functionalities of our system. Static hand pose recognizer is used in conjunction with sketch classification algorithm to develop a robust and effective system for natural and intuitive interaction. To evaluate the performance of the system user studies were performed on multiple participants. The efficacy of the presented system is demonstrated using multiple interfaces developed for different tasks including computer-aided design modeling.

Cite

CITATION STYLE

APA

Huang, J., Jaiswal, P., & Rai, R. (2019). Gesture-based system for next generation natural and intuitive interfaces. Artificial Intelligence for Engineering Design, Analysis and Manufacturing: AIEDAM, 33(1), 54–68. https://doi.org/10.1017/S0890060418000045

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free