Multimodal Fusion of Voice and Gesture Data for UAV Control

11Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

To enable unmanned aerial vehicle (UAV) operators to efficiently and intuitively convey their commands to a swarm of UAVs, we propose the use of natural and human-centric input modalities, such as voices and gestures. This paper addresses the fusion of input modalities such as voice and gesture data, which are captured through a microphone and a Leap Motion controller, respectively, to control UAV swarms. The obtained experimental results are presented, and the achieved performance (accuracy) is analyzed. Finally, combined human factor ergonomics test with a questionnaire to verify the method’s validity.

Cite

CITATION STYLE

APA

Xiang, X., Tan, Q., Zhou, H., Tang, D., & Lai, J. (2022). Multimodal Fusion of Voice and Gesture Data for UAV Control. Drones, 6(8). https://doi.org/10.3390/drones6080201

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free