Recognizing the brazilian signs language alphabet with neural networks over visual 3D data sensor

3Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper describes an investigation of the Brazilian Signs Language (LInguagem BRAsileira de Sinais - LIBRAS) alphabet recognition using neural networks. The LIBRAS alphabet is represented by static postures and dynamic gestures. In this investigation, gestures were recorded with a 3D camera sensor and its associated library, providing coordinates from hands' fingertips. Deaf people, LIBRAS teachers and students were involved in the recording process. The pre-processing data involved frames sampling, normalization and 3D geometric transformations. Neural network models with different settings were trained, compared, and assessed to verify classification accuracy.

Cite

CITATION STYLE

APA

de Souza Pereira Moreira, G., Matuck, G. R., Saotome, O., & da Cunha, A. M. (2014). Recognizing the brazilian signs language alphabet with neural networks over visual 3D data sensor. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 8864, 637–648. https://doi.org/10.1007/978-3-319-12027-0_51

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free