A framework for recognizing the simultaneous aspects of American sign language

241Citations
Citations of this article
122Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The major challenge that faces American Sign Language (ASL) recognition now is developing methods that will scale well with increasing vocabulary size. Unlike in spoken languages, phonemes can occur simultaneously in ASL. The number of possible combinations of phonemes is approximately 1.5 × 109, which cannot be tackled by conventional hidden Markov model-based methods. Gesture recognition, which is less constrained than ASL recognition, suffers from the same problem. In this paper we present a novel framework to ASL recognition that aspires to being a solution to the scalability problems. It is based on breaking down the signs into their phonemes and modeling them with parallel hidden Markov models. These model the simultaneous aspects of ASL independently. Thus, they can be trained independently, and do not require consideration of the different combinations at training time. We show in experiments with a 22-sign-vocabulary how to apply this framework in practice. We also show that parallel hidden Markov models outperform conventional hidden Markov models.

Cite

CITATION STYLE

APA

Vogler, C., & Metaxas, D. (2001). A framework for recognizing the simultaneous aspects of American sign language. Computer Vision and Image Understanding, 81(3), 358–384. https://doi.org/10.1006/cviu.2000.0895

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free