Non-manual cues in automatic sign language recognition

Citations of this article
Mendeley users who have this article in their library.
Get full text
This PDF is freely available from an open access repository. It may not have been peer-reviewed.


Present work deals with the incorporation of non-manual cues in automatic sign language recognition. More specifically, eye gaze, head pose, and facial expressions are discussed in relation to their grammatical and syntactic function and means of including them in the recognition phase are investigated. Computer vision issues related to extracting facial features, eye gaze, and head pose cues are presented and classification approaches for incorporating these non-manual cues into the overall Sign Language recognition architecture are introduced. © 2012 Springer-Verlag London.




Caridakis, G., Asteriadis, S., & Karpouzis, K. (2014). Non-manual cues in automatic sign language recognition. In Personal and Ubiquitous Computing (Vol. 18, pp. 37–46). Springer London.

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free