Continuous Sign Language Recognition Based on Spatial-Temporal Graph Attention Network

5Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

Continuous sign language recognition (CSLR) is challenging due to the complexity of video background, hand gesture variability, and temporal modeling difficulties. This work proposes a CSLR method based on a spatial-temporal graph attention network to focus on essential features of video series. The method considers local details of sign language movements by taking the information on joints and bones as inputs and constructing a spatial-temporal graph to reflect inter-frame relevance and physical connections between nodes. The graph-based multi-head attention mechanism is utilized with adjacent matrix calculation for better local-feature exploration, and short-term motion correlation modeling is completed via a temporal convolutional network. We adopted BLSTM to learn the long-term dependence and connectionist temporal classification to align the word-level sequences. The proposed method achieves competitive results regarding word error rates (1.59%) on the Chinese Sign Language dataset and the mean Jaccard Index (65.78%) on the ChaLearn LAP Continuous Gesture Dataset.

Cite

CITATION STYLE

APA

Guo, Q., Zhang, S., & Li, H. (2023). Continuous Sign Language Recognition Based on Spatial-Temporal Graph Attention Network. CMES - Computer Modeling in Engineering and Sciences, 134(3), 1653–1670. https://doi.org/10.32604/cmes.2022.021784

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free