Abstract
Emotions significantly impact human physical and mental health, and, therefore, emotion recognition has been a popular research area in neuroscience, psychology, and medicine. In this paper, we preprocess the raw signals acquired by millimeter-wave radar to obtain high-quality heartbeat and respiration signals. Then, we propose a deep learning model incorporating a convolutional neural network and gated recurrent unit neural network in combination with human face expression images. The model achieves a recognition accuracy of 84.5% in person-dependent experiments and 74.25% in person-independent experiments. The experiments show that it outperforms a single deep learning model compared to traditional machine learning algorithms.
Author supplied keywords
Cite
CITATION STYLE
Dang, X., Chen, Z., Hao, Z., Ga, M., Han, X., Zhang, X., & Yang, J. (2023). Wireless Sensing Technology Combined with Facial Expression to Realize Multimodal Emotion Recognition. Sensors, 23(1). https://doi.org/10.3390/s23010338
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.