JOURNAL ARTICLE

User Emotion Recognition Method Based on Facial Expression and Speech Signal Fusion

Abstract

In human-computer interaction, it is an urgent problem to use facial expressions and speech information to identify the user's continuous emotions, and the key factors affecting the recognition accuracy are the data deficiencies during the fusion of speech and facial information, and the abnormal frames in the video. In order to solve these problems, a user emotion recognition system based on the fusion of facial expressions and speech multimodality is designed. In the part of facial expressions, Gabor transform continuous emotion recognition method based on data increments is proposed. In the part of speech information, Mel-scale Frequency Cepstral Coefficients (MFCC) is used to extract speech features, and user emotions are recognize through transfer learning. Finally, in the late fusion, multiple linear regression is used for multi-modality to verify the method in this paper. This paper uses the AVEC2013 dataset with Arousal-Valence label to conduct a valid experiment on the proposed method. The experimental results prove that the method improves the accuracy of user emotion recognition.

Keywords:
Speech recognition Computer science Facial expression Mel-frequency cepstrum Artificial intelligence Emotion recognition Modality (human–computer interaction) Pattern recognition (psychology) Emotion classification Feature extraction

Metrics

3
Cited By
0.20
FWCI (Field Weighted Citation Impact)
21
Refs
0.55
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Emotion and Mood Recognition
Social Sciences →  Psychology →  Experimental and Cognitive Psychology
Face and Expression Recognition
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Face recognition and analysis
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

JOURNAL ARTICLE

Feature Fusion Algorithm for Multimodal Emotion Recognition from Speech and Facial Expression Signal

Zhiyan HanJian Wang

Journal:   MATEC Web of Conferences Year: 2016 Vol: 61 Pages: 03012-03012
JOURNAL ARTICLE

Multimodal emotion recognition from facial expression and speech based on feature fusion

Guichen TangYue XieKe LiRuiyu LiangZhao Li

Journal:   Multimedia Tools and Applications Year: 2022 Vol: 82 (11)Pages: 16359-16373
JOURNAL ARTICLE

Bimodal emotion recognition based on facial expression and speech

Jie YanGuojun LuHaibo LiShuai Wang

Journal:   Journal of Nanjing University of Posts and Telecommunications Year: 2018 Vol: 38 (1)Pages: 60-65
© 2026 ScienceGate Book Chapters — All rights reserved.