Abstract

The interaction between human beings and computers will be more natural if computers are able to perceive and respond to human non-verbal communication such as emotions. Although several approaches have been proposed to recognize human emotions based on facial expressions or speech, relatively limited work has been done to fuse these two, and other, modalities to improve the accuracy and robustness of the emotion recognition system. This paper analyzes the strengths and the limitations of systems based only on facial expressions or acoustic information. It also discusses two approaches used to fuse these two modalities: decision level and feature level integration. Using a database recorded from an actress, four emotions were classified: sadness, anger, happiness, and neutral state. By the use of markers on her face, detailed facial motions were captured with motion capture, in conjunction with simultaneous speech recordings. The results reveal that the system based on facial expression gave better performance than the system based on just acoustic information for the emotions considered. Results also show the complementarily of the two modalities and that when these two modalities are fused, the performance and the robustness of the emotion recognition system improve measurably.

Keywords:
Sadness Modalities Facial expression Speech recognition Computer science Robustness (evolution) Happiness Anger Facial recognition system Emotion recognition Artificial intelligence Affective computing Feature extraction Psychology

Metrics

824
Cited By
8.45
FWCI (Field Weighted Citation Impact)
26
Refs
0.98
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Face and Expression Recognition
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Emotion and Mood Recognition
Social Sciences →  Psychology →  Experimental and Cognitive Psychology
Speech and Audio Processing
Physical Sciences →  Computer Science →  Signal Processing

Related Documents

JOURNAL ARTICLE

Multimodal Emotion Recognition Based on Facial Expressions, Speech, and EEG

Jiahui PanWeijie FangZhihang ZhangBingzhi ChenZheng ZhangShuihua Wang

Journal:   IEEE Open Journal of Engineering in Medicine and Biology Year: 2023 Vol: 5 Pages: 396-403
JOURNAL ARTICLE

Emotion Detection using facial Expressions and speech recognition

B. Apurva G. Harshitha

Journal:   Zenodo (CERN European Organization for Nuclear Research) Year: 2022
JOURNAL ARTICLE

Emotion Detection using facial Expressions and speech recognition

G. Harshitha, B. Apurva

Journal:   Zenodo (CERN European Organization for Nuclear Research) Year: 2022
JOURNAL ARTICLE

SPEECH EMOTION RECOGNITION WITH FACIAL EXPRESSIONS

Dr.R PremaA.Vamsi KumarB.Sai Eswar Reddy

Journal:   INTERANTIONAL JOURNAL OF SCIENTIFIC RESEARCH IN ENGINEERING AND MANAGEMENT Year: 2023 Vol: 07 (03)
© 2026 ScienceGate Book Chapters — All rights reserved.