JOURNAL ARTICLE

Texture and Shape Information Fusion for Facial Action Unit Recognition

Abstract

A novel method that fuses texture and shape information to achieve Facial Action Unit (FAU) recognition from video sequences is proposed. In order to extract the texture information, a subspace method based on Discriminant Non- negative Matrix Factorization (DNMF) is applied on the difference images of the video sequence, calculated taking under consideration the neutral and the most expressive frame, to extract the desired classification label. The shape information consists of the deformed Candide facial grid (more specifically the grid node displacements between the neutral and the most expressive facial expression frame) that corresponds to the facial expression depicted in the video sequence. The shape information is afterwards classified using a two-class Support Vector Machine (SVM) system. The fusion of texture and shape information is performed using Median Radial Basis Functions (MRBFs) Neural Networks (NNs) in order to detect the set of present FAUs. The accuracy achieved in the Cohn-Kanade database is equal to 92.1% when recognizing the 17 FAUs that are responsible for facial expression development.

Keywords:
Artificial intelligence Computer science Pattern recognition (psychology) Computer vision Support vector machine Facial expression Texture (cosmology) Discriminant Feature extraction Subspace topology Facial recognition system Image (mathematics)

Metrics

9
Cited By
0.59
FWCI (Field Weighted Citation Impact)
12
Refs
0.74
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Face and Expression Recognition
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Emotion and Mood Recognition
Social Sciences →  Psychology →  Experimental and Cognitive Psychology
Advanced Computing and Algorithms
Social Sciences →  Social Sciences →  Urban Studies

Related Documents

© 2026 ScienceGate Book Chapters — All rights reserved.