This thesis investigates the use of deep learning techniques to address the problem of machine understanding of human affective behaviour and improve the accuracy of both unimodal and multimodal human emotion recognition. The objective was to explore how best to configure deep learning networks to capture individually and jointly, the key features contributing to human emotions from three modalities (speech, face, and bodily movements) to accurately classify the expressed human emotion. The outcome of the research should be useful for several applications including the design of social robots.
Sharmeen M.Saleem Abdullah AbdullahSiddeeq Y. AmeenMohammed A. M. SadeeqSubhi R. M. Zeebaree
Wei LiuWei‐Long ZhengBao‐Liang Lu
Noora Al RokenGerassimos Barlas
Hina IramArslan ShaukatMuhammad Usman Akram