JOURNAL ARTICLE

Interactive Multimodal Attention Network for Emotion Recognition in Conversation

Minjie RenXiangdong HuangXiaoqi ShiWeizhi Nie

Year: 2021 Journal:   IEEE Signal Processing Letters Vol: 28 Pages: 1046-1050   Publisher: Institute of Electrical and Electronics Engineers

Abstract

In this letter, we propose a novel Interactive Multimodal Attention Network (IMAN) for emotion recognition in conversations. IMAN introduces a cross-modal attention fusion module to capture cross-modal interactions of multimodal information, and employs a conversational modeling module to explore the context information and speaker dependency of the whole conversation. Concretely, the cross-modal attention fusion module captures the cross-modal interactions and complementary information among the pre-extracted unimodal features from textual, visual, acoustic modalities based on the cross-modal attention block. Afterward, the updated features from each modality are fused to concentrate more on the informative modality and achieve a refined feature for each constituent utterance. The conversational modeling module defines three different gated recurrent units (GRUs) with respect to the context information, the speaker dependency, and the emotional state of utterances. In this way, we exploit the speaker dependency and contextual information to obtain the emotional state of utterances for emotion classification. Empirical evaluations on the multimodal benchmark IEMOCAP dataset demonstrate that our IMAN achieves competitive performance compared to the state-of-the-art approaches.

Keywords:
Computer science Dependency (UML) Conversation Utterance Modal Context (archaeology) Feature (linguistics) Modality (human–computer interaction) Modalities Benchmark (surveying) Speech recognition Artificial intelligence Natural language processing Exploit Linguistics

Metrics

29
Cited By
4.88
FWCI (Field Weighted Citation Impact)
39
Refs
0.94
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Emotion and Mood Recognition
Social Sciences →  Psychology →  Experimental and Cognitive Psychology
Sentiment Analysis and Opinion Mining
Physical Sciences →  Computer Science →  Artificial Intelligence
Human Pose and Action Recognition
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

JOURNAL ARTICLE

SIA-Net: Sparse Interactive Attention Network for Multimodal Emotion Recognition

Shuzhen LiTong ZhangC. L. Philip Chen

Journal:   IEEE Transactions on Computational Social Systems Year: 2024 Vol: 11 (5)Pages: 6782-6794
JOURNAL ARTICLE

Speaker-Aware Interactive Graph Attention Network for Emotion Recognition in Conversation

Zhaohong JiaYunwei ShiWeifeng LiuZhenhua HuangXiao Sun

Journal:   ACM Transactions on Asian and Low-Resource Language Information Processing Year: 2023 Vol: 22 (12)Pages: 1-18
JOURNAL ARTICLE

HAAN-ERC: hierarchical adaptive attention network for multimodal emotion recognition in conversation

Tao ZhangZhenhua TanXiaoer Wu

Journal:   Neural Computing and Applications Year: 2023 Vol: 35 (24)Pages: 17619-17632
© 2026 ScienceGate Book Chapters — All rights reserved.