JOURNAL ARTICLE

A Lightweight Multi-modal Emotion Recognition Network Based on Multi-task Learning

Abstract

Human emotion recognition is a very important part of the human-computer interaction process, and its application scenarios are very wide, which has received more and more attention in recent years. In this paper, a lightweight multimodal emotion recognition network is proposed, which makes the network model as small as possible under the premise of ensuring network accuracy, so that human emotion recognition can be well applied to mobile devices. Specifically, this article uses three modalities: audio, video, and text as input data. The audio signal is converted into MFCC and video signal using MobileNet for feature extraction, thereby reducing the amount of network parameters. For text data, Bert is used for feature extraction, and features extracted from the three modalities are combined through the attention mechanism. Finally, in order to improve the recognition rate and generalization ability of the network, a multi-task structure is also introduced. The experimental results show that the lightweight model can effectively reduce the amount of network parameters, greatly reduce the requirements for equipment, and make it possible to apply emotion recognition on the mobile terminal.

Keywords:
Computer science Feature extraction Mel-frequency cepstrum Emotion recognition Task (project management) Process (computing) Artificial intelligence Modalities Feature (linguistics) Speech recognition Modal Generalization Pattern recognition (psychology) Machine learning

Metrics

2
Cited By
0.41
FWCI (Field Weighted Citation Impact)
29
Refs
0.64
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Emotion and Mood Recognition
Social Sciences →  Psychology →  Experimental and Cognitive Psychology
Sentiment Analysis and Opinion Mining
Physical Sciences →  Computer Science →  Artificial Intelligence
Advanced Computing and Algorithms
Social Sciences →  Social Sciences →  Urban Studies
© 2026 ScienceGate Book Chapters — All rights reserved.