JOURNAL ARTICLE

Multi-Scale Masked Autoencoders for Cross-Session Emotion Recognition

Miaoqi PangHongtao WangJiayang HuangChi‐Man VongZhiqiang ZengChuangquan Chen

Year: 2024 Journal:   IEEE Transactions on Neural Systems and Rehabilitation Engineering Vol: 32 Pages: 1637-1646   Publisher: Institute of Electrical and Electronics Engineers

Abstract

Affective brain-computer interfaces (aBCIs) have garnered widespread applications, with remarkable advancements in utilizing electroencephalogram (EEG) technology for emotion recognition. However, the time-consuming process of annotating EEG data, inherent individual differences, non-stationary characteristics of EEG data, and noise artifacts in EEG data collection pose formidable challenges in developing subject-specific cross-session emotion recognition models. To simultaneously address these challenges, we propose a unified pre-training framework based on multi-scale masked autoencoders (MSMAE), which utilizes large-scale unlabeled EEG signals from multiple subjects and sessions to extract noise-robust, subject-invariant, and temporal-invariant features. We subsequently fine-tune the obtained generalized features with only a small amount of labeled data from a specific subject for personalization and enable cross-session emotion recognition. Our framework emphasizes: 1) Multi-scale representation to capture diverse aspects of EEG signals, obtaining comprehensive information; 2) An improved masking mechanism for robust channel-level representation learning, addressing missing channel issues while preserving inter-channel relationships; and 3) Invariance learning for regional correlations in spatial-level representation, minimizing inter-subject and inter-session variances. Under these elaborate designs, the proposed MSMAE exhibits a remarkable ability to decode emotional states from a different session of EEG data during the testing phase. Extensive experiments conducted on the two publicly available datasets, i.e., SEED and SEED-IV, demonstrate that the proposed MSMAE consistently achieves stable results and outperforms competitive baseline methods in cross-session emotion recognition.

Keywords:
Computer science Electroencephalography Session (web analytics) Artificial intelligence Pattern recognition (psychology) Representation (politics) Speech recognition Machine learning Psychology

Metrics

21
Cited By
14.76
FWCI (Field Weighted Citation Impact)
43
Refs
0.98
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

EEG and Brain-Computer Interfaces
Life Sciences →  Neuroscience →  Cognitive Neuroscience
Emotion and Mood Recognition
Social Sciences →  Psychology →  Experimental and Cognitive Psychology
ECG Monitoring and Analysis
Health Sciences →  Medicine →  Cardiology and Cardiovascular Medicine

Related Documents

JOURNAL ARTICLE

Emotion Recognition Using Multi-Scale Auto-Encoders with Cross Session Adoption

G ChennaKesava Reddyet al.

Journal:   Zenodo (CERN European Organization for Nuclear Research) Year: 2025
JOURNAL ARTICLE

Emotion Recognition Using Multi-Scale Auto-Encoders with Cross Session Adoption

G ChennaKesava Reddyet al.

Journal:   Zenodo (CERN European Organization for Nuclear Research) Year: 2025
BOOK-CHAPTER

MultiMAE: Multi-modal Multi-task Masked Autoencoders

Roman BachmannDavid MizrahiAndrei AtanovAmir Zamir

Lecture notes in computer science Year: 2022 Pages: 348-367
JOURNAL ARTICLE

Multi-Scale Hyperbolic Contrastive Learning for Cross-Subject EEG Emotion Recognition

Chang JiangZhixin ZhangYuhua QianPan Lin

Journal:   IEEE Transactions on Affective Computing Year: 2025 Vol: 16 (3)Pages: 1716-1731
© 2026 ScienceGate Book Chapters — All rights reserved.