JOURNAL ARTICLE

Auditory Attention Decoding with Task-Related Multi-View Contrastive Learning

Abstract

The human brain can easily focus on one speaker and suppress others in scenarios such as a cocktail party. Recently, researchers found that auditory attention can be decoded from the electroencephalogram (EEG) data. However, most existing deep learning methods are difficult to use prior knowledge of different views (that is attended speech and EEG are task-related views) and extract an unsatisfactory representation. Inspired by Broadbent's filter model, we decode auditory attention in a multi-view paradigm and extract the most relevant and important information utilizing the missing view. Specifically, we propose an auditory attention decoding (AAD) method based on multi-view VAE with task-related multi-view contrastive (TMC) learning. Employing TMC learning in multi-view VAE can utilize the missing view to accumulate prior knowledge of different views into the fusion of representation, and extract the approximate task-related representation. We examine our method on two popular AAD datasets, and demonstrate the superiority of our method by comparing it to the state-of-the-art method.

Keywords:
Computer science Decoding methods Task (project management) Representation (politics) Focus (optics) Artificial intelligence Feature learning Speech recognition Electroencephalography Filter (signal processing) Multi-task learning Machine learning Natural language processing Psychology Computer vision

Metrics

3
Cited By
0.79
FWCI (Field Weighted Citation Impact)
27
Refs
0.66
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

EEG and Brain-Computer Interfaces
Life Sciences →  Neuroscience →  Cognitive Neuroscience
Blind Source Separation Techniques
Physical Sciences →  Computer Science →  Signal Processing
Neural dynamics and brain function
Life Sciences →  Neuroscience →  Cognitive Neuroscience
© 2026 ScienceGate Book Chapters — All rights reserved.