JOURNAL ARTICLE

Multi-Modal Graph-Aware Transformer with Contrastive Fusion for Brain Tumor Segmentation

Rini ChowdhuryPrashant KumarR. SuganthiV. AmmuR. Evance LeethialC. Roopa

Year: 2025 Journal:   Journal of Electronics Electromedical Engineering and Medical Informatics Vol: 7 (4)Pages: 1226-1239

Abstract

Accurate segmentation of brain tumors in MRI images is critical for early diagnosis, surgical planning, and effective treatment strategies. Traditional deep learning models such as U-Net, Attention U-Net, and Swin-U-Net have demonstrated commendable success in tumor segmentation by leveraging Convolutional Neural Networks (CNNs) and transformer-based encoders. However, these models often fall short in effectively capturing complex inter-modality interactions and long-range spatial dependencies, particularly in tumor regions with diffuse or poorly defined boundaries. Additionally, they suffer from limited generalization capabilities and demand substantial computational resources. AIM: To overcome these limitations, a novel approach named Graph-Aware Transformer with Contrastive Fusion (GAT-CF) is introduced. This model enhances segmentation performance by integrating spatial attention mechanisms of transformers with graph-based relational reasoning across multiple MRI modalities, namely T1, T2, FLAIR, and T1CE. The graph-aware structure models inter-slice and intra-slice relationships more effectively, promoting better structural understanding of tumor regions. Furthermore, a multi-modal contrastive learning strategy is employed to align semantic features and distinguish complementary modality-specific information, thereby improving the model’s discriminative power. The fusion of these techniques facilitates improved contextual understanding and more accurate boundary delineation in complex tumor regions. When evaluated on the BraTS2021 dataset, the proposed GAT-CF model achieved a Dice score of 99.1% and an IoU of 98.4%, surpassing the performance of state-of-the-art architectures like Swin-UNet and SegResNet. It also demonstrated superior accuracy in detecting and enhancing tumor voxels and core tumor regions, highlighting its robustness, precision, and potential for clinical adoption in neuroimaging applications

Keywords:

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
0
Refs
0.50
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Brain Tumor Detection and Classification
Life Sciences →  Neuroscience →  Neurology
Advanced Computing and Algorithms
Social Sciences →  Social Sciences →  Urban Studies
Medical Image Segmentation Techniques
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

BOOK-CHAPTER

Multi-modal Transformer for Brain Tumor Segmentation

Ji‐Hoon ChoJinah Park

Lecture notes in computer science Year: 2023 Pages: 138-148
JOURNAL ARTICLE

RFNet: Region-aware Fusion Network for Incomplete Multi-modal Brain Tumor Segmentation

Yuhang DingXin YuYi Yang

Journal:   2021 IEEE/CVF International Conference on Computer Vision (ICCV) Year: 2021 Pages: 3955-3964
JOURNAL ARTICLE

Boundary-aware and cross-modal fusion network for enhanced multi-modal brain tumor segmentation

Tongxue Zhou

Journal:   Pattern Recognition Year: 2025 Vol: 165 Pages: 111637-111637
JOURNAL ARTICLE

Multi-modal brain tumor segmentation via disentangled representation learning and region-aware contrastive learning

Tongxue Zhou

Journal:   Pattern Recognition Year: 2024 Vol: 149 Pages: 110282-110282
© 2026 ScienceGate Book Chapters — All rights reserved.