JOURNAL ARTICLE

Adjacent-Scale Multimodal Fusion Networks for Semantic Segmentation of Remote Sensing Data

Xianping MaXichen XuXiaokang ZhangMan-On Pun

Year: 2024 Journal:   IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing Vol: 17 Pages: 20116-20128   Publisher: Institute of Electrical and Electronics Engineers

Abstract

Semantic segmentation is a fundamental task in remote sensing image analysis. The accurate delineation of objects within such imagery serves as the cornerstone for a wide range of applications. To address this issue, edge detection, cross-modal data, large intraclass variability, and limited interclass variance must be considered. Traditional convolutional-neural-network-based models are notably constrained by their local receptive fields, Nowadays, transformer-based methods show great potential to learn features globally, while they ignore positional cues easily and are still unable to cope with multimodal data. Therefore, this work proposes an adjacent-scale multimodal fusion network (ASMFNet) for semantic segmentation of remote sensing data. ASMFNet stands out not only for its innovative interaction mechanism across adjacent-scale features, effectively capturing contextual cues while maintaining low computational complexity but also for its remarkable cross-modal capability. It seamlessly integrates different modalities, enriching feature representation. Its hierarchical scale attention (HSA) module bolsters the association between ground objects and their surrounding scenes through learning discriminative features at higher level abstractions, thereby linking the broad structural information. Adaptive modality fusion module is equipped by HSA with valuable insights into the interrelationships between cross-model data, and it assigns spatial weights at the pixel level and seamlessly integrates them into channel features to enhance fusion representation through an evaluation of modality importance via feature concatenation and filtering. Extensive experiments on representative remote sensing semantic segmentation datasets, including the ISPRS Vaihingen and Potsdam datasets, confirm the impressive performance of the proposed ASMFNet.

Keywords:
Computer science Segmentation Scale (ratio) Sensor fusion Artificial intelligence Fusion Remote sensing Geology Cartography Geography

Metrics

13
Cited By
8.30
FWCI (Field Weighted Citation Impact)
63
Refs
0.96
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Advanced Computational Techniques and Applications
Physical Sciences →  Computer Science →  Artificial Intelligence
Remote-Sensing Image Classification
Physical Sciences →  Engineering →  Media Technology

Related Documents

BOOK-CHAPTER

Deep Multimodal Fusion for Semantic Segmentation of Remote Sensing Earth Observation Data

Ivica DimitrovskiVlatko SpasevIvan Kitanovski

Communications in computer and information science Year: 2025 Pages: 106-120
JOURNAL ARTICLE

Learning Frequency-Domain Fusion for Multimodal Remote Sensing Semantic Segmentation

Guangsheng ChenFangyu SunWeipeng JingWeitao ZouDonglin DiYang SongLei Fan

Journal:   IEEE Transactions on Geoscience and Remote Sensing Year: 2025 Vol: 63 Pages: 1-16
JOURNAL ARTICLE

A Multilevel Multimodal Fusion Transformer for Remote Sensing Semantic Segmentation

Xianping MaXiaokang ZhangMan-On PunMing Liu

Journal:   IEEE Transactions on Geoscience and Remote Sensing Year: 2024 Vol: 62 Pages: 1-15
© 2026 ScienceGate Book Chapters — All rights reserved.