Zhifu ZhuXiping YuanShu GanRaobo LiWeidong LuoRui BiCheng Chen
Generative adversarial networks (GANs) possess powerful image translation capabilities. They can transform images acquired from different sensors into a unified domain, effectively mitigating the incomparability problem caused by imaging discrepancies in multimodal remote sensing change detection (CD). However, existing approaches predominantly emphasize domain unification while neglecting the loss of fine-grained features inherent in the translation process, consequently compromising both image translation quality and CD accuracy. To overcome these limitations, we propose a novel texture and structure interaction guided GAN (TSIG-GAN). This network establishes interactive guidance between image texture and structural features through a carefully designed dual-stream cross encoder-decoder architecture, enabling in-depth mining of fine-grained features and significantly improving the fidelity of translated images. Furthermore, to address the spatial scale diversity and complexity of remote sensing images, we develop a multi-scale adaptive feature pyramid (MAFP) module and a contextual semantic interaction guidance (CSIG) mechanism, aiming to further strengthen the model's robust representation of fine-grained features across multiple scales and complex scenes. Specifically, the MAFP module effectively captures spatial details of targets at different resolutions by dynamically integrating multi-scale features, thereby preventing detail loss in small objects due to scale discrepancies. The CSIG mechanism achieves deep interaction between texture and structural features at the contextual semantic level, further promoting their mutual cooperation, thereby enhancing the consistency of fine-grained features representation and semantic integrity in complex scenes. Finally, the translated fine-grained images are fed into a custom CD network to extract changes. To evaluate the effectiveness of the proposed method, we conducted systematic experiments on five representative real-world datasets and performed comparative analysis with sixteen state-of-the-art multimodal CD methods. The experimental results demonstrate that TSIG-GAN achieves significant improvements in both image translation and CD performance, exhibiting superior fine-grained restoration capability and change identification capability.
Zhifu ZhuXiping YuanShu GanRaobo LiRui BiWeidong LuoCheng Chen
Shou-Quan Che Shou-Quan CheJian-Feng Lu Shou-Quan Che
Song GaoDeren LiErting PanHaonan GuoJinjiang WeiJunyi LiuZijie ChenKaimin Sun
Zhifu ZhuXiping YuanShu GanRaobo LiWeidong LuoCheng ChenRui Bi
Baoyu ZhuQunbo LvYuanbo YangKai ZhangXuefu SuiYinhui TangZheng Tan