Semantic segmentation of high-resolution remote sensing images (RSIs) is developing rapidly. Multispectral images can provide rich spectral information for semantic segmentation, while 3D LiDAR point cloud data can provide depth information. Thus, semantic segmentation accuracy could be improved by fusing multispectral images and 3D LiDAR point cloud. In this paper, we propose a method titled Direct LiDAR-Aerial Fusion Network (DLAFNet) which directly uses RSIs and LiDAR point cloud for semantic segmentation tasks. In particular, owing to the fact that sparse features extracted from the KPConv branch are not as essential as features from RSIs, we design LiDAR Assisted Attention Module (L-AAM). Our experiments on the modified GRSS18 dataset prove that our method is proper and can obtain the best results by comparing with its components and other methods.
Wei LiuHe WangYicheng QiaoHaopeng ZhangJunli Yang
Poliyapram VinayarajWeimin WangRyosuke Nakamura
Ali AbdelKaderMohamed Moustafa
Rui XiangMu ZhouNan DuXiaolong Yang
Lu RenJianwei NiuZhenchao OuyangZhibin ZhangSiyi Zheng