Weitao LiHui GaoYi SuBiffon Manyura Momanyi
With the development of deep learning, the performance of image semantic segmentation in remote sensing has been constantly improved. However, the performance usually degrades while testing on different datasets because of the domain gap. To achieve feasible performance, extensive pixel-wise annotations are acquired in a new environment, which is time-consuming and labor-intensive. Therefore, unsupervised domain adaptation (UDA) has been proposed to alleviate the effort of labeling. However, most previous approaches are based on outdated network architectures that hinder the improvement of performance in UDA. Since the effects of recent architectures for UDA have been barely studied, we reveal the potential of Transformer in UDA for remote sensing with a self-training framework. Additionally, two training strategies have been proposed to enhance the performance of UDA: (1) Gradual Class Weights (GCW) to stabilize the model on the source domain by addressing the class-imbalance problem; (2) Local Dynamic Quality (LDQ) to improve the quality of the pseudo-labels via distinguishing the discrete and clustered pseudo-labels on the target domain. Overall, our proposed method improves the state-of-the-art performance by 8.23% mIoU on Potsdam→Vaihingen and 9.2% mIoU on Vaihingen→Potsdam and facilitates learning even for difficult classes such as clutter/background.
Xianping MaXiaokang ZhangXingchen DingMan-On PunSiwei Ma
Kangjian CaoSheng WangZiheng WeiKexin ChenRong‐Guey ChangFu Xu
Yuxiang CaiYingchun YangQiyi ZhengZhengwei ShenYongheng ShangJianwei YinZhongtian Shi
Yikun LiuXudong KangYuwen HuangKuikui WangGongping Yang
Hao CuiGuo ZhangJi QiHaifeng LiChao TaoXue LiShasha HouDeren Li