Rui YangDi ZhangYanhe GuoShuang Wang
Cross-modal remote sensing image-text retrieval (CMRSITR) can retrieve images of interest from a vast amount of remote sensing images and has received significant attention in recent years. However, existing methods do not consider saliency and texture information, which are essential for remote sensing images when extracting image features. Therefore, this paper proposes a novel texture and saliency enhanced image learning method for CMRSITR. We constructed a multi-task image feature extractor in this new method. A texture map and a saliency map are created by extracting texture and detecting the saliency of each RS image. Both maps are set as supervised information during training to make the extracted saliency and texture features gradually reconstructed to a saliency map and a texture map, respectively. At the same time, the retrieval features of each RS image are obtained from the retrieval feature branch of the image. Experiments conducted on two commonly used CMRSITR datasets, RSICD and UCM, showed that the proposed method is effective in improving retrieval performance and achieved state-of-the-art retrieval performance compared to existing methods.
Jie ShaoYiran XiePengda WangGuohao Feng
Zhuoyue WangXueqian WangGang Li
W ZhangJihao LiShuoke LiJialiang ChenWenkai ZhangXin GaoXian Sun
Zuopeng ZhaoXiaoran MiaoChen HeJianfeng HuBingbing MinYumeng GaoYing LiuKanyaphakphachsorn Pharksuwan
Lingxin XuLuyao WangJinzhi ZhangDa HaHaisu Zhang