The attenuation, scattering, and absorption of light in water make underwater imaging a challenging task, and global contextual information is key to enhancing underwater images. However, most existing algorithms rely on simple convolutional neural networks (CNNs), which inherently have locality due to convolutional operations, are difficult to directly obtain global contextual information. The problem is addressed by embedding Swin-Transformer, which has global modeling capability, into U-Net. Specifically, a feature fusion module is constructed to change the way feature propagation occurs between network layers and to fuse context feature information between different network layers. The proposed method is evaluated on the constructed dataset. Experimental results show that the proposed method outperforms existing methods in objective metrics such as peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM), and significantly improved the quality and visibility of underwater images.
Yuzhen LiuMeiyi LiuSen LinZhiyong Tao
Shuai ShenHaoyi WangWeitao ChenPingkang WangQianyong LiangXuwen Qin
Jiaxing ZhangJiawei WuZuoyong LiShenghua TengFeng Guo
Peng LinZihao FanYafei WangXudong SunYuán-RuìYángXianping Fu