Cheng YuJunfang ZhouLin WangGuizhen LiuZhongjun Ding
Images captured underwater frequently suffer from color casts, blurring, and distortion, which are mainly attributable to the unique optical characteristics of water. Although conventional UIE methods rooted in physics are available, their effectiveness is often constrained, particularly in challenging aquatic and illumination conditions. More recently, deep learning has become a leading paradigm for UIE, recognized for its superior performance and operational efficiency. This paper proposes UCA-Net, a lightweight CNN-Transformer hybrid network. It incorporates multiple attention mechanisms and utilizes composite attention to effectively enhance textures, reduce blur, and correct color. A novel adaptive sparse self-attention module is introduced to jointly restore global color consistency and fine local details. The model employs a U-shaped encoder–decoder architecture with three-stage up- and down-sampling, facilitating multi-scale feature extraction and global context fusion for high-quality enhancement. Experimental results on multiple public datasets demonstrate UCA-Net’s superior performance, achieving a PSNR of 24.75 dB and an SSIM of 0.89 on the UIEB dataset, while maintaining an extremely low computational cost with only 1.44M parameters. Its effectiveness is further validated by improvements in various downstream image tasks. UCA-Net achieves an optimal balance between performance and efficiency, offering a robust and practical solution for underwater vision applications.
Jyotirmaya TembhurneRahul Katarya
Gang LiJiaqing FanChen‐Yu Cheng
Wen ZhangYikang ZhaoFeng GaoHao SuYuan RaoJunyu Dong
Pengfei TangLiangliang LiYuan XueMing LvZhenhong JiaHongbing Ma