WaveUIE: Policy-Driven Contrastive Learning for Underwater Image Enhancement in the Wavelet Domain

1 Beijing University of Posts and Telecommunications, 2 South China University of Technology
3 University of Oxford, 4 Guangzhou University

# Indicates Corresponding Author

🌊✨ WaveUIE enhances raw YouTube wild underwater video into clear visuals ⬅️➡️

Abstract

Underwater image enhancement remains challenging in computer vision due to its complex and diverse degradation phenomena. Existing state-of-the-art methods mainly focus on spatial domain modelling and demonstrate superiority, but they fall short in two key aspects: 1) underutilization of frequency domain features, and 2) insufficient exploitation of the relationship between the positive and negative samples during the training. To overcome these limitations, we propose a wavelet-driven framework called WaveUIE. Specifically, our method decomposes images into high/low-frequency components via the wavelet transform. To better mine the features of these two component patterns, a high-frequency enhancement block (HFEBlock) is introduced to restore fine textures of the high-frequency component. In addition, we also introduce a low-frequency enhancement block (LFEBlock) to effectively eliminates color casts and blur of the low-frequency features. Moreover, we design a dynamic frequency fusion block (DFFBlock), which orchestrates cross-frequency interactions by first calibrating high-frequency data with the enhanced low-frequency features, and then synthesising them into coherent visual outputs. Finally, we design a wavelet fine-grained contrastive policy (WFCPolicy) to decompose images into four sub-bands and impose contrastive constraints on the enhanced result (anchor) within the sub-band space. These constraints pull the anchor closer to the ground-truth (positive samples) while pushing it away from the original degraded image (negative samples), thereby explicitly leveraging the degradation features present in different frequency sub-bands to improve enhancement performance. Experimental evaluations demonstrate state-of-the-art performance across multiple benchmarks, with significant improvements in both SSIM and UIQM metrics over existing methods.

Frequency-Domain Motivation

Wavelet Domain Analysis
Comparative wavelet analysis revealing two critical observations: (1) Low-frequency components (LL-subband) dominate color distribution and global structure preservation, while (2) high-frequency components (HL/HH/LH-subbands) exhibit localized impact on texture details. Hybrid reconstruction experiments demonstrate that low-frequency substitution significantly alters RGB histograms whereas high-frequency exchange preserves overall color characteristics.

Method Overview

WaveUIE Framework
Architecture of WaveUIE, directly motivated by our wavelet-domain observations: (1) LFEBlock conducts global compensation on low-frequency components (dominant in color/structure), (2) HFEBlock performs local refinement of high-frequency details (minor texture residuals), (3) DFFBlock orchestrates cross-frequency guidance where enhanced low-frequency features calibrate high-frequency restoration, followed by coherent multi-band fusion.

Experimental Results

UIEB Test Results
Visual comparisons on underwater images sampled from UIEB dataset. (a) RAWS. (b) Fusion. (c) WaterNet. (d) UWCNN-Type I. (e) UIEC²Net. (f) PUIE-Net. (g) U-Shape. (h) P2CNet. (i) Proposed WaveUIE. (j) Ground truth.
UIEB Challenge Results
Visual comparisons on UIEB Challenge dataset. (a) RAWS. (b) Fusion. (c) WaterNet. (d) UWCNN-Type I. (e) UIEC²Net. (f) PUIE-Net. (g) U-Shape. (h) P2CNet. (i) Proposed WaveUIE.
LSUI Results
Visual comparisons on LSUI dataset. (a) RAWS. (b) Fusion. (c) WaterNet. (d) UWCNN-Type I. (e) UIEC²Net. (f) PUIE-Net. (g) U-Shape. (h) P2CNet. (i) Proposed WaveUIE. (j) Ground truth.

BibTeX

BibTex Code Here