pith. machine review for the scientific record. sign in

arxiv: 2605.07146 · v1 · submitted 2026-05-08 · 💻 cs.CV

Recognition: 2 theorem links

· Lean Theorem

UniV2D: Bridging Visual Restoration and Semantic Perception for Underwater Salient Object Detection

Authors on Pith no claims yet

Pith reviewed 2026-05-11 01:06 UTC · model grok-4.3

classification 💻 cs.CV
keywords underwater salient object detectionvisual restorationsemantic perceptionunified networkdual-branch architecturejoint optimizationmarine vision
0
0 comments X

The pith

A unified network lets high-level saliency semantics guide low-level image restoration to improve underwater object detection over sequential pipelines.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Underwater images lose contrast and color from absorption and scattering, making it hard to spot key objects. Standard methods clean the image first in one network then run detection in another, but the cleaned result often fails to support accurate detection and can add noise. UniV2D instead trains a single model so that predicted saliency masks steer the restoration steps while the restored details in turn sharpen the saliency output. The design uses staged modules that first produce rough saliency and restored content, then refine both together through cross-level modulation. Experiments on standard underwater benchmarks show higher detection scores than prior separate or joint approaches.

Core claim

UniV2D is a unified vision-to-detection network that jointly optimizes visual restoration and salient object detection. It replaces the conventional enhance-then-detect sequence with a semantic-driven loop: high-level saliency semantics actively guide the restoration process, while the restored visual cues reciprocally enhance saliency perception. The architecture begins with a self-calibrated decoder that produces initial saliency masks and a mask-aware restoration module that reconstructs image content, followed by a saliency-guided refinement module that aligns structural fidelity with semantic consistency.

What carries the argument

Hierarchical dual-branch architecture that couples a self-calibrated decoder for initial saliency prediction with a mask-aware restoration module and a saliency-guided refinement stage using cross-level modulation.

If this is right

  • Restored underwater images become more consistent with the saliency task rather than optimized only for visual quality metrics.
  • The mutual reinforcement loop reduces introduction of task-irrelevant noise during restoration.
  • Cross-level modulation allows structural details and semantic masks to correct each other at multiple scales.
  • The single-model approach eliminates the need to train and align separate restoration and detection networks.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same semantic-guidance idea could be tested on other degraded domains such as low-light or foggy scenes where restoration and recognition interact.
  • Adding explicit physical scattering models into the mask-aware module might further stabilize training without losing the joint benefit.
  • The staged refinement could be adapted to video by propagating saliency masks across frames to maintain temporal consistency during restoration.

Load-bearing premise

That the joint semantic-guided restoration produces images that measurably improve downstream detection accuracy compared with images restored by independent networks.

What would settle it

Detection performance measured on images restored by UniV2D versus the same detector run on images restored by a standalone restoration network; if the scores show no gain or a drop, the joint-guidance premise fails.

Figures

Figures reproduced from arXiv: 2605.07146 by Bo Du, Chang Xu, Kui Jiang, Laibin Chang, Shaodong Wang, Xu Zhang, Yunke Wang.

Figure 1
Figure 1. Figure 1: Comparison of different paradigms for joint Underwater [PITH_FULL_IMAGE:figures/full_fig_p001_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Overview of the proposed UniV2D. It features a semantic-driven dual-branch design, consisting of Self-Calibrated Saliency [PITH_FULL_IMAGE:figures/full_fig_p003_2.png] view at source ↗
Figure 4
Figure 4. Figure 4: Architecture of the MACR module. It utilizes the pre [PITH_FULL_IMAGE:figures/full_fig_p005_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Architecture of the CLFM module. It facilitates bidi [PITH_FULL_IMAGE:figures/full_fig_p005_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Qualitative comparisons between UniV2D and SOTA methods across diverse underwater degradation types ( [PITH_FULL_IMAGE:figures/full_fig_p006_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Qualitative comparisons between UniV2D and SOTA methods across diverse biological categories and object scales ( [PITH_FULL_IMAGE:figures/full_fig_p007_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Efficiency evaluation of each compared method in terms [PITH_FULL_IMAGE:figures/full_fig_p008_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Visual ablation of the SCSM and MACR modules. [PITH_FULL_IMAGE:figures/full_fig_p009_9.png] view at source ↗
read the original abstract

Underwater salient object detection (USOD) plays a vital role in marine vision tasks but remains fundamentally challenging due to severe visual degradation, such as selective absorption and medium scattering. Conventional pipelines typically adopt a sequential "enhance-then-detect" paradigm. However, isolating low-level visual restoration from high-level semantic perception often leads to semantic inconsistency, where the restored images may not be optimal for detection and can even introduce task-irrelevant noise. To break this sequential bottleneck, we propose UniV2D, a Unified Vision-to-Detection Network that jointly optimizes visual restoration and salient object detection within a mutually beneficial framework. Unlike traditional methods that rely on disjointed pipelines or rigid physical priors, UniV2D introduces a semantic-driven learning paradigm: high-level saliency semantics actively guide the restoration process, while the restored visual cues reciprocally enhance saliency perception. Specifically, UniV2D features a hierarchical dual-branch architecture. It first employs a self-calibrated decoder to predict initial saliency masks alongside a mask-aware restoration module to reconstruct image content. Subsequently, a saliency-guided refinement module equipped with cross-level modulation is utilized to align structural fidelity with semantic consistency. Extensive experiments across multiple benchmarks demonstrate that UniV2D significantly outperforms state-of-the-art methods in both quantitative and qualitative evaluations, establishing a new standard for joint underwater perception.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper introduces UniV2D, a unified end-to-end network for underwater salient object detection that jointly optimizes visual restoration and semantic perception via a hierarchical dual-branch architecture. It uses a self-calibrated decoder for initial saliency masks, a mask-aware restoration module, and a saliency-guided refinement module with cross-level modulation to enforce mutual benefit between low-level restoration and high-level detection, addressing semantic inconsistency in sequential enhance-then-detect pipelines. Extensive experiments on multiple benchmarks are claimed to show significant quantitative and qualitative gains over state-of-the-art methods.

Significance. If the reported gains hold under rigorous validation, the work would demonstrate that semantic-driven joint optimization can measurably outperform independent restoration followed by detection in underwater settings, providing a concrete alternative to rigid physical priors and sequential pipelines. This has potential value for marine robotics and vision tasks where degradation is severe.

major comments (2)
  1. [§3.3] §3.3 and Eq. (7): the cross-level modulation mechanism is described at a high level but the precise formulation of how saliency features modulate restoration features (or vice versa) is not fully specified; without this, it is difficult to assess whether the claimed mutual benefit is realized or whether the module reduces to standard feature concatenation.
  2. [§4.3] §4.3, Tables 3-5: the ablation studies isolate the contribution of the saliency-guided refinement but do not include a controlled comparison against a strong sequential baseline that uses the same backbone and training data; this leaves open whether the joint training itself, rather than architectural capacity, drives the reported gains.
minor comments (2)
  1. [§1] The abstract and §1 repeatedly use 'semantic inconsistency' without a precise definition or quantitative measure; a short formalization would clarify the problem the method targets.
  2. [§4.4] Figure 4 caption and §4.4: qualitative examples would benefit from side-by-side restored images from both UniV2D and the strongest competing restoration-then-detection pipeline to visually substantiate the semantic-consistency claim.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive comments, which have helped us improve the clarity and rigor of our manuscript. We address each major comment point by point below.

read point-by-point responses
  1. Referee: [§3.3] §3.3 and Eq. (7): the cross-level modulation mechanism is described at a high level but the precise formulation of how saliency features modulate restoration features (or vice versa) is not fully specified; without this, it is difficult to assess whether the claimed mutual benefit is realized or whether the module reduces to standard feature concatenation.

    Authors: We acknowledge that the description of the cross-level modulation in §3.3 and Equation (7) is presented at a relatively high level. To address this, we will revise the manuscript to provide a more precise and detailed formulation of the modulation process. This will include explicit equations showing how saliency features are used to modulate restoration features (and vice versa) through the cross-level mechanism, clarifying the operations involved and demonstrating that it implements a semantic-guided interaction rather than simple feature concatenation. revision: yes

  2. Referee: [§4.3] §4.3, Tables 3-5: the ablation studies isolate the contribution of the saliency-guided refinement but do not include a controlled comparison against a strong sequential baseline that uses the same backbone and training data; this leaves open whether the joint training itself, rather than architectural capacity, drives the reported gains.

    Authors: We appreciate this point and agree that a controlled comparison would better isolate the benefits of joint training. In the revised manuscript, we will add an ablation experiment that trains a sequential 'enhance-then-detect' pipeline using the identical backbone and training data as UniV2D, but without the proposed dual-branch interactions and cross-level modulation. The performance of this baseline will be reported alongside our ablations in updated Tables 3-5 to directly address whether the joint optimization contributes to the observed gains. revision: yes

Circularity Check

0 steps flagged

No significant circularity detected

full rationale

The paper presents an empirical deep-learning architecture (hierarchical dual-branch network with semantic-driven guidance, self-calibrated decoder, mask-aware restoration, and cross-level modulation) for joint underwater restoration and detection. Its claims rest on end-to-end training and benchmark experiments rather than any closed-form derivation, first-principles prediction, or parameter that is fitted to a subset and then re-used as an output. No equations, uniqueness theorems, or self-citation chains are invoked to force the central result; performance is externally validated against independent datasets and prior methods. The derivation chain is therefore self-contained and does not reduce to its own inputs by construction.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The central claim rests on standard supervised deep-learning assumptions (availability of paired degraded-clean and saliency-annotated underwater images) and the unstated premise that joint optimization yields better task-specific restoration than independent restoration. No new physical axioms or invented entities are introduced.

axioms (1)
  • domain assumption Paired underwater image and saliency ground-truth data exist and are representative of real deployment conditions.
    The method is trained and evaluated on existing benchmarks; performance claims assume these benchmarks capture the target distribution.

pith-pipeline@v0.9.0 · 5556 in / 1201 out tokens · 27294 ms · 2026-05-11T01:06:19.793016+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Reference graph

Works this paper leans on

96 extracted references · 7 canonical work pages · 1 internal anchor

  1. [1]

    Frequency-tuned salient region detec- tion

    Radhakrishna Achanta, Sheila Hemami, Francisco Estrada, and Sabine Susstrunk. Frequency-tuned salient region detec- tion. InProceedings of the IEEE/CVF Conference on Com- puter Vision and Pattern Recognition, pages 1597–1604,

  2. [2]

    Seadiff: Under- water image enhancement with degradation-aware diffusion model.IEEE Transactions on Circuits and Systems for Video Technology, 35(12):12212–12226, 2025

    Hengyue Bi, Long Chen, Jingchao Cao, Jingyang Wang, Jinghao Sun, Yuan Rao, and Junyu Dong. Seadiff: Under- water image enhancement with degradation-aware diffusion model.IEEE Transactions on Circuits and Systems for Video Technology, 35(12):12212–12226, 2025. 2

  3. [3]

    Erd: Encoder-residual- decoder neural network for underwater image enhancement

    Jingchao Cao, Wangzhen Peng, Yutao Liu, Junyu Dong, Patrick Le Callet, and Sam Kwong. Erd: Encoder-residual- decoder neural network for underwater image enhancement. IEEE Transactions on Circuits and Systems for Video Tech- nology, 35(9):8958–8972, 2025. 2

  4. [4]

    Laibin Chang, Huajun Song, Mingjie Li, and Ming Xiang. Uidef: A real-world underwater image dataset and a color- contrast complementary image enhancement framework.IS- PRS Journal of Photogrammetry and Remote Sensing, 196: 415–428, 2023. 1

  5. [5]

    Waterdiffusion: Learning a prior-involved un- rolling diffusion for joint underwater saliency detection and visual restoration

    Laibin Chang, Yunke Wang, Longxiang Deng, Bo Du, and Chang Xu. Waterdiffusion: Learning a prior-involved un- rolling diffusion for joint underwater saliency detection and visual restoration. InProceedings of the AAAI Conference on Artificial Intelligence, pages 1998–2006, 2025. 2, 4, 7

  6. [6]

    Rectan- gling and enhancing underwater stitched image via content- aware warping and perception balancing.Neural Networks, 181:106809, 2025

    Laibin Chang, Yunke Wang, Bo Du, and Chang Xu. Rectan- gling and enhancing underwater stitched image via content- aware warping and perception balancing.Neural Networks, 181:106809, 2025. 7

  7. [7]

    Ma- rine saliency segmenter: Object-focused conditional diffusion with region-level semantic knowledge distillation,

    Laibin Chang, Yunke Wang, JiaXing Huang, Longxiang Deng, Bo Du, and Chang Xu. Marine saliency seg- menter: Object-focused conditional diffusion with region- level semantic knowledge distillation.arXiv preprint arXiv:2504.02391, 2025. 2

  8. [8]

    Color correction meets cross-spectral refinement: a distribution- aware diffusion for underwater image restoration.IEEE Transactions on Multimedia, 2026

    Laibin Chang, Yunke Wang, Bo Du, and Chang Xu. Color correction meets cross-spectral refinement: a distribution- aware diffusion for underwater image restoration.IEEE Transactions on Multimedia, 2026. 2

  9. [9]

    Percep- tual underwater image enhancement with deep learning and physical priors.IEEE Transactions on Circuits and Systems for Video Technology, 31(8):3078–3092, 2020

    Long Chen, Zheheng Jiang, Lei Tong, Zhihua Liu, Aite Zhao, Qianni Zhang, Junyu Dong, and Huiyu Zhou. Percep- tual underwater image enhancement with deep learning and physical priors.IEEE Transactions on Circuits and Systems for Video Technology, 31(8):3078–3092, 2020. 2

  10. [10]

    Bauodnet for class imbal- ance learning in underwater object detection.IEEE Trans- actions on Emerging Topics in Computational Intelligence, 9(3):2452–2461, 2024

    Long Chen, Haohan Yu, Xirui Dong, Yaxin Li, Jialie Shen, Jiangrong Shen, and Qi Xu. Bauodnet for class imbal- ance learning in underwater object detection.IEEE Trans- actions on Emerging Topics in Computational Intelligence, 9(3):2452–2461, 2024. 3

  11. [11]

    Fusion-based channel-wise isotropic convergent real-time underwater im- age enhancement.IEEE Transactions on Circuits and Sys- tems for Video Technology, 35(10):9763–9774, 2025

    Yuehan Chen, Jiqing Zhang, Yafeng Li, Yudong Li, Haom- ing Tang, Huibing Wang, and Xianping Fu. Fusion-based channel-wise isotropic convergent real-time underwater im- age enhancement.IEEE Transactions on Circuits and Sys- tems for Video Technology, 35(10):9763–9774, 2025. 2

  12. [12]

    Point-aware interaction and cnn-induced refinement network for rgb-d salient object detection

    Runmin Cong, Hongyu Liu, Chen Zhang, Wei Zhang, Feng Zheng, Ran Song, and Sam Kwong. Point-aware interaction and cnn-induced refinement network for rgb-d salient object detection. InProceedings of the 31st ACM international con- ference on multimedia, pages 406–416, 2023. 1

  13. [13]

    Trnet: Two- tier recursion network for co-salient object detection.IEEE Transactions on Circuits and Systems for Video Technology, 35(6):5844–5857, 2025

    Runmin Cong, Ning Yang, Hongyu Liu, Dingwen Zhang, Qingming Huang, Sam Kwong, and Wei Zhang. Trnet: Two- tier recursion network for co-salient object detection.IEEE Transactions on Circuits and Systems for Video Technology, 35(6):5844–5857, 2025. 1

  14. [14]

    Uis-mamba: exploring mamba for underwater in- stance segmentation via dynamic tree scan and hidden state weaken

    Runmin Cong, Zongji Yu, Hao Fang, Haoyan Sun, and Sam Kwong. Uis-mamba: exploring mamba for underwater in- stance segmentation via dynamic tree scan and hidden state weaken. InProceedings of the 33rd ACM International Con- ference on Multimedia, pages 343–352, 2025. 3

  15. [15]

    Runmin Cong, Zhiyang Chen, Hao Fang, Sam Kwong, and Wei Zhang. Breaking barriers, localizing saliency: A large- scale benchmark and baseline for condition-constrained salient object detection.IEEE Transactions on Pattern Anal- ysis and Machine Intelligence, 48(4):4167–4183, 2026. 1

  16. [16]

    Frequency- driven diffusion: A hierarchical attention weighting frame- work for underwater image restoration.Computational In- telligence, 41(4):e70095, 2025

    Longxiang Deng, Laibin Chang, and Wei Liu. Frequency- driven diffusion: A hierarchical attention weighting frame- work for underwater image restoration.Computational In- telligence, 41(4):e70095, 2025. 2

  17. [17]

    Recurrent multi-scale transformer for high-resolution salient object detection

    Xinhao Deng, Pingping Zhang, Wei Liu, and Huchuan Lu. Recurrent multi-scale transformer for high-resolution salient object detection. InProceedings of the 31st ACM Interna- tional Conference on Multimedia, pages 7413–7423, 2023. 2, 7

  18. [18]

    Enhanced-alignment measure for binary foreground map evaluation,

    Dengping Fan, Cheng Gong, Yang Cao, Bo Ren, Mingming Cheng, and Ali Borji. Enhanced-alignment measure for bi- nary foreground map evaluation.arXiv:1805.10421, 2018. 7

  19. [19]

    Structure-measure: A new way to evaluate foreground maps

    Deng-Ping Fan, Ming-Ming Cheng, Yun Liu, Tao Li, and Ali Borji. Structure-measure: A new way to evaluate foreground maps. InProceedings of the IEEE/CVF International Con- ference on Computer Vision, pages 4548–4557, 2017. 7

  20. [20]

    Llava-based semantic feature modulation diffusion model for underwater image enhancement.Infor- mation Fusion, page 103566, 2025

    Guodong Fan, Shengning Zhou, Zhen Hua, Jinjiang Li, and Jingchun Zhou. Llava-based semantic feature modulation diffusion model for underwater image enhancement.Infor- mation Fusion, page 103566, 2025. 2

  21. [21]

    Multi-scale and detail-enhanced segment anything model for salient object detection

    Shixuan Gao, Pingping Zhang, Tianyu Yan, and Huchuan Lu. Multi-scale and detail-enhanced segment anything model for salient object detection. InProceedings of the 32nd ACM International Conference on Multimedia, pages 9894–9903, 2024. 1

  22. [22]

    A simple yet effective network based on vision transformer for camouflaged object and salient object detec- tion.IEEE Transactions on Image Processing, 34:608–622,

    Chao Hao, Zitong Yu, Xin Liu, Jun Xu, Huanjing Yue, and Jingyu Yang. A simple yet effective network based on vision transformer for camouflaged object and salient object detec- tion.IEEE Transactions on Image Processing, 34:608–622,

  23. [23]

    Object detection in hyperspectral image 10 via unified spectral–spatial feature aggregation.IEEE Trans- actions on Geoscience and Remote Sensing, 61:1–13, 2023

    Xiao He, Chang Tang, Xinwang Liu, Wei Zhang, Kun Sun, and Jiangfeng Xu. Object detection in hyperspectral image 10 via unified spectral–spatial feature aggregation.IEEE Trans- actions on Geoscience and Remote Sensing, 61:1–13, 2023. 3

  24. [24]

    Multispec- tral object detection via cross-modal conflict-aware learning

    Xiao He, Chang Tang, Xin Zou, and Wei Zhang. Multispec- tral object detection via cross-modal conflict-aware learning. InProceedings of the 31st ACM International Conference on Multimedia, pages 1465–1474, 2023. 3

  25. [25]

    Lin Hong, Xin Wang, De-Sheng Zhang, Ming Zhao, and Hang Xu. Vision-based underwater inspection with portable autonomous underwater vehicle: Development, control, and evaluation.IEEE Transactions on Intelligent Vehicles, 9(1): 2197–2209, 2024. 2

  26. [26]

    Usis16k: high-quality dataset for underwater salient instance segmen- tation.arXiv preprint arXiv:2506.19472, 2025

    Lin Hong, Xin Wang, Yihao Li, and Xia Wang. Usis16k: high-quality dataset for underwater salient instance segmen- tation.arXiv preprint arXiv:2506.19472, 2025. 2

  27. [27]

    Usod10k: a new benchmark dataset for underwater salient object de- tection.IEEE Transactions on Image Processing, 34:1602– 1615, 2025

    Lin Hong, Xin Wang, Gan Zhang, and Ming Zhao. Usod10k: a new benchmark dataset for underwater salient object de- tection.IEEE Transactions on Image Processing, 34:1602– 1615, 2025. 2, 3, 7

  28. [28]

    Cross-modal fusion and progressive decoding network for rgb-d salient object detection.International Journal of Computer Vision, 132(8):3067–3085, 2024

    Xihang Hu, Fuming Sun, Jing Sun, Fasheng Wang, and Hao- jie Li. Cross-modal fusion and progressive decoding network for rgb-d salient object detection.International Journal of Computer Vision, 132(8):3067–3085, 2024. 1

  29. [29]

    Shallow-water image enhancement using relative global histogram stretching based on adaptive parameter acquisition

    Dongmei Huang, Yan Wang, Wei Song, Jean Sequeira, and S´ebastien Mavromatis. Shallow-water image enhancement using relative global histogram stretching based on adaptive parameter acquisition. In2018 International Conference on MultiMedia Modeling, pages 453–465, 2018. 2

  30. [30]

    Contrastive semi-supervised learning for underwa- ter image restoration via reliable bank

    Shirui Huang, Keyan Wang, Huan Liu, Jun Chen, and Yun- song Li. Contrastive semi-supervised learning for underwa- ter image restoration via reliable bank. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18145–18155, 2023. 2, 7

  31. [31]

    Semantic segmentation of underwater im- agery: Dataset and benchmark

    Md Jahidul Islam, Chelsey Edge, Yuyang Xiao, Peigen Luo, Muntaqim Mehtaz, Christopher Morse, Sadman Sakib Enan, and Junaed Sattar. Semantic segmentation of underwater im- agery: Dataset and benchmark. In2020 IEEE/RSJ Interna- tional Conference on Intelligent Robots and Systems, pages 1769–1776, 2020. 7

  32. [32]

    Fast un- derwater image enhancement for improved visual percep- tion.IEEE Robotics and Automation Letters, 5(2):3227– 3234, 2020

    Md Jahidul Islam, Youya Xia, and Junaed Sattar. Fast un- derwater image enhancement for improved visual percep- tion.IEEE Robotics and Automation Letters, 5(2):3227– 3234, 2020. 7

  33. [33]

    Svam: Saliency-guided visual attention modeling by autonomous underwater robots.arXiv:2011.06252, 2021

    Md Jahidul Islam, Ruobing Wang, and Junaed Sattar. Svam: Saliency-guided visual attention modeling by autonomous underwater robots.arXiv:2011.06252, 2021. 7

  34. [34]

    Integrating qdwd with pattern distinctness and local contrast for underwater saliency detection.Journal of Visual Communication and Image Representation, 53:31–41,

    Muwei Jian, Qiang Qi, Junyu Dong, Yilong Yin, and Kin- Man Lam. Integrating qdwd with pattern distinctness and local contrast for underwater saliency detection.Journal of Visual Communication and Image Representation, 53:31–41,

  35. [35]

    Underwater salient object detection via dual- stage self-paced learning and depth emphasis.IEEE Trans- actions on Circuits and Systems for Video Technology, 35(3): 2147–2160, 2025

    Jianhui Jin, Qiuping Jiang, Qingyuan Wu, Binwei Xu, and Runmin Cong. Underwater salient object detection via dual- stage self-paced learning and depth emphasis.IEEE Trans- actions on Circuits and Systems for Video Technology, 35(3): 2147–2160, 2025. 2, 7

  36. [36]

    Un- veiling underwater structures: pyramid saliency detection via homomorphic filtering.Multimedia Tools and Applications, pages 1–18, 2024

    Maria Kanwal, M Mohsin Riaz, and Abdul Ghafoor. Un- veiling underwater structures: pyramid saliency detection via homomorphic filtering.Multimedia Tools and Applications, pages 1–18, 2024. 2

  37. [37]

    Saliency based shape extraction of objects in unconstrained underwater environment.Multimedia Tools and Applica- tions, 78:15121–15139, 2019

    Nitin Kumar, Harish Kumar Sardana, and SN Shome. Saliency based shape extraction of objects in unconstrained underwater environment.Multimedia Tools and Applica- tions, 78:15121–15139, 2019. 2

  38. [38]

    An underwater image enhancement benchmark dataset and beyond.IEEE Trans- actions on Image Processing, 29:4376–4389, 2020

    Chongyi Li, Chunle Guo, Wenqi Ren, Runmin Cong, Junhui Hou, Sam Kwong, and Dacheng Tao. An underwater image enhancement benchmark dataset and beyond.IEEE Trans- actions on Image Processing, 29:4376–4389, 2020. 7

  39. [39]

    Uwsam: Segment anything model guided under- water instance segmentation and a large-scale benchmark dataset.arXiv e-prints, pages arXiv–2505, 2025

    Hua Li, Shijie Lian, Zhiyuan Li, Runmin Cong, and Sam Kwong. Uwsam: Segment anything model guided under- water instance segmentation and a large-scale benchmark dataset.arXiv e-prints, pages arXiv–2505, 2025. 2

  40. [40]

    Fscdiff: Frequency-spatial entangled conditional diffusion model for underwater salient object detection

    Hua Li, Gaowei Lin, Zhiyuan Li, Sam Kwong, and Run- min Cong. Fscdiff: Frequency-spatial entangled conditional diffusion model for underwater salient object detection. In Proceedings of the 33rd ACM International Conference on Multimedia, pages 8379–8388, 2025. 1

  41. [41]

    Kunqian Li, Hongtao Fan, Qi Qi, Chi Yan, Kun Sun, and QM Jonathan Wu. Tctl-net: Template-free color transfer learning for self-attention driven underwater image enhance- ment.IEEE Transactions on Circuits and Systems for Video Technology, 34(6):4682–4697, 2024. 7

  42. [42]

    Efficient underwater object de- tection with enhanced feature extraction and fusion.IEEE Transactions on Industrial Informatics, 21(6):4904–4914,

    Shaoming Li, Ziyi Wang, Rong Dai, Yaqing Wang, Fangxun Zhong, and Yunhui Liu. Efficient underwater object de- tection with enhanced feature extraction and fusion.IEEE Transactions on Industrial Informatics, 21(6):4904–4914,

  43. [43]

    Watermask: Instance segmentation for under- water imagery

    Shijie Lian, Hua Li, Runmin Cong, Suqi Li, Wei Zhang, and Sam Kwong. Watermask: Instance segmentation for under- water imagery. InProceedings of the IEEE/CVF Interna- tional Conference on Computer Vision, pages 1305–1315,

  44. [44]

    Diving into underwater: Segment anything model guided underwater salient instance segmentation and a large-scale dataset.arXiv preprint arXiv:2406.06039, 2024

    Shijie Lian, Ziyi Zhang, Hua Li, Wenjie Li, Laurence Tian- ruo Yang, Sam Kwong, and Runmin Cong. Diving into underwater: Segment anything model guided underwater salient instance segmentation and a large-scale dataset.arXiv preprint arXiv:2406.06039, 2024. 2

  45. [45]

    Sam-daq: Seg- ment anything model with depth-guided adaptive queries for rgb-d video salient object detection

    Jia Lin, Xiaofei Zhou, Jiyuan Liu, Runmin Cong, Guo- dao Zhang, Zhi Liu, and Jiyong Zhang. Sam-daq: Seg- ment anything model with depth-guided adaptive queries for rgb-d video salient object detection. InProceedings of the AAAI Conference on Artificial Intelligence, pages 6952– 6960, 2026. 1

  46. [46]

    Twin adversarial contrastive learning for underwater image enhancement and beyond.IEEE Transactions on Image Pro- cessing, 31:4922–4936, 2022

    Risheng Liu, Zhiying Jiang, Shuzhou Yang, and Xin Fan. Twin adversarial contrastive learning for underwater image enhancement and beyond.IEEE Transactions on Image Pro- cessing, 31:4922–4936, 2022. 2, 3

  47. [47]

    Auto-usod: searching topology for underwa- ter salient object detection

    Tingwei Liu, Runyu Wang, Miao Zhang, Yongri Piao, and Huchuan Lu. Auto-usod: searching topology for underwa- ter salient object detection. InChinese Conference on Pat- tern Recognition and Computer Vision (PRCV), pages 3–16,

  48. [48]

    Toward better than pseudo-reference in underwater image enhancement.IEEE Transactions on Image Processing, 34: 6168–6179, 2025

    Yi Liu, Qiuping Jiang, Xingbo Li, Ting Luo, and Wenqi Ren. Toward better than pseudo-reference in underwater image enhancement.IEEE Transactions on Image Processing, 34: 6168–6179, 2025. 2 11

  49. [49]

    Hdanet: Enhancing underwa- ter salient object detection with physics-inspired multimodal joint learning.IEEE Transactions on Geoscience and Re- mote Sensing, 63:1–14, 2025

    Yiwen Liu, Xiaoyu Zhang, Jinchao Zhu, Biting Ma, Yu- tai Duan, and Panlong Tan. Hdanet: Enhancing underwa- ter salient object detection with physics-inspired multimodal joint learning.IEEE Transactions on Geoscience and Re- mote Sensing, 63:1–14, 2025. 2

  50. [50]

    Un- derwater image enhancement method based on denoising diffusion probabilistic model.Journal of Visual Communi- cation and Image Representation, 96:103926, 2023

    Siqi Lu, Fengxu Guan, Hanyu Zhang, and Haitao Lai. Un- derwater image enhancement method based on denoising diffusion probabilistic model.Journal of Visual Communi- cation and Image Representation, 96:103926, 2023. 2

  51. [51]

    Stamf: Synergistic trans- former and mamba fusion network for rgb-polarization based underwater salient object detection.Information Fusion, 122:103182, 2025

    Qianwen Ma, Xiaobo Li, Bincheng Li, Zhen Zhu, Jing Wu, Feng Huang, and Haofeng Hu. Stamf: Synergistic trans- former and mamba fusion network for rgb-polarization based underwater salient object detection.Information Fusion, 122:103182, 2025. 1

  52. [52]

    Human-visual- system-inspired underwater image quality measures.IEEE Journal of Oceanic Engineering, 41(3):541–551, 2016

    Karen Panetta, Chen Gao, and Sos Agaian. Human-visual- system-inspired underwater image quality measures.IEEE Journal of Oceanic Engineering, 41(3):541–551, 2016. 7

  53. [53]

    U-shape trans- former for underwater image enhancement.IEEE Transac- tions on Image Processing, 32:3066–3079, 2023

    Lintao Peng, Chunli Zhu, and Liheng Bian. U-shape trans- former for underwater image enhancement.IEEE Transac- tions on Image Processing, 32:3066–3079, 2023. 7

  54. [54]

    Blurriness-guided underwater salient object detection and data augmentation.IEEE Journal of Oceanic Engineer- ing, 49(3):1089–1103, 2024

    Yan-Tsung Peng, Yu-Cheng Lin, Wen-Yi Peng, and Chen-Yu Liu. Blurriness-guided underwater salient object detection and data augmentation.IEEE Journal of Oceanic Engineer- ing, 49(3):1089–1103, 2024. 2

  55. [55]

    Saliency filters: Contrast based filtering for salient region detection

    Federico Perazzi, Philipp Kr ¨ahenb¨uhl, Yael Pritch, and Alexander Hornung. Saliency filters: Contrast based filtering for salient region detection. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 733–740, 2012. 7

  56. [56]

    Mfnet: Multi-filter directive network for weakly supervised salient object detection

    Yongri Piao, Jian Wang, Miao Zhang, and Huchuan Lu. Mfnet: Multi-filter directive network for weakly supervised salient object detection. InProceedings of the IEEE/CVF International Conference on Computer Vision, pages 4136– 4145, 2021. 7

  57. [57]

    Underwater image co-enhancement with correlation feature matching and joint learning.IEEE Transactions on Circuits and Systems for Video Technology, 32(3):1133–1147, 2022

    Qi Qi, Yongchang Zhang, Fei Tian, QM Jonathan Wu, Kun- qian Li, Xin Luan, and Dalei Song. Underwater image co-enhancement with correlation feature matching and joint learning.IEEE Transactions on Circuits and Systems for Video Technology, 32(3):1133–1147, 2022. 2

  58. [58]

    U2-net: Go- ing deeper with nested u-structure for salient object detec- tion.Pattern Recognition, 106:107404, 2020

    Xuebin Qin, Zichen Zhang, Chenyang Huang, Masood De- hghan, Osmar R Zaiane, and Martin Jagersand. U2-net: Go- ing deeper with nested u-structure for salient object detec- tion.Pattern Recognition, 106:107404, 2020. 7

  59. [59]

    Dif- fuie: Learning latent global priors in diffusion models for underwater image enhancement.IEEE Transactions on Mul- timedia, pages 1–14, 2024

    Yuhao Qing, Si Liu, Hai Wang, and Yueying Wang. Dif- fuie: Learning latent global priors in diffusion models for underwater image enhancement.IEEE Transactions on Mul- timedia, pages 1–14, 2024. 2

  60. [60]

    U-net: Convolutional networks for biomedical image segmentation

    Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. InInternational Conference on Medical Image Computing and Computer-Assisted Intervention, pages 234–241, 2015. 3

  61. [61]

    Very Deep Convolutional Networks for Large-Scale Image Recognition

    Karen Simonyan and Andrew Zisserman. Very deep convo- lutional networks for large-scale image recognition.arXiv preprint arXiv:1409.1556, 2014. 6

  62. [62]

    Huajun Song, Laibin Chang, Ziwei Chen, and Peng Ren. Enhancement-registration-homogenization (erh): A compre- hensive underwater visual reconstruction paradigm.IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(10):6953–6967, 2022. 2

  63. [63]

    Dual-model: Revised imaging network and visual perception correction for underwater image enhancement.Engineering Applications of Artificial Intelligence, 125:106731, 2023

    Huajun Song, Laibin Chang, Hao Wang, and Peng Ren. Dual-model: Revised imaging network and visual perception correction for underwater image enhancement.Engineering Applications of Artificial Intelligence, 125:106731, 2023. 2

  64. [64]

    Catnet: A cascaded and aggregated transformer network for rgb-d salient object detection.IEEE Transac- tions on Multimedia, 26:2249–2262, 2023

    Fuming Sun, Peng Ren, Bowen Yin, Fasheng Wang, and Haojie Li. Catnet: A cascaded and aggregated transformer network for rgb-d salient object detection.IEEE Transac- tions on Multimedia, 26:2249–2262, 2023. 7

  65. [65]

    Divide-and-conquer: Confluent triple-flow net- work for rgb-t salient object detection.IEEE Transactions on Pattern Analysis and Machine Intelligence, 47(3):1958– 1974, 2024

    Hao Tang, Zechao Li, Dong Zhang, Shengfeng He, and Jin- hui Tang. Divide-and-conquer: Confluent triple-flow net- work for rgb-t salient object detection.IEEE Transactions on Pattern Analysis and Machine Intelligence, 47(3):1958– 1974, 2024. 1

  66. [66]

    Hao Wang, Shixin Sun, Laibin Chang, Huanyu Li, Wenwen Zhang, Alejandro C Frery, and Peng Ren. Inspiration: A re- inforcement learning-based human visual perception-driven image enhancement paradigm for underwater scenes.Engi- neering Applications of Artificial Intelligence, 133:108411,

  67. [67]

    Pixels, regions, and objects: Multiple en- hancement for salient object detection

    Yi Wang, Ruili Wang, Xin Fan, Tianzhu Wang, and Xi- angjian He. Pixels, regions, and objects: Multiple en- hancement for salient object detection. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10031–10040, 2023. 1

  68. [68]

    Domain adaptation for underwater im- age enhancement.IEEE Transactions on Image Processing, 32:1442–1457, 2023

    Zhengyong Wang, Liquan Shen, Mai Xu, Mei Yu, Kun Wang, and Yufei Lin. Domain adaptation for underwater im- age enhancement.IEEE Transactions on Image Processing, 32:1442–1457, 2023. 2

  69. [69]

    High- resolution underwater creature segmentation.IEEE Trans- actions on Image Processing, 34:7759–7772, 2025

    Huiyang Wu, Qiuping Jiang, Zongwei Wu, Runmin Cong, C´edric Demonceaux, Yi Yang, and Xiangyang Ji. High- resolution underwater creature segmentation.IEEE Trans- actions on Image Processing, 34:7759–7772, 2025. 2

  70. [70]

    Effiseanet: Pioneering lightweight network for underwater salient object detection

    Qingyao Wu, Zhenqi Fu, Hong Lin, Chenyu Ma, Xiaotong Tu, and Xinghao Ding. Effiseanet: Pioneering lightweight network for underwater salient object detection. InPro- ceedings of the Asian Conference on Computer Vision, pages 1486–1501, 2024. 2

  71. [71]

    An underwater color im- age quality evaluation metric.IEEE Transactions on Image Processing, 24(12):6062–6071, 2015

    Miao Yang and Arcot Sowmya. An underwater color im- age quality evaluation metric.IEEE Transactions on Image Processing, 24(12):6062–6071, 2015. 7

  72. [72]

    Saliency detection of turbid underwater im- ages based on depth attention adversarial network

    Shudi Yang, Xing Cui, Sen Zhu, Senqi Tan, Jiaxiong Wu, and Fu Chang. Saliency detection of turbid underwater im- ages based on depth attention adversarial network. InIn- ternational Conference on Autonomous Unmanned Systems, pages 154–163, 2023. 2

  73. [73]

    If-usod: Mul- timodal information fusion interactive feature enhancement architecture for underwater salient object detection.Infor- mation Fusion, 117:102806, 2025

    Genji Yuan, Jintao Song, and Jinjiang Li. If-usod: Mul- timodal information fusion interactive feature enhancement architecture for underwater salient object detection.Infor- mation Fusion, 117:102806, 2025. 2

  74. [74]

    Heterogeneous experts and hierarchical perception for un- derwater salient object detection.IEEE Transactions on Im- age Processing, 34:3703–3717, 2025

    Mingfeng Zha, Guoqing Wang, Yunqiang Pei, Tianyu Li, Xiongxin Tang, Chongyi Li, Yang Yang, and Heng Tao Shen. Heterogeneous experts and hierarchical perception for un- derwater salient object detection.IEEE Transactions on Im- age Processing, 34:3703–3717, 2025. 2, 3 12

  75. [75]

    Cross-modality discrepant in- teraction network for rgb-d salient object detection

    Chen Zhang, Runmin Cong, Qinwei Lin, Lin Ma, Feng Li, Yao Zhao, and Sam Kwong. Cross-modality discrepant in- teraction network for rgb-d salient object detection. InPro- ceedings of the 29th ACM international conference on mul- timedia, pages 2094–2102, 2021. 1

  76. [76]

    Synergistic multiscale detail refinement via intrinsic supervision for underwater image enhancement

    Dehuan Zhang, Jingchun Zhou, Chunle Guo, Weishi Zhang, and Chongyi Li. Synergistic multiscale detail refinement via intrinsic supervision for underwater image enhancement. InProceedings of the AAAI Conference on Artificial Intelli- gence, pages 7033–7041, 2024. 7

  77. [77]

    Fantastic animals and where to find them: segment any ma- rine animal with dual sam

    Pingping Zhang, Tianyu Yan, Yang Liu, and Huchuan Lu. Fantastic animals and where to find them: segment any ma- rine animal with dual sam. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2578–2587, 2024. 7

  78. [78]

    Underwater image enhance- ment via minimal color loss and locally adaptive contrast en- hancement.IEEE Transactions on Image Processing, 31: 3997–4010, 2022

    Weidong Zhang, Peixian Zhuang, Hai-Han Sun, Guohou Li, Sam Kwong, and Chongyi Li. Underwater image enhance- ment via minimal color loss and locally adaptive contrast en- hancement.IEEE Transactions on Image Processing, 31: 3997–4010, 2022. 7

  79. [79]

    Underwater image enhancement via princi- pal component fusion of foreground and background.IEEE Transactions on Circuits and Systems for Video Technology, 34(11):10930–10943, 2024

    Weidong Zhang, Qingmin Liu, Yikun Feng, Lei Cai, and Peixian Zhuang. Underwater image enhancement via princi- pal component fusion of foreground and background.IEEE Transactions on Circuits and Systems for Video Technology, 34(11):10930–10943, 2024. 2

  80. [80]

    Underwater im- age enhancement via weighted wavelet visual perception fu- sion.IEEE Transactions on Circuits and Systems for Video Technology, 34(4):2469–2483, 2024

    Weidong Zhang, Ling Zhou, Peixian Zhuang, Guohou Li, Xipeng Pan, Wenyi Zhao, and Chongyi Li. Underwater im- age enhancement via weighted wavelet visual perception fu- sion.IEEE Transactions on Circuits and Systems for Video Technology, 34(4):2469–2483, 2024. 7

Showing first 80 references.