Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Underwater scene prior inspired deep underwater image and video enhancement

Published: 01 February 2020 Publication History

Highlights

Underwater image and video synthesis approach is desired by data-driven methods.
Underwater scene prior is helpful for underwater image and video enhancement.
Light-weight network structure can be easily extended to underwater video.

Abstract

In underwater scenes, wavelength-dependent light absorption and scattering degrade the visibility of images and videos. The degraded underwater images and videos affect the accuracy of pattern recognition, visual understanding, and key feature extraction in underwater scenes. In this paper, we propose an underwater image enhancement convolutional neural network (CNN) model based on underwater scene prior, called UWCNN. Instead of estimating the parameters of underwater imaging model, the proposed UWCNN model directly reconstructs the clear latent underwater image, which benefits from the underwater scene prior which can be used to synthesize underwater image training data. Besides, based on the light-weight network structure and effective training data, our UWCNN model can be easily extended to underwater videos for frame-by-frame enhancement. Specifically, combining an underwater imaging physical model with optical properties of underwater scenes, we first synthesize underwater image degradation datasets which cover a diverse set of water types and degradation levels. Then, a light-weight CNN model is designed for enhancing each underwater scene type, which is trained by the corresponding training data. At last, this UWCNN model is directly extended to underwater video enhancement. Experiments on real-world and synthetic underwater images and videos demonstrate that our method generalizes well to different underwater scenes.

References

[1]
D. Akkaynak, T. Treibitz, A revised underwater image formation, Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit. (CVPR), IEEE, 2018, pp. 6723–6732.
[2]
S. Chikkerur, N. Cartwright, V. Govindaraju, Fingerprint enhancement using stft analysis, Pattern Recognit. 40 (1) (2007) 198–211.
[3]
G. Lore, A. Akintayo, S. Sarkar, Llnet: a deep autoencoder approach to natural low-light image enhancement, Pattern Recognit. 61 (2017) 650–662.
[4]
W. Wang, Y. Xu, J. Shen, S. Zhu, Attentive fashion grammar network for fashion landmark detection and clothing category classification, Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit. (CVPR), IEEE, 2018, pp. 4271–4280.
[5]
F. Yang, K. Yan, S. Lu, H. Jia, X. Xie, W. Gao, Attentive driven person re-identification, Pattern Recognit. 86 (2019) 143–155.
[6]
W. Wang, J. Shen, F. Porikli, R. Yang, Semi-supervised video object segmentation with super-trajectories, IEEE Trans. Pattern Anal. Mach. Intell. 41 (4) (2019) 985–998.
[7]
C. Guo, C. Li, J. Guo, R. Cong, H. Fu, P. Han, Hierarchical features driven residual learning for depth map super-resolution, IEEE Trans. Image Process. 28 (5) (2019) 2545–2557.
[8]
Z. Wu, C. Shen, A.V.D. Hengel, Wider or deeper: revisiting the resnet model for visual recognition, Pattern Recognit. 90 (2019) 119–133.
[9]
D. Berman, T. Treibitz, S. Avidan, Diving into haze-lines: color restoration of underwater images, Proc. Brit. Mach. Vis. Conf. (BMVC), Springer, 2017, pp. 1–11.
[10]
W. Wang, J. Shen, Deep visual attention prediction, IEEE Trans. Image Process. 27 (5) (2018) 2368–2378.
[11]
H. Song, W. Wang, S. Zhao, J.S.K. Lam, Pyramid dilated deeper convlstm for video salient object detection, Proc. Eur. Conf. Comput. Vis. (ECCV), Springer, 2018, pp. 715–731.
[12]
J. Gu, Z. Wang, J. Kuen, L. Ma, A. Shahroudy, B. Shuai, T. Liu, X. Wang, J. Cai, T. Chen, Recent advances in convolutional neural networks, Pattern Recognit. 77 (2018) 354–377.
[13]
C. Li, J. Guo, Underwater image enhancement by dehazing and color correction, J. Electron. Imag. 24 (3) (2015).
[14]
C. Ancuti, C.O. Ancuti, Enhancing underwater images and videos by fusion, Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit. (CVPR), IEEE, 2012, pp. 81–88.
[15]
C. Li, J. Guo, C. Guo, R. Cong, J. Gong, A hybrid method for underwater image correction, Pattern Recognit. Lett. 94 (2017) 62–67.
[16]
C. Li, J. Guo, C. Guo, Emerging from water: underwater image color correction based on weakly supervised color transfer, IEEE Signal Process. Lett. 25 (3) (2018) 323–327.
[17]
Y. Guo, H. Li, P. Zhuang, Underwater image enhancement using a multiscale dense generative adversarial network, IEEE J. Ocean. Engineer. (2019) 1–9.
[18]
C. Ancuti, C.O. Ancuti, C. Vleeschouwer, Color balance and fusion for underwater image enhancement, IEEE Trans. Image Process. 27 (1) (2018) 379–393.
[19]
C. Li, C. Guo, W. Ren, R. Cong, J. Hou, S. Kwong, An underwater image enhancement benchmark dataset and beyond (2019).
[20]
J. Chiang, Y. Chen, Underwater image enhancement by wavelength compensation and dehazing, IEEE Trans. Image Process. 21 (4) (2012) 1756–1769.
[21]
A. Galdran, D. Pardo, A. Picn, A. Alvarez-Gila, Automatic red-channel underwater image restoration, Vis. Commun. Image Rep. 26 (2015) 132–145.
[22]
P. Drews, E. Nascimento, S. Botelho, M. Campos, Underwater depth estimation and image restoration based on single images, IEEE Comput. Graph. Appl. 36 (2) (2016) 24–35.
[23]
K. He, J. Sun, X. Tang, Single image haze removal using dark channel prior, IEEE Trans. Pattern Anal. Mach. Intell. 33 (12) (2011) 2341–2343.
[24]
C. Li, J. Guo, S. Chen, Y. Tang, Y. Pang, J. Wang, Underwater image restoration based on minimum information loss principle and optical properties of underwater imaging, Proc. IEEE Int. Conf. Image Process. (ICIP), IEEE, 2016, pp. 1993–1997.
[25]
C. Li, J. Guo, R. Cong, Y. Pang, B. Wang, Underwater image enhancement by dehazing with minimum information loss and histogram distribution prior, IEEE Trans. Image Process. 25 (12) (2016) 5664–5677.
[26]
Y. Peng, P. Cosman, Underwater image restoration based on image blurriness and light absorption, IEEE Trans. Image Process. 26 (4) (2017) 1579–1594.
[27]
J. Li, K. Skinner, R. Eustice, M. Roberson, Watergan: unsupervised generative network to enable real-time color correction of monocular underwater images, IEEE Robot. Autom. Lett. 3 (1) (2017) 387–394.
[28]
M. Sheinin, Y. Schechner, The next best underwater view, Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit. (CVPR), IEEE, 2017, pp. 1436–1443.
[29]
C. Li, C. Guo, J. Guo, P. Han, H. Fu, R. Cong, Pdr-net: perception-inspired single image dehazing network with refinement, IEEE Trans. Multimed. (2019) 1.
[30]
Z. Shen, W.L.T. Xu, J. Kautz, M. Yang, Deep semantic face deblurring, Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit. (CVPR), IEEE, 2018, pp. 8260–8269.
[31]
H. Huang, H. He, X. Fan, J. Zhang, Super-resolution of human face image using canonical correlation analysis, Pattern Recognit. 43 (7) (2010) 2532–2543.
[32]
T. Lopes, E. de Aguiar, F.D. Souza, T. Oliveira-Santos, Facial expression recognition with convolutional neural networks: croping with few data and the training sample order, Pattern Recognit. 61 (1) (2017) 610–628.
[33]
G. Huang, Z. Liu, L. van der Matten, K. Weinberger, Densely connected convolutional networks, Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit. (CVPR), IEEE, 2017, pp. 4700–4708.
[34]
W. Wang, J.S.F. Guo, M. Cheng, A. Borji, Revisiting video saliency: a large-scale benchmark and a new model, Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit. (CVPR), IEEE, 2018, pp. 4894–4903.
[35]
C. Zhou, J. Yuan, Multi-label learning of part detectors for occluded pedestrian detection, Pattern Recognit. 86 (2019) 99–111.
[36]
C. Li, R. Cong, J. Hou, S. Zhang, Y. Qian, S. Kwong, Nested network with two-stream pyramid for salient object detection in optical remote sensing images, IEEE Trans. Geosci. Remote Sens. (2019) 1.
[37]
N. Silberman, D. Hoiem, P. Kohli, R. Fergus, Indoor segmentation and support inference from rgbd images, Proc. Eur. Conf. Comput. Vis. (ECCV), Springer, 2012, pp. 746–760.
[38]
Z. Wang, A. Bovik, H. Sherikh, E. Simoncelli, Image quality assessment: from error visibility to structural similarity, IEEE Trans. Image Process. 13 (4) (2004) 600–612.
[39]
M. Yang, A. Sowmya, An underwater color image quality evaluation metric, IEEE Trans. Image Process. 24 (12) (2015) 6062–6071.
[40]
K. Panetta, C. Gao, S. Agaian, Human-visual-system-inspired underwater image quality measures, IEEE J. Ocean. Eng. 41 (3) (2016) 541–551.

Cited By

View all
  • (2024)Multi-Scale and Multi-Layer Lattice Transformer for Underwater Image EnhancementACM Transactions on Multimedia Computing, Communications, and Applications10.1145/368880220:11(1-24)Online publication date: 14-Aug-2024
  • (2024)Underwater Fuzzy Image Enhancement Method Based on CycleGANProceedings of the 2024 3rd International Symposium on Control Engineering and Robotics10.1145/3679409.3679437(131-137)Online publication date: 24-May-2024
  • (2024)Enhancing Underwater Images via Asymmetric Multi-Scale Invertible NetworksProceedings of the 32nd ACM International Conference on Multimedia10.1145/3664647.3681098(6182-6191)Online publication date: 28-Oct-2024
  • Show More Cited By

Index Terms

  1. Underwater scene prior inspired deep underwater image and video enhancement
    Index terms have been assigned to the content through auto-classification.

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image Pattern Recognition
    Pattern Recognition  Volume 98, Issue C
    Feb 2020
    385 pages

    Publisher

    Elsevier Science Inc.

    United States

    Publication History

    Published: 01 February 2020

    Author Tags

    1. Underwater image and video enhancement and restoration
    2. Underwater image synthesis
    3. Pattern recognition
    4. Deep learning

    Qualifiers

    • Research-article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 23 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Multi-Scale and Multi-Layer Lattice Transformer for Underwater Image EnhancementACM Transactions on Multimedia Computing, Communications, and Applications10.1145/368880220:11(1-24)Online publication date: 14-Aug-2024
    • (2024)Underwater Fuzzy Image Enhancement Method Based on CycleGANProceedings of the 2024 3rd International Symposium on Control Engineering and Robotics10.1145/3679409.3679437(131-137)Online publication date: 24-May-2024
    • (2024)Enhancing Underwater Images via Asymmetric Multi-Scale Invertible NetworksProceedings of the 32nd ACM International Conference on Multimedia10.1145/3664647.3681098(6182-6191)Online publication date: 28-Oct-2024
    • (2024)UIERL: Internal-External Representation Learning Network for Underwater Image EnhancementIEEE Transactions on Multimedia10.1109/TMM.2024.338776026(9252-9267)Online publication date: 12-Apr-2024
    • (2024)Underwater Color Correction Network With Knowledge TransferIEEE Transactions on Multimedia10.1109/TMM.2024.337459826(8088-8103)Online publication date: 12-Mar-2024
    • (2024)Underwater Image Quality Assessment: Benchmark Database and Objective MethodIEEE Transactions on Multimedia10.1109/TMM.2024.337121826(7734-7747)Online publication date: 28-Feb-2024
    • (2024)Multi-Scale Fusion and Decomposition Network for Single Image DerainingIEEE Transactions on Image Processing10.1109/TIP.2023.333455633(191-204)Online publication date: 1-Jan-2024
    • (2024)Semi-Supervised Feature Distillation and Unsupervised Domain Adversarial Distillation for Underwater Image EnhancementIEEE Transactions on Circuits and Systems for Video Technology10.1109/TCSVT.2024.337825234:8(7671-7682)Online publication date: 1-Aug-2024
    • (2024)Underwater Image Quality Improvement via Color, Detail, and Contrast RestorationIEEE Transactions on Circuits and Systems for Video Technology10.1109/TCSVT.2023.329752434:3(1726-1742)Online publication date: 1-Mar-2024
    • (2024)Non-Uniform Illumination Underwater Image Restoration via Illumination Channel Sparsity PriorIEEE Transactions on Circuits and Systems for Video Technology10.1109/TCSVT.2023.329036334:2(799-814)Online publication date: 1-Feb-2024
    • Show More Cited By

    View Options

    View options

    Login options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media