Nothing Special   »   [go: up one dir, main page]

CN112767279A - Underwater image enhancement method for generating countermeasure network based on discrete wavelet integration - Google Patents

Underwater image enhancement method for generating countermeasure network based on discrete wavelet integration Download PDF

Info

Publication number
CN112767279A
CN112767279A CN202110134075.9A CN202110134075A CN112767279A CN 112767279 A CN112767279 A CN 112767279A CN 202110134075 A CN202110134075 A CN 202110134075A CN 112767279 A CN112767279 A CN 112767279A
Authority
CN
China
Prior art keywords
image
network
convolution
layer
underwater
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110134075.9A
Other languages
Chinese (zh)
Other versions
CN112767279B (en
Inventor
牛玉贞
张宇杰
林煌伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fuzhou University
Original Assignee
Fuzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fuzhou University filed Critical Fuzhou University
Priority to CN202110134075.9A priority Critical patent/CN112767279B/en
Publication of CN112767279A publication Critical patent/CN112767279A/en
Application granted granted Critical
Publication of CN112767279B publication Critical patent/CN112767279B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/73Deblurring; Sharpening
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/10Image enhancement or restoration using non-spatial domain filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/90Dynamic range modification of images or parts thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20048Transform domain processing
    • G06T2207/20064Wavelet transform [DWT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to an underwater image enhancement method for generating a countermeasure network based on discrete wavelet integration. The method comprises the following steps: carrying out pairing processing, data enhancement and normalization processing on data to be trained, then zooming to a fixed size, and downsampling a label image, namely a real underwater image enhancement result to generate images with different sizes; designing a multi-level image enhancement generation network, and training an image enhancement model capable of enhancing the underwater image by using the designed network; designing a multi-level image identification neural network, and training an image identification model capable of predicting the probability that an input image is a real image by using the designed network; designing a generating network and identifying a neural network target loss function; alternately training the generating network and the discriminating neural network to converge to Nash balance by using the paired images; and inputting the underwater image into the trained image enhancement generation model, and outputting the enhanced image. The invention solves the problems of color distortion and blurring of the underwater image and can obviously improve the quality of the underwater image.

Description

Underwater image enhancement method for generating countermeasure network based on discrete wavelet integration
Technical Field
The invention relates to the field of image processing and computer vision, in particular to an underwater image enhancement method for generating a countermeasure network based on discrete wavelet integration.
Background
The marine environment containing a large amount of resources is one of the core components of human sustainable development, however, the marine complex imaging environment causes the underwater image acquired by the optical vision system to be generally accompanied with various types of degradation, which brings great influence on underwater operation depending on vision, such as submarine resource exploration, marine archaeology, underwater target detection and the like. The main reason for the degradation of underwater images is the energy attenuation and scattering phenomena of light as it travels underwater. When light is transmitted underwater, energy attenuation is accompanied, the light attenuation degrees of different spectral regions are different, red light attenuation is the most serious, green light is the second, and blue light is the least. The underwater image generally appears to be bluish-green accompanied by a color deviation. The scattering of light in water is classified into forward scattering and backward scattering. Forward scattering refers to a scattering phenomenon that small angle deviation is generated when light reflected by an object in water is transmitted to a camera, and image details are blurred. Backscattering refers to the fact that when light rays irradiate on objects in water, the light rays encounter impurities in the water and are scattered and directly received by a camera, and the contrast of an image is reduced.
At present, some imaging devices specially designed for underwater imaging environments can overcome part of underwater imaging problems and acquire clear images. However, in practical application, the requirements for the quality of underwater images are far more than that, the requirements are still difficult to meet only by means of acquisition equipment, and advanced underwater imaging equipment has extremely high cost, so that the underwater imaging equipment is difficult to be widely applied to various underwater tasks. Compared with expensive underwater imaging equipment, the image processing technology has the advantages of low cost, easiness in operation and the like.
The existing underwater image enhancement methods are mainly divided into two types: firstly, mathematical modeling is carried out on the degradation process of an underwater image, the degradation process is inverted to obtain a clear underwater image by estimating model parameters, the method needs to establish a model according to information such as optical parameters of a water body, camera parameters and the distance between a camera and a target object, the information needs to be measured by manual measurement or other methods to obtain numerical values, the obtained information is generally inaccurate, and therefore the enhanced image is prone to the problems of detail blurring, color deviation and the like. Meanwhile, although the methods can achieve good effects in certain fixed scenes, underwater scenes are complex and changeable, underwater image degradation types are various, the methods are difficult to match all the degradation types, generalization capability is poor, and when the scenes are changed greatly, the models need to be built again.
Secondly, a method based on deep learning is used, the method is a data-driven method, has strong learning characteristic representation capability, is lower in complexity than a non-deep learning method, and is widely applied to the field of image processing, so that a plurality of underwater image enhancement algorithms based on deep learning are proposed in recent years. However, although the existing underwater image enhancement method based on deep learning can effectively improve the quality of underwater images, the existing underwater image enhancement method based on deep learning has shortcomings in local and detail enhancement of underwater images due to the fact that the underwater environment is too complex, and the recovered images have problems of color deviation, detail blurring and the like.
In summary, the underwater image is complex in imaging environment and low in imaging image quality, and the existing method is easy to generate the phenomena of detail blurring and color deviation on the result of underwater image enhancement. The method comprises the steps of dividing an image into high-frequency information and low-frequency information and processing the high-frequency information and the low-frequency information separately, effectively reducing information loss by utilizing the advantages of discrete wavelets in the aspect of signal reconstruction, generating an underwater image enhancement result which is more in line with human subjective visual effect through countertraining of a generation model and an identification model, and remarkably improving the quality of the underwater image.
Disclosure of Invention
The invention aims to provide an underwater image enhancement method for generating a countermeasure network based on discrete wavelet integration, which is beneficial to improving the quality of underwater images.
In order to achieve the purpose, the technical scheme of the invention is as follows: an underwater image enhancement method based on discrete wavelet integration generation countermeasure network comprises the following steps:
step S1, data preprocessing is carried out on the data to be trained: firstly, matching an underwater image with a corresponding label image, then performing data enhancement and normalization processing, then scaling the normalized image to a fixed size, and downsampling the label image, namely a real underwater image enhancement result, to generate images with different sizes;
s2, designing a multi-level image enhancement generation network, and training an image enhancement model for enhancing the underwater image by using the designed network;
step S3, designing a multi-level image identification neural network, and using the designed network to train an image identification model for predicting the probability that an input image is a real image;
step S4, designing a generating network and identifying a neural network target loss function;
step S5, alternately training the generation network and the discrimination neural network by using the paired images to converge to Nash balance;
and step S6, inputting the underwater image into the trained image enhancement generation model, and outputting the enhanced image.
In an embodiment of the present invention, the step S1 is implemented as follows:
step S11, matching the underwater image with the corresponding label image;
step S12, carrying out uniform random turning operation on all paired images to be trained, and enhancing data;
step S13, all the enhanced images to be trained are normalized, an image I (I, j) is given, and a normalization value is calculated
Figure BDA0002926229630000021
The formula of (1) is as follows:
Figure BDA0002926229630000022
wherein (i, j) represents the position of the pixel;
step S14, scaling all normalized images to a fixed size H multiplied by W;
step S15, down-sampling the label image to generate different size images, and giving image G(H,W)A label image G of H × W size is represented, and an image G 'is generated by down-sampling'kK is 2,3, wherein
Figure BDA0002926229630000023
Downsampling uses the nearest neighbor interpolation algorithm, and G'1G; for any label image G, label image sets { G 'of different sizes are obtained'1,G′2,G′3}。
In an embodiment of the present invention, the step S2 is implemented as follows:
step S21, designing a multi-level neural network structure for extracting effective features of images, inputting the feature extraction network into an underwater image, wherein the network comprises a plurality of convolution blocks and wavelet pooling layers; the convolution blocks consist of convolution layers, a normalization layer and an activation layer, the feature extraction network has three layers, each layer comprises a plurality of convolution blocks and a wavelet pooling layer, and the number of the convolution blocks in each layer can be different; the convolution layer uses convolution with convolution kernel of 3x3 and step length of 1, the normalization layer uses batch normalization, and the activation layer uses ReLu function; the wavelet pooling layer uses discrete Haar wavelets to decompose the features, using four decomposition kernels LLT,LHT,HLT,HHTWherein the low frequency filter and the high frequency filter are respectively
Figure BDA0002926229630000031
The wavelet pooling layer decomposes the input features into low and high frequency components, llk,lhk,hlk,hhkK is 1,2,3, wherein llkFor low frequency components, lhk,hlk,hhkAre all high frequency components; the low-frequency components are transmitted to the next level of the network, and each high-frequency component is directly transmitted into the image reconstruction network after passing through a feature thinning processing module; the feature refinement processing module consists of a plurality of residual attention modules, each residual attention module comprises two layers of convolution kernels of 3x3 and step lengthConvolution of 1 and a channel attention module; the output of the final feature extraction network is ll3Features and three levels of high frequency components lhk,hlk,hhk,k=1,2,3;
Step S22, designing a multi-level neural network structure for image reconstruction, wherein the input of the image reconstruction network is ll extracted in the step S213Features and three levels of high frequency components lhk,hlk,hhkH is 1,2, 3; the image reconstruction network comprises a plurality of volume blocks, reconstruction blocks and a wavelet anti-pooling layer; the convolution block consists of a convolution layer, a normalization layer and an activation layer, and the reconstruction block consists of a convolution layer and an activation layer; the image reconstruction network has three levels which respectively correspond to the three levels of the feature extraction network, each level comprises a rolling block, a wavelet anti-pooling layer and a reconstruction block, and the number of the rolling blocks in each level can be different; convolution layers in the convolution blocks use convolution with convolution kernels of 3x3 and step length of 1, the normalization layer uses batch normalization, and the activation layer uses a ReLu function; convolution layers in the reconstruction blocks use convolution with convolution kernels of 3x3 and step length of 1, and activation layers use Tanh functions; the first level of input is ll3Features, low-frequency component ll is obtained after convolution block4Features as low frequency components of the wavelet anti-pooling layer; discrete Haar wavelet pair ll is used by wavelet anti-pooling layer4Characteristic and high frequency components lhk,hlk,hhkThe k is 3, and the output of the wavelet anti-pooling layer is used as the input of the reconstruction block on one hand and as the input characteristic of the second layer on the other hand; the second layer is rolled to obtain a low-frequency component ll5Characteristic, wavelet anti-pooling layer pair ll5Characteristic high frequency component lhk,hlk,hhkK is 2; similarly to the previous layer, the output of the wavelet anti-pooling layer is used as the input of the reconstruction block on one hand and as the input characteristic of the third layer on the other hand; the third layer is rolled to obtain a low-frequency component ll6Characteristic, wavelet anti-pooling layer pair ll6Characteristic high frequency component lhk,hlk,hhkCombining k with 1, and generating an H multiplied by W underwater image enhanced image through a rolling block and a reconstruction block; in addition, the method can be used for producing a composite materialThe features after the first two layers are combined generate an underwater image enhancement image with the current size through a reconstruction block; final output of three sizes of enhanced image I'kK is 1,2 and 3, and the sizes are H multiplied by W,1/2H multiplied by 1/2W,1/4H multiplied by 1/4W and G 'respectively'kAnd k is equal to 1,2 and 3.
In an embodiment of the present invention, the step S3 is specifically implemented as follows:
designing a multi-level neural network structure for image identification, wherein the image identification network inputs three-size enhanced images I 'generated by a generation network'kK is 1,2,3 or tag image G'kWhen k is 1,2,3, the network needs to be the same type of image for each input, namely, the same enhanced image or the same label image; the image identification network comprises a plurality of rolling blocks, a splicing layer and a pooling layer; the convolution block consists of a convolution layer and an activation layer, wherein the convolution layer uses convolution with a convolution kernel of 3x3 and a step length of 1, the activation layer uses a LeakyReLu function, and the pooling layer uses convolution with a convolution kernel of 4x4 and a step length of 2; image authentication network input l'1Or G'1Representing the underwater image and the label image respectively, firstly obtaining a characteristic vector d with the size of 1/2H multiplied by 1/2W through a pooling layer and two convolution blocks2And then picture I 'is processed using one convolution block'2Or G'2Extracting corresponding dimension characteristics f2The two characteristics are fused by using a characteristic splicing layer in a way of
Figure BDA0002926229630000041
Figure BDA0002926229630000042
Wherein
Figure BDA0002926229630000043
Representing a splicing operation; the fused features are subjected to a pooling layer and two convolution blocks to obtain a feature vector d with the size of 1/4H multiplied by 1/4W3And then picture I 'is processed using one convolution block'3Or G'3Extracting corresponding dimension characteristics f3Fusing the two characteristics by using a characteristic splicing layer in a fusing modeIs consistent with the previous layer; the fused feature vector passes through a pooling layer and two convolution blocks to obtain a feature vector with the size of 1/8H multiplied by 1/8W, and finally passes through a convolution layer with 1 channel to predict the probability of being a real image.
In an embodiment of the present invention, the step S4 is implemented as follows:
step S41, designing and generating a network target loss function, and generating a network total target loss function as follows:
Figure BDA0002926229630000044
wherein,
Figure BDA0002926229630000045
and
Figure BDA0002926229630000046
respectively L1 loss and generative countermeasure loss, λ1And λ2Is each loss balance coefficient, is a real number dot product operation; the specific calculation formula of each loss is as follows:
Figure BDA0002926229630000047
the generation network outputs three sizes of enhanced images, and the L1 loss only calculates the loss between the maximum size generated image and the reference label image; i'1Is to generate the largest size image, G 'in the network enhancement result'1Is the maximum size image in the reference label image, | | |1Is an absolute value operation;
Figure BDA0002926229630000048
wherein z represents an underwater image, Pz(z) probability distribution of underwater image, G generation network, D identification network, G (z) enhancement result of underwater image by generation network, D (G:)z)) represents the identification result of the output of the generation network by the identification network, a constant c is a numerical value used by the identification network for judging the difference between the generated image and the label image, wherein the value is c is 1, the generation network outputs enhanced images with three sizes in total, and all the enhanced images need to be sent to the identification network for calculating loss;
step S42, designing and identifying a neural network target loss function, wherein the identifying network target loss function is as follows:
Figure BDA0002926229630000051
where x denotes a reference label image, Pdata(x) Representing the probability distribution of the reference label image, z representing the underwater image, Pz(z) represents the probability distribution of the underwater image, G represents a generation network, D represents an identification network, D (x) represents the identification result of the identification neural network to the label image, G (z) represents the enhancement result of the generation network to the underwater image, D (G (z)) represents the identification result of the identification neural network to the output of the generation network, constants a and b represent the label image and the mark of the generation image respectively, the value is 1, b is 0, the generation network outputs the enhanced images of three sizes in total, all the enhanced images of three sizes corresponding to the reference label image need to be sent to the identification neural network to calculate loss, and all the enhanced images of three sizes corresponding to the reference label image need to be sent to the identification neural network to calculate loss.
In an embodiment of the present invention, the step S5 is implemented as follows:
step S51, randomly dividing the matched underwater images and label images into a plurality of batches, wherein each batch comprises N pairs of images;
step S52, inputting the underwater image into a generation network to obtain enhanced images I 'of three sizes'kK is 1,2,3, and the size is H × W,1/2H × 1/2W,1/4H × 1/4W in sequence, wherein H × W is the size specified in step S1;
step S53, inputting the generated enhanced image and the label image into an identifying neural network to obtain a judgment result;
step S54, calculating the gradient of each parameter in the generated network by using a back propagation method according to the total target loss function of the generated network, and updating the parameters of the generated network by using a random gradient descent method;
step S55, calculating the gradient of each parameter in the identified neural network by using a back propagation method according to the target loss function of the identified neural network, and updating the parameter of the identified neural network by using a random gradient descent method;
and S56, repeating the steps S51 to S54 of generating the network and identifying the neural network training by taking batches as units until the target loss function value of the generated network and the target loss function value of the identifying neural network converge to Nash balance, storing the network parameters, and finishing the training process of the generated network and the identifying neural network.
Compared with the prior art, the invention has the following beneficial effects:
the method is suitable for underwater image enhancement in various complex environments, can effectively restore the distorted color of the image, remove the image blur, and improve the contrast and brightness of the image. The enhanced image conforms to human subjective visual perception. The invention provides an underwater image enhancement network based on discrete wavelets, which can effectively reduce information loss in the image transmission process, retain image detail information, avoid detail blurring and be suitable for most complex scenes.
Drawings
FIG. 1 is a flow chart of an implementation of the method of the present invention.
Fig. 2 is a diagram of a network model architecture in an embodiment of the present invention.
Fig. 3 is a structural diagram of a feature refinement module in an embodiment of the present invention.
FIG. 4 is a diagram of a residual attention module in the feature refinement module according to an embodiment of the present invention.
Detailed Description
The technical scheme of the invention is specifically explained below with reference to the accompanying drawings.
The invention relates to an underwater image enhancement method for generating a countermeasure network based on discrete wavelet integration, which comprises the following steps:
step S1, data preprocessing is carried out on the data to be trained: firstly, matching an underwater image with a corresponding label image, then performing data enhancement and normalization processing, then scaling the normalized image to a fixed size, and downsampling the label image, namely a real underwater image enhancement result, to generate images with different sizes;
s2, designing a multi-level image enhancement generation network, and training an image enhancement model for enhancing the underwater image by using the designed network;
step S3, designing a multi-level image identification neural network, and using the designed network to train an image identification model for predicting the probability that an input image is a real image;
step S4, designing a generating network and identifying a neural network target loss function;
step S5, alternately training the generation network and the discrimination neural network by using the paired images to converge to Nash balance;
and step S6, inputting the underwater image into the trained image enhancement generation model, and outputting the enhanced image.
The following is a specific implementation of the present invention.
As shown in fig. 1, an underwater image enhancement method for generating a countermeasure network based on discrete wavelet integration includes the following steps:
s1, preprocessing data to be trained, firstly pairing the data, then enhancing and normalizing the data, then scaling the normalized image to a fixed size, and downsampling a label image, namely a real underwater image enhancement result to generate images with different sizes;
s2, designing a multi-level image enhancement generation network, and training an image enhancement model capable of enhancing the underwater image by using the designed network;
step S3, designing a multi-level image identification neural network, and using the designed network to train an image identification model capable of predicting the probability that an input image is a real image;
step S4, designing and generating a network and identifying a network target loss function;
step S5, alternately training the generation network and the identification network by using the paired images to converge to Nash balance;
and step S6, inputting the underwater image into the trained image enhancement generation model, and outputting the enhanced image.
Further, step S1 includes the steps of:
step S11: and matching the underwater image with the corresponding label image.
Step S12: and carrying out uniform random overturning operation on all paired images to be trained, and enhancing the data.
Step S13: all images to be trained are normalized, an image I (I, j) is given, and a normalization value is calculated
Figure BDA0002926229630000061
The formula of (1) is as follows:
Figure BDA0002926229630000071
where (i, j) represents the position of the pixel.
Step S14, scale all normalized images to a fixed size H × W.
Step S15, down-sampling the label image to generate different size images, and giving image G(H,W)A label image G of H × W size is represented, and an image G 'is generated by down-sampling'kK is 2,3, wherein
Figure BDA0002926229630000072
The downsampling uses a nearest neighbor interpolation algorithm. And G'1G. For any label image G, label image sets { G 'of different sizes are obtained'1,G′2,G′3}。
Further, step S2 includes the steps of:
s21, designing a multi-level neural network structure for extracting effective features of the image, wherein the feature extraction network is input as an underwater imageLike this, the network includes a plurality of volume blocks and a wavelet pooling layer. The convolution blocks are composed of convolution layers, a normalization layer and an activation layer, the feature extraction network has three layers, each layer comprises a plurality of convolution blocks and a wavelet pooling layer, and the number of the convolution blocks in each layer can be different. The convolutional layer uses convolution with a convolution kernel of 3x3 and a step size of 1, the normalization layer uses batch normalization, and the activation layer uses the ReLu function. The wavelet pooling layer uses discrete Haar wavelets to decompose the features, using four decomposition kernels LLT,LHT,HLT,HHTWherein the high-frequency filters are respectively
Figure BDA0002926229630000073
The wavelet pooling layer decomposes the input features into low and high frequency components, llk,lhk,hlk,hhkK is 1,2,3, wherein llkFor low frequency components, lhk,hlk,hhkAre all high frequency components. The low-frequency components are transmitted to the next level of the network, and each high-frequency component is directly transmitted into the image reconstruction network after passing through a feature thinning processing module. The feature refinement processing module consists of a plurality of residual attention modules, each of which consists of two layers of convolution kernel of 3x3 with step size of 1 and one channel attention module. The output of the final feature extraction network is ll3Features and three levels of high frequency components lhk,hlk,hhk,k=1,2,3。
Step S22, designing a multi-level neural network structure for image reconstruction, wherein the input of the image reconstruction network is ll extracted in the step S213Features and three levels of high frequency components lhk,hlk,hhkAnd k is 1,2, 3. The image reconstruction network includes a plurality of volume blocks, reconstruction blocks, and a wavelet anti-pooling layer. The convolution block is composed of a convolution layer, a normalization layer and an activation layer, and the reconstruction block is composed of a convolution layer and an activation layer. The image reconstruction network has three levels which respectively correspond to the three levels of the feature extraction network, each level comprises a rolling block, a wavelet anti-pooling layer and a reconstruction block, and the number of the rolling blocks in each level can be different. Convolutional layer in convolutional blockConvolution with a convolution kernel of 3x3 and a step size of 1 was used, the normalization layer used batch normalization, and the activation layer used the ReLu function. The convolutional layers in the reconstruction block use convolution with a convolution kernel of 3x3 and step size 1, and the active layers use the Tanh function. The first level of input is ll3Features that after several convolution blocks, the low-frequency component ll is obtained4And the characteristic is used as the low-frequency component of the wavelet anti-pooling layer. Discrete Haar wavelet pair ll is used by wavelet anti-pooling layer4Characteristic and high frequency components lhk,hlk,hhkAnd k is 3, the inverse pooling kernel parameters are the same as those in step S21, and the output of the wavelet inverse pooling layer is used as the input of the reconstruction block on one hand and as the input feature of the second layer on the other hand. The second layer is rolled to obtain a low-frequency component ll5Characteristic, wavelet anti-pooling layer pair ll5Characteristic high frequency component lhk,hlk,hhkAnd k is 2. Similarly to the previous layer, the output of the wavelet anti-pooling layer is used as the input of the reconstruction block on one hand and as the input feature of the third layer on the other hand. The third layer is rolled to obtain a low-frequency component ll6Characteristic, wavelet anti-pooling layer pair ll6Characteristic high frequency component lhk,hlk,hhkAnd k is combined to be 1, and then the H multiplied by W underwater image enhanced image is generated through several volume blocks and a reconstruction block. In addition, the features of the first two layers after combination are subjected to reconstruction to generate an underwater image enhanced image with the current size. Final output of three sizes of enhanced image I'kK is 1,2 and 3, and the sizes are H multiplied by W,1/2H multiplied by 1/2W,1/4H multiplied by 1/4W and G 'respectively'kAnd k is equal to 1,2, and 3, wherein H × W is the size specified in step S1.
Further, step S3 is implemented as follows:
designing a multi-level neural network structure for image identification, wherein the image identification network inputs three-size enhanced images I 'generated by a generation network'kK is 1,2,3 or tag image G'kAnd k is 1,2 and 3, the network needs to be the same type of image, namely the enhanced image or the label image, every time the network inputs the image, and the image size is the size specified in the step S1 and the step S2. Image authentication networkComprising a plurality of volume blocks, a splice layer and a pooling layer. The convolutional block consists of convolutional layers and active layers, wherein the convolutional layers use convolution with a convolution kernel of 3x3 and a step length of 1, the active layers use LeakyReLu functions, and the pooling layers use convolution with a convolution kernel of 4x4 and a step length of 2. Network inputs I'1Or G'1Representing the underwater image and the label image respectively, firstly obtaining a characteristic vector d with the size of 1/2H multiplied by 1/2W through a pooling layer and two convolution blocks2And then picture I 'is processed using one convolution block'2Or G'2Extracting corresponding dimension characteristics f2The two characteristics are fused by using a characteristic splicing layer in a way of
Figure BDA0002926229630000081
Wherein
Figure BDA0002926229630000082
Indicating a splicing operation. The fused features are subjected to a pooling layer and two convolution blocks to obtain a feature vector d with the size of 1/4H multiplied by 1/4W3And then picture I 'is processed using one convolution block'3Or G'3Extracting corresponding dimension characteristics f3And fusing the two characteristics by using the characteristic splicing layer, wherein the fusing mode is consistent with that of the previous layer. The fused feature vector passes through a pooling layer and two convolution blocks to obtain a feature vector with the size of 1/8H multiplied by 1/8W, and finally passes through a convolution layer with 1 channel to predict the probability of being a real image.
Further, step S4 includes the steps of:
step S41, designing and generating a network target loss function, and generating a network total target loss function as follows:
Figure BDA0002926229630000083
wherein,
Figure BDA0002926229630000084
and
Figure BDA0002926229630000085
respectively L1 loss and generative countermeasure loss, λ1And λ2Is each loss balance coefficient, is a real number dot product operation; the specific calculation formula of each loss is as follows:
Figure BDA0002926229630000086
the generation network outputs three sizes of enhanced images, and the L1 penalty only calculates the penalty between the largest size generated image and the reference label image. I'1Is to generate the largest size image, G 'in the network enhancement result'1Is the maximum size image in the reference label image, | | |1Is an absolute value operation;
Figure BDA0002926229630000091
wherein z represents an underwater image, Pz(z) represents the probability distribution of the underwater image, G represents a generation network, D represents an identification network, G (z) represents the enhancement result of the generation network on the underwater image, D (G (z)) represents the output identification result of the identification network on the generation network, a constant c is a numerical value used by the identification network for judging the difference between the generation image and the label image, the value is c-1, the generation network outputs the enhancement images with three sizes in total, and all the enhancement images need to be sent to the identification network for calculating loss.
Step S42, designing and identifying a network target loss function, wherein the network target loss function is as follows:
Figure BDA0002926229630000092
where x denotes a reference label picture, Pdata(x) Representing the probability distribution of the reference label image, z representing the underwater image, Pz(z)
The method comprises the steps of representing underwater image probability distribution, G representing a generation network, D representing an identification network, D (x) representing identification results of the identification network on label images, G (z) representing enhancement results of the generation network on the underwater images, D (G (z)) representing identification results of the identification network on output of the generation network, constants a and b respectively representing label images and marks of the generation images, wherein the values are a-1, b-0, the generation network outputs three-size enhancement images in total, all the three-size enhancement images need to be sent to the identification network for calculation loss, and all the three-size images corresponding to reference label images need to be sent to the identification network for calculation loss.
Further, step S5 includes the steps of:
step S51, randomly dividing the matched underwater images and label images into a plurality of batches, wherein each batch comprises N pairs of images;
step S52, inputting the underwater image into the generation network of step S2 to obtain three-size enhanced images I'kAnd k is 1,2 and 3, and the sizes are H multiplied by W,1/2H multiplied by 1/2W and 1/4H multiplied by 1/4W in sequence, wherein H multiplied by W is the size specified in the step S1.
Step S53, the generated enhanced image and label image are input to the discrimination network in step S3 to obtain the discrimination result.
Step S54, calculating the gradient of each parameter in the generated network by using a back propagation method according to the total target loss function of the generated network, and updating the parameters of the generated network by using a random gradient descent method;
step S55, calculating the gradient of each parameter in the identification network by using a back propagation method according to the target loss function of the identification network, and updating the parameters of the identification network by using a random gradient descent method;
and S56, repeating the steps from S51 to S54 for generating the network and identifying the network training by taking batches as units until the target loss function value of the generated network and the target loss function value of the identified network converge to Nash balance, storing the network parameters, and finishing the training process of the generated network and the identified network.
The above are preferred embodiments of the present invention, and all changes made according to the technical scheme of the present invention that produce functional effects do not exceed the scope of the technical scheme of the present invention belong to the protection scope of the present invention.

Claims (6)

1. An underwater image enhancement method for generating a countermeasure network based on discrete wavelet integration is characterized by comprising the following steps:
step S1, data preprocessing is carried out on the data to be trained: firstly, matching an underwater image with a corresponding label image, then performing data enhancement and normalization processing, then scaling the normalized image to a fixed size, and downsampling the label image, namely a real underwater image enhancement result, to generate images with different sizes;
s2, designing a multi-level image enhancement generation network, and training an image enhancement model for enhancing the underwater image by using the designed network;
step S3, designing a multi-level image identification neural network, and using the designed network to train an image identification model for predicting the probability that an input image is a real image;
step S4, designing a generating network and identifying a neural network target loss function;
step S5, alternately training the generation network and the discrimination neural network by using the paired images to converge to Nash balance;
and step S6, inputting the underwater image into the trained image enhancement generation model, and outputting the enhanced image.
2. The underwater image enhancement method for generating the countermeasure network based on the discrete wavelet integration according to claim 1, wherein the step S1 is implemented by the following steps:
step S11, matching the underwater image with the corresponding label image;
step S12, carrying out uniform random turning operation on all paired images to be trained, and enhancing data;
step S13, all the enhanced images to be trained are normalized, an image I (I, j) is given, and a normalization value is calculated
Figure FDA0002926229620000011
Is disclosedThe formula is as follows:
Figure FDA0002926229620000012
wherein (i, j) represents the position of the pixel;
step S14, scaling all normalized images to a fixed size H multiplied by W;
step S15, down-sampling the label image to generate different size images, and giving image G(H,W)A label image G of H × W size is represented, and an image G 'is generated by down-sampling'kK is 2,3, wherein
Figure FDA0002926229620000013
Downsampling uses the nearest neighbor interpolation algorithm, and G'1G; for any label image G, label image sets { G 'of different sizes are obtained'1,G′2,G′3}。
3. The underwater image enhancement method for generating the countermeasure network based on the discrete wavelet integration according to claim 2, wherein the step S2 is implemented by the following steps:
step S21, designing a multi-level neural network structure for extracting effective features of images, inputting the feature extraction network into an underwater image, wherein the network comprises a plurality of convolution blocks and wavelet pooling layers; the convolution blocks consist of convolution layers, a normalization layer and an activation layer, the feature extraction network has three layers, each layer comprises a plurality of convolution blocks and a wavelet pooling layer, and the number of the convolution blocks in each layer can be different; the convolution layer uses convolution with convolution kernel of 3x3 and step length of 1, the normalization layer uses batch normalization, and the activation layer uses ReLu function; the wavelet pooling layer uses discrete Haar wavelets to decompose the features, using four decomposition kernels LLT,LHT,HLT,HHTWherein the low frequency filter and the high frequency filter are respectively
Figure FDA0002926229620000021
The wavelet pooling layer decomposes the input features into low and high frequency components, llk,lhk,hlk,hhkK is 1,2,3, wherein llkFor low frequency components, lhk,hlk,hhkAre all high frequency components; the low-frequency components are transmitted to the next level of the network, and each high-frequency component is directly transmitted into the image reconstruction network after passing through a feature thinning processing module; the characteristic thinning processing module consists of a plurality of residual error attention modules, and each residual error attention module consists of two layers of convolution kernels with the convolution rate of 3x3 and the step length of 1 and a channel attention module; the output of the final feature extraction network is ll3Features and three levels of high frequency components lhk,hlk,hhk,k=1,2,3;
Step S22, designing a multi-level neural network structure for image reconstruction, wherein the input of the image reconstruction network is ll extracted in the step S213Features and three levels of high frequency components lhk,hlk,hhkK is 1,2, 3; the image reconstruction network comprises a plurality of volume blocks, reconstruction blocks and a wavelet anti-pooling layer; the convolution block consists of a convolution layer, a normalization layer and an activation layer, and the reconstruction block consists of a convolution layer and an activation layer; the image reconstruction network has three levels which respectively correspond to the three levels of the feature extraction network, each level comprises a rolling block, a wavelet anti-pooling layer and a reconstruction block, and the number of the rolling blocks in each level can be different; convolution layers in the convolution blocks use convolution with convolution kernels of 3x3 and step length of 1, the normalization layer uses batch normalization, and the activation layer uses a ReLu function; convolution layers in the reconstruction blocks use convolution with convolution kernels of 3x3 and step length of 1, and activation layers use Tanh functions; the first level of input is ll3Features, low-frequency component ll is obtained after convolution block4Features as low frequency components of the wavelet anti-pooling layer; discrete Haar wavelet pair ll is used by wavelet anti-pooling layer4Characteristic and high frequency components lhk,hlk,hhkThe k is 3, and the output of the wavelet anti-pooling layer is used as the input of the reconstruction block on one hand and as the input characteristic of the second layer on the other hand; the second layer is rolled to obtain a low-frequency component ll5Characteristic, wavelet anti-pooling layer pair ll5Characteristic high frequency component lhk,hlk,hhkK is 2; similarly to the previous layer, the output of the wavelet anti-pooling layer is used as the input of the reconstruction block on one hand and as the input characteristic of the third layer on the other hand; the third layer is rolled to obtain a low-frequency component ll6Characteristic, wavelet anti-pooling layer pair ll6Characteristic high frequency component lhk,hlk,hhkCombining k with 1, and generating an H multiplied by W underwater image enhanced image through a rolling block and a reconstruction block; in addition, the features of the first two layers after combination generate an underwater image enhanced image with the current size through a reconstruction block; final output of three sizes of enhanced image I'kK is 1,2 and 3, and the sizes are H multiplied by W,1/2H multiplied by 1/2W,1/4H multiplied by 1/4W and G 'respectively'kAnd k is equal to 1,2 and 3.
4. The underwater image enhancement method for generating a countermeasure network based on discrete wavelet integration according to claim 3, wherein the step S3 is implemented as follows:
designing a multi-level neural network structure for image identification, wherein the image identification network inputs three-size enhanced images I 'generated by a generation network'kK is 1,2,3 or tag image G'kWhen k is 1,2,3, the network needs to be the same type of image for each input, namely, the same enhanced image or the same label image; the image identification network comprises a plurality of rolling blocks, a splicing layer and a pooling layer; the convolution block consists of a convolution layer and an activation layer, wherein the convolution layer uses convolution with a convolution kernel of 3x3 and a step length of 1, the activation layer uses a LeakyReLu function, and the pooling layer uses convolution with a convolution kernel of 4x4 and a step length of 2; image authentication network input l'1Or G'1Representing the underwater image and the label image respectively, firstly obtaining a characteristic vector d with the size of 1/2H multiplied by 1/2W through a pooling layer and two convolution blocks2And then picture I 'is processed using one convolution block'2Or G'2Extracting corresponding dimension characteristics f2The two characteristics are fused by using a characteristic splicing layer in a way of
Figure FDA0002926229620000031
Figure FDA0002926229620000032
Wherein
Figure FDA0002926229620000033
Representing a splicing operation; the fused features are subjected to a pooling layer and two convolution blocks to obtain a feature vector d with the size of 1/4H multiplied by 1/4W3And then picture I 'is processed using one convolution block'3Or G'3Extracting corresponding dimension characteristics f3Fusing the two characteristics by using a characteristic splicing layer, wherein the fusing mode is consistent with that of the previous layer; the fused feature vector passes through a pooling layer and two convolution blocks to obtain a feature vector with the size of 1/8H multiplied by 1/8W, and finally passes through a convolution layer with 1 channel to predict the probability of being a real image.
5. The underwater image enhancement method for generating the countermeasure network based on discrete wavelet integration according to claim 4, wherein the step S4 is implemented by the following steps:
step S41, designing and generating a network target loss function, and generating a network total target loss function as follows:
Figure FDA0002926229620000034
wherein,
Figure FDA0002926229620000035
and
Figure FDA0002926229620000036
respectively L1 loss and generative countermeasure loss, λ1And λ2Is each loss balance coefficient, is a real number dot product operation; the specific calculation formula of each loss is as follows:
Figure FDA0002926229620000037
the generation network outputs three sizes of enhanced images, and the L1 loss only calculates the loss between the maximum size generated image and the reference label image; i'1Is to generate the largest size image, G 'in the network enhancement result'1Is the maximum size image in the reference label image, | | |1Is an absolute value operation;
Figure FDA0002926229620000038
wherein z represents an underwater image, Pz(z) representing underwater image probability distribution, G representing a generation network, D representing an identification network, G (z) representing an enhancement result of the generation network on the underwater image, D (G (z)) representing an identification result of the identification network on the output of the generation network, and a constant c being a numerical value used by the identification network for judging the difference between the generation image and the label image, wherein the value is c being 1, the generation network outputs the enhancement images of three sizes in total, and the enhancement images are required to be sent to the identification network for calculating loss;
step S42, designing and identifying a neural network target loss function, wherein the identifying network target loss function is as follows:
Figure FDA0002926229620000039
where x denotes a reference label image, Pdata(x) Representing the probability distribution of the reference label image, z representing the underwater image, Pz(z) represents underwater image probability distribution, G represents generation network, D represents identification network, D (x) represents identification result of identification neural network to label image, G (z) represents enhancement result of generation network to underwater image, D (G (z)) represents identification result of identification neural network to output of generation network, constants a and b represent label image and label of generation image respectively, wherein, the values are a-1, b-0, and generation networkThe three-size enhanced images are output by the network, and all the three-size enhanced images need to be sent to the discrimination neural network for calculating loss, and all the three-size enhanced images corresponding to the reference label image need to be sent to the discrimination neural network for calculating loss.
6. The underwater image enhancement method for generating the countermeasure network based on discrete wavelet integration according to claim 5, wherein the step S5 is implemented by the following steps:
step S51, randomly dividing the matched underwater images and label images into a plurality of batches, wherein each batch comprises N pairs of images;
step S52, inputting the underwater image into a generation network to obtain enhanced images I 'of three sizes'kK is 1,2,3, and the size is H × W,1/2H × 1/2W,1/4H × 1/4W in sequence, wherein H × W is the size specified in step S1;
step S53, inputting the generated enhanced image and the label image into an identifying neural network to obtain a judgment result;
step S54, calculating the gradient of each parameter in the generated network by using a back propagation method according to the total target loss function of the generated network, and updating the parameters of the generated network by using a random gradient descent method;
step S55, calculating the gradient of each parameter in the identified neural network by using a back propagation method according to the target loss function of the identified neural network, and updating the parameter of the identified neural network by using a random gradient descent method;
and S56, repeating the steps S51 to S54 of generating the network and identifying the neural network training by taking batches as units until the target loss function value of the generated network and the target loss function value of the identifying neural network converge to Nash balance, storing the network parameters, and finishing the training process of the generated network and the identifying neural network.
CN202110134075.9A 2021-02-01 2021-02-01 Underwater image enhancement method for generating countermeasure network based on discrete wavelet integration Active CN112767279B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110134075.9A CN112767279B (en) 2021-02-01 2021-02-01 Underwater image enhancement method for generating countermeasure network based on discrete wavelet integration

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110134075.9A CN112767279B (en) 2021-02-01 2021-02-01 Underwater image enhancement method for generating countermeasure network based on discrete wavelet integration

Publications (2)

Publication Number Publication Date
CN112767279A true CN112767279A (en) 2021-05-07
CN112767279B CN112767279B (en) 2022-06-14

Family

ID=75704416

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110134075.9A Active CN112767279B (en) 2021-02-01 2021-02-01 Underwater image enhancement method for generating countermeasure network based on discrete wavelet integration

Country Status (1)

Country Link
CN (1) CN112767279B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113222953A (en) * 2021-05-21 2021-08-06 西安电子科技大学 Natural image enhancement method based on depth gamma transformation
CN113240606A (en) * 2021-05-24 2021-08-10 博奥生物集团有限公司 Traditional Chinese medicine inspection image color correction method and system
CN113284061A (en) * 2021-05-17 2021-08-20 大连海事大学 Underwater image enhancement method based on gradient network
CN114092793A (en) * 2021-11-12 2022-02-25 杭州电子科技大学 End-to-end biological target detection method suitable for complex underwater environment
CN114612330A (en) * 2022-03-15 2022-06-10 中国人民解放军国防科技大学 Image enhancement method of multiple cameras
CN114881879A (en) * 2022-05-17 2022-08-09 燕山大学 Underwater image enhancement method based on brightness compensation residual error network
CN117152445A (en) * 2023-10-31 2023-12-01 暨南大学 Real-time image semantic segmentation method and system based on multi-connection coding wavelet pooling

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109685072A (en) * 2018-12-22 2019-04-26 北京工业大学 A kind of compound degraded image high quality method for reconstructing based on generation confrontation network
US20190295223A1 (en) * 2018-03-22 2019-09-26 Adobe Inc. Aesthetics-guided image enhancement
CN111242865A (en) * 2020-01-10 2020-06-05 南京航空航天大学 Fundus image enhancement method based on generation type countermeasure network
CN111275640A (en) * 2020-01-17 2020-06-12 天津大学 Image enhancement method for fusing two-dimensional discrete wavelet transform and generating countermeasure network
US20200286208A1 (en) * 2019-03-08 2020-09-10 International Business Machines Corporation Neural network based enhancement of intensity images
CN111798400A (en) * 2020-07-20 2020-10-20 福州大学 Non-reference low-illumination image enhancement method and system based on generation countermeasure network
CN112288658A (en) * 2020-11-23 2021-01-29 杭州师范大学 Underwater image enhancement method based on multi-residual joint learning

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190295223A1 (en) * 2018-03-22 2019-09-26 Adobe Inc. Aesthetics-guided image enhancement
CN109685072A (en) * 2018-12-22 2019-04-26 北京工业大学 A kind of compound degraded image high quality method for reconstructing based on generation confrontation network
US20200286208A1 (en) * 2019-03-08 2020-09-10 International Business Machines Corporation Neural network based enhancement of intensity images
CN111242865A (en) * 2020-01-10 2020-06-05 南京航空航天大学 Fundus image enhancement method based on generation type countermeasure network
CN111275640A (en) * 2020-01-17 2020-06-12 天津大学 Image enhancement method for fusing two-dimensional discrete wavelet transform and generating countermeasure network
CN111798400A (en) * 2020-07-20 2020-10-20 福州大学 Non-reference low-illumination image enhancement method and system based on generation countermeasure network
CN112288658A (en) * 2020-11-23 2021-01-29 杭州师范大学 Underwater image enhancement method based on multi-residual joint learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
QILI DENG ET AL.: "HardGAN: A Haze-Aware Representation Distillation GAN for Single Image Dehazing", 《COMPUTER VISION–ECCV 2020》 *
沈亮: "复杂天气条件下水天场景图像增强技术研究", 《中国优秀博硕士学位论文全文数据库(硕士)工程科技Ⅱ辑》 *

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113284061A (en) * 2021-05-17 2021-08-20 大连海事大学 Underwater image enhancement method based on gradient network
CN113284061B (en) * 2021-05-17 2024-04-05 大连海事大学 Underwater image enhancement method based on gradient network
CN113222953B (en) * 2021-05-21 2023-10-20 西安电子科技大学 Natural image enhancement method based on depth gamma transformation
CN113222953A (en) * 2021-05-21 2021-08-06 西安电子科技大学 Natural image enhancement method based on depth gamma transformation
CN113240606A (en) * 2021-05-24 2021-08-10 博奥生物集团有限公司 Traditional Chinese medicine inspection image color correction method and system
CN114092793A (en) * 2021-11-12 2022-02-25 杭州电子科技大学 End-to-end biological target detection method suitable for complex underwater environment
CN114092793B (en) * 2021-11-12 2024-05-17 杭州电子科技大学 End-to-end biological target detection method suitable for complex underwater environment
CN114612330B (en) * 2022-03-15 2024-02-02 中国人民解放军国防科技大学 Image enhancement method for multiple cameras
CN114612330A (en) * 2022-03-15 2022-06-10 中国人民解放军国防科技大学 Image enhancement method of multiple cameras
CN114881879A (en) * 2022-05-17 2022-08-09 燕山大学 Underwater image enhancement method based on brightness compensation residual error network
CN114881879B (en) * 2022-05-17 2024-09-24 燕山大学 Underwater image enhancement method based on brightness compensation residual error network
CN117152445A (en) * 2023-10-31 2023-12-01 暨南大学 Real-time image semantic segmentation method and system based on multi-connection coding wavelet pooling
CN117152445B (en) * 2023-10-31 2024-01-12 暨南大学 Real-time image semantic segmentation method and system based on multi-connection coding wavelet pooling

Also Published As

Publication number Publication date
CN112767279B (en) 2022-06-14

Similar Documents

Publication Publication Date Title
CN112767279B (en) Underwater image enhancement method for generating countermeasure network based on discrete wavelet integration
CN112233038B (en) True image denoising method based on multi-scale fusion and edge enhancement
CN112507997B (en) Face super-resolution system based on multi-scale convolution and receptive field feature fusion
CN112767251B (en) Image super-resolution method based on multi-scale detail feature fusion neural network
CN108932536B (en) Face posture reconstruction method based on deep neural network
CN106548159A (en) Reticulate pattern facial image recognition method and device based on full convolutional neural networks
CN111861894A (en) Image motion blur removing method based on generating type countermeasure network
CN113284061B (en) Underwater image enhancement method based on gradient network
Liu et al. Learning hadamard-product-propagation for image dehazing and beyond
CN111179196B (en) Multi-resolution depth network image highlight removing method based on divide-and-conquer
Lu et al. Underwater image enhancement method based on denoising diffusion probabilistic model
CN116703752A (en) Image defogging method and device of near infrared fused transducer structure
CN112614070A (en) DefogNet-based single image defogging method
CN112164010A (en) Multi-scale fusion convolution neural network image defogging method
CN114004766A (en) Underwater image enhancement method, system and equipment
Pan et al. ChebyLighter: Optimal Curve Estimation for Low-light Image Enhancement
Zhuang et al. A dense stereo matching method based on optimized direction-information images for the real underwater measurement environment
CN115170944A (en) NR underwater enhanced image quality evaluation method based on CNN
CN113628143A (en) Weighted fusion image defogging method and device based on multi-scale convolution
CN114820395B (en) Underwater image enhancement method based on multi-field information fusion
CN107301625A (en) Image defogging algorithm based on brightness UNE
Huang et al. Attention-based for multiscale fusion underwater image enhancement
CN115272131B (en) Image mole pattern removing system and method based on self-adaptive multispectral coding
CN116703750A (en) Image defogging method and system based on edge attention and multi-order differential loss
Natarajan A review on underwater image enhancement techniques

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant