Nothing Special   »   [go: up one dir, main page]

CN114387170A - Image repairing method for improving edge incoherence phenomenon of repairing area - Google Patents

Image repairing method for improving edge incoherence phenomenon of repairing area Download PDF

Info

Publication number
CN114387170A
CN114387170A CN202011116700.9A CN202011116700A CN114387170A CN 114387170 A CN114387170 A CN 114387170A CN 202011116700 A CN202011116700 A CN 202011116700A CN 114387170 A CN114387170 A CN 114387170A
Authority
CN
China
Prior art keywords
network
image
relative
repairing
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011116700.9A
Other languages
Chinese (zh)
Inventor
张文强
陈飞宇
张睿
张传法
邓苇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fudan University
Original Assignee
Fudan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fudan University filed Critical Fudan University
Priority to CN202011116700.9A priority Critical patent/CN114387170A/en
Publication of CN114387170A publication Critical patent/CN114387170A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/77Retouching; Inpainting; Scratch removal
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides an image repairing method for improving the edge incoherence phenomenon of a repairing area, belonging to the field of image repairing in the computer vision, and the method comprises the following steps: preprocessing a data set, constructing a network model, optimizing the network model, screening the network model and repairing a defective image to be repaired, wherein the construction of the network model is mainly divided into the construction of a generation network model by utilizing image multi-resolution information and a soft attention mechanism and the construction of a relative judgment network model for judgment. The method can automatically and effectively repair the defect-removed picture, and compared with the prior method, the repaired picture has more continuity of texture, more detailed picture content and more vivid overall visual effect.

Description

Image repairing method for improving edge incoherence phenomenon of repairing area
Technical Field
The invention belongs to the field of computer vision image restoration, and particularly relates to an image restoration method for improving the edge incoherence phenomenon of a restoration area.
Background
In recent years, with the rapid development of the internet and the popularization of intelligent mobile devices, people increasingly like to record and share their drip life in the form of images and short videos. On the other hand, the way of recording history and inheriting culture with images and video as carriers is also getting more and more attention from governments and common people. However, old images are often damaged or there are often unwanted people or objects in the images for display. Solving these problems not only requires a skilled technician, but also wastes a lot of time. This makes the image restoration work more and more regarded by the scientific community. Image Inpainting (Image Inpainting) refers to the reconstruction of lost or damaged portions of images and video so that the viewer cannot perceive that the Image was once missing or repaired. The method aims to automatically recover the lost information according to the existing information of the image, and can be used for recovering the lost information in the old photo, removing watermark characters, hiding wrong objects and the like.
At present, with the continuous and deep research of artificial intelligence, image restoration and artificial intelligence are gradually becoming mainstream research directions. Image restoration methods based on Convolutional Neural Networks (CNNs) and generative countermeasure networks (GANs) have achieved good performance in the field of image restoration. However, the conventional image restoration method combining artificial intelligence still has the problems of inconsistent image texture of the restored area and the image texture of the original area, missing details of the restored content, distorted object structure of the restored picture and the like.
Disclosure of Invention
The present invention has been made to solve the above problems, and an object of the present invention is to provide an image inpainting method for improving an edge discontinuity phenomenon in an inpainting region.
The invention provides an image repairing method for improving the edge incoherence phenomenon of a repairing area, which is used for repairing a defective image and has the characteristics that the method comprises the following steps: step 1, acquiring an original image data set, dividing the original image data set into a test set and a training set, and then preprocessing an original image in the test set to obtain a preprocessed test set with incomplete images; step 2, constructing a relative countermeasure network model, wherein the relative countermeasure network model comprises a generating network and a relative judging network which utilize image multi-resolution information and a soft attention mechanism, the generating network is used for repairing incomplete images to obtain complete images, and the judging network is used for judging the complete images to obtain probability feature vectors for describing the truth and the falseness of the complete images; step 3, inputting original images of a training set into a relative countermeasure network model, then respectively carrying out network optimization training on a generation network and a relative discrimination network, respectively calculating relative countermeasure loss, deeply supervised space loss and regional space loss, optimizing network parameters of the generation network and the relative discrimination network based on a back propagation algorithm and an Adam optimizer, then carrying out iterative updating on the network parameters according to a set learning rate, and alternately updating the generation network and the relative discrimination network, wherein when a complete image of the generated network restoration is close to a picture of the original image when the complete image is not incomplete, the network training is completed; step 4, inputting the preprocessed incomplete images of the test set into a trained relative confrontation network model, judging whether the relative confrontation generation network is over-fitted or not through evaluation indexes, returning to the step 3 if judging that the relative confrontation generation network is over-fitted, and reserving the generation network after training as an image restoration network if judging that the relative confrontation generation network is not over-fitted; and 5, inputting the incomplete image to be repaired into an image repairing network to obtain a repaired complete image.
The image inpainting method for improving the edge incoherence of the inpainting area provided by the invention also has the following characteristics: wherein, the pretreatment in the step 1 specifically comprises the following steps: and generating a rectangular shielding surface to randomly shield any position picture of the original image in the test set so as to obtain the incomplete image.
The image inpainting method for improving the edge incoherence of the inpainting area provided by the invention also has the following characteristics: wherein the generating network in step 2 has a multi-resolution encoder and a dual channel decoder.
The image inpainting method for improving the edge incoherence of the inpainting area provided by the invention also has the following characteristics: wherein the multi-resolution encoder has 5 sub-networks for feature extraction, and each sub-network for feature extraction includes 4 convolutional layers, the convolution operations in the convolutional layers all use convolution kernels of size, and the ELU functions are all used in the convolutional layers as activation functions, in the same layer of feature extraction sub-networks, the first convolutional layer uses a convolutional kernel with the moving step length of 2, the next three convolutional layers use a convolutional kernel with the moving step length of 1, the first three convolutional layers of the previous layer of feature extraction sub-networks and the first three convolutional layers of the next layer of feature extraction sub-networks perform one-to-one corresponding feature transmission, meanwhile, the second convolutional layer and the third convolutional layer are subjected to one-time reduction to half of the original size in the transmission process, and the fourth convolutional layer of the feature extraction sub-network of the previous layer transmits the output features to the first convolutional layer of the feature extraction sub-network of the next layer.
The image inpainting method for improving the edge incoherence of the inpainting area provided by the invention also has the following characteristics: the dual-channel decoder is provided with two feature reduction sub-networks, the first feature reduction sub-network is formed by connecting four soft attention mechanism network layers and a deconvolution network layer in series, the second feature reduction sub-network is formed by connecting five deconvolution network layers in series, the first reduction sub-network is in jump connection with the multi-resolution decoder, and the corresponding network layers between the first reduction sub-network and the second reduction sub-network have feature transfer relations.
The image inpainting method for improving the edge incoherence of the inpainting area provided by the invention also has the following characteristics: the relative discrimination network in the step 2 divides the regions of the complete image, measures the truth of each region through a relative countermeasure loss function, and discriminates by adopting a LeakyReLU activation function.
The image inpainting method for improving the edge incoherence of the inpainting area provided by the invention also has the following characteristics: wherein, the learning rate in step 3 is 0.0001, the relative confrontation loss is calculated by a relative confrontation loss function, and the relative confrontation loss function is:
Figure BDA0002730480330000041
Figure BDA0002730480330000042
the deeply supervised spatial loss is calculated from a spatial loss function acting on the characteristics of each layer resolution output of the dual channel decoder, and the formula is as follows:
Figure BDA0002730480330000043
the regional spatial loss is calculated by a regional spatial loss function, wherein the regional spatial loss function is as follows:
Lloral=||(xr-xf)⊙(1-M)||1
in the formula (I), the compound is shown in the specification,
Figure BDA0002730480330000044
and
Figure BDA0002730480330000045
for loss of the generating network and loss of the relative discriminating network, x, respectivelyrAnd xfRespectively an original image and a complete image restored with a generating network,
Figure BDA0002730480330000051
and
Figure BDA0002730480330000052
are respectively based on xrAnd xfAverage value of (D)RaFor relative discriminant networks, x is the number of missed portionsrAnd xfThe picture contents in (1) are consistent,
Figure BDA0002730480330000053
for the original image of the corresponding nth layer resolution size, pn(fn) Representing the result of compressing the multi-channel feature from the n-th layer into a three-channel feature, which indicates the multiplication of the corresponding elements of the matrix, M is such that the original image xrBecome incomplete pictureFor the occlusion plane of the image, M is a single-channel tensor, the numerical value corresponding to the missing region is 0, and the numerical value corresponding to the non-missing region is 1.
The image inpainting method for improving the edge incoherence of the inpainting area provided by the invention also has the following characteristics: wherein, the evaluation indexes in the step 4 are peak signal-to-noise ratio PSNR, average absolute error MAE, mean square error MSE and structure similarity SSIM.
Action and Effect of the invention
According to the image repairing method for improving the edge incoherence phenomenon of the repairing area, the advantages of the confrontation network in deep learning are combined, the generation network and the relative judgment network of multi-resolution information and a soft attention mechanism are adopted, the image repairing is completed through the iterative network model, the repairing accuracy is high, and a better repairing result is obtained.
Therefore, the image repairing method for improving the edge incoherence phenomenon of the repairing area can automatically and effectively repair the defect-removed image. Compared with the prior method, the repaired image texture has more continuity, the picture content has more detail, and the overall visual effect is more vivid.
Drawings
FIG. 1 is a flow chart of an image inpainting method of the present invention based on a relative confrontation generation network using image multi-resolution information and a soft attention mechanism;
FIG. 2 is a schematic diagram of an original image in an embodiment of the invention;
FIG. 3 is a diagram of a masked image in an embodiment of the invention;
FIG. 4 is a mask diagram of a mask in an embodiment of the invention;
fig. 5 is a schematic structural diagram of a relative countermeasure generation network in an embodiment of the invention;
FIG. 6 is a schematic diagram of a soft attention mechanism according to an embodiment of the present invention;
fig. 7 is a diagram of an image restoration result in an embodiment of the present invention.
Detailed Description
In order to make the technical means and functions of the present invention easy to understand, the present invention is specifically described below with reference to the embodiments and the accompanying drawings.
FIG. 1 is a flow chart of an image inpainting method of the present invention based on a relative confrontation generation network that utilizes image multi-resolution information and a soft attention mechanism.
As shown in fig. 1, an image inpainting method for improving the edge incoherence of an inpainting region according to the present invention is used for inpainting a defective image, and includes the following steps:
step 1, acquiring an original image data set, dividing the original image data set into a test set and a training set, and then preprocessing the original image in the test set to obtain a preprocessed test set with incomplete images.
In the invention, the pretreatment specifically comprises the following steps: and generating a rectangular shielding surface to randomly shield any position picture of the original image in the test set so as to obtain the incomplete image.
In the invention, the preprocessing of the data set can provide a proper data image for network model optimization and network model screening.
And 2, constructing a relative countermeasure network model, wherein the relative countermeasure network model comprises a generation network and a relative judgment network which utilize image multi-resolution information and a soft attention mechanism, the generation network is used for repairing the incomplete image to obtain a complete image, and the judgment network is used for judging the complete image to obtain a probability feature vector for describing the truth and the falseness of the complete image.
In the present invention, the soft attention mechanism is a method for constructing the similarity between the characteristics of the missing region and the reserved region. The soft idea is mainly reflected in that firstly, the sense field with the size of 3 × 3 and the sense field with the size of 5 × 5 are used to determine and calculate the value on the occlusion surface, secondly, the reverse thinking is used to determine whether the feature square block corresponding to the calculated value of the occlusion surface belongs to the non-reserved area, the feature square block corresponding to the non-reserved area and the feature square block corresponding to the reserved area are distinguished, and then the feature square block corresponding to the non-reserved area and the feature square block corresponding to the reserved area are used again, and the similarity between the feature square block in the reserved area and the feature square block in the non-reserved area is calculated through the cosine similarity theorem. And finally, taking the similarity as a weight, and carrying out weighted summation on the feature blocks of the reserved area to form a new feature block of the unreserved area.
In addition, the generation network is divided into an encoder and a decoder, and the encoder can effectively utilize the multi-resolution information of the image, while the generation network in the invention has a multi-resolution encoder and a dual-channel decoder.
Wherein the multi-resolution encoder has 5 sub-networks for feature extraction, and each sub-network for feature extraction includes 4 convolutional layers, the convolution operations in the convolutional layers all use convolution kernels of size, and the ELU functions are all used in the convolutional layers as activation functions, in the same layer of feature extraction sub-networks, the first convolutional layer uses a convolutional kernel with the moving step length of 2, the next three convolutional layers use a convolutional kernel with the moving step length of 1, the first three convolutional layers of the previous layer of feature extraction sub-networks and the first three convolutional layers of the next layer of feature extraction sub-networks perform one-to-one corresponding feature transmission, meanwhile, the second convolutional layer and the third convolutional layer are reduced to half of the original size in a one-time mode in the transmission process, and the fourth convolutional layer of the feature extraction sub-network of the previous layer transmits the output features to the first convolutional layer of the feature extraction sub-network of the next layer; the dual-channel decoder is provided with two feature reduction sub-networks, the first feature reduction sub-network is formed by connecting four soft attention mechanism network layers and a deconvolution network layer in series, the second feature reduction sub-network is formed by connecting five deconvolution network layers in series, the first reduction sub-network is in jump connection with the multi-resolution decoder, and the corresponding network layers between the first reduction sub-network and the second reduction sub-network have feature transfer relations.
In the invention, the relative discrimination network divides the complete image into regions, measures the truth of each region through a relative countermeasure loss function, and discriminates by adopting a LeakyReLU activation function.
And 3, inputting the original images of the training set into a relative countermeasure network model, then respectively carrying out network optimization training on the generation network and the relative judgment network, respectively calculating relative countermeasure loss, deeply supervised space loss and regional space loss, optimizing network parameters of the generation network and the relative judgment network based on a back propagation algorithm and an Adam optimizer, then carrying out iterative updating on the network parameters according to a set learning rate, and alternately updating the generation network and the relative judgment network, wherein when the generated network repaired complete image is close to the original image frame without defects, the network training is completed.
In the invention, the optimization of the relative confrontation network model specifically finds out proper network parameters, so that the output restored complete image of the generated network is more consistent with the original image. The screening of the relative confrontation network model mainly finds the image which can not participate in the model optimization process through proper indexes, so that the relative confrontation network model with the best repairing effect is obtained.
In the present invention, the learning rate is set to 0.0001.
Secondly, the relative confrontation loss is calculated by a relative confrontation loss function, and the formula of the relative confrontation loss function is as follows:
Figure BDA0002730480330000091
Figure BDA0002730480330000092
thirdly, the spatial loss of the deep supervision is calculated by a spatial loss function, and the spatial loss function acts on the characteristics of each layer resolution output of the dual-channel decoder, and the formula is as follows:
Figure BDA0002730480330000093
finally, the regional spatial loss is calculated by a regional spatial loss function, which is:
Llocal=||(xr-xf)⊙(1-M)||1
in the above-mentioned formula, the compound of formula,
Figure BDA0002730480330000094
and
Figure BDA0002730480330000095
for loss of the generating network and loss of the relative discriminating network, x, respectivelyrAnd xfRespectively an original image and a complete image restored with a generating network,
Figure BDA0002730480330000096
and
Figure BDA0002730480330000097
are respectively based on xrAnd xfAverage value of (D)RaFor relative discriminant networks, x is the number of missed portionsrAnd xfThe picture contents in (1) are consistent,
Figure BDA0002730480330000098
for the original image of the corresponding nth layer resolution size, pn(fn) Representing the result of compressing the multi-channel feature from the n-th layer into a three-channel feature, which indicates the multiplication of the corresponding elements of the matrix, M is such that the original image xrAnd M is a single-channel tensor, the numerical value corresponding to the missing area is 0, and the numerical value corresponding to the non-missing area is 1.
And 4, inputting the preprocessed incomplete images of the test set into the trained relative confrontation network model, judging whether the relative confrontation generation network is over-fitted or not through the evaluation indexes, returning to the step 3 if judging that the relative confrontation generation network is over-fitted, and reserving the generation network after training as an image restoration network if judging that the relative confrontation generation network is not over-fitted.
In the invention, the evaluation indexes are peak signal-to-noise ratio PSNR, mean absolute error MAE, mean square error MSE and structure similarity SSIM.
And 5, inputting the incomplete image to be repaired into an image repairing network to obtain a repaired complete image.
Example (b):
the embodiment takes a network model of the CelebA-HQ face data set as an example.
1. Data set preprocessing
Fig. 2 is a schematic diagram of an original image in an embodiment of the present invention, fig. 3 is a schematic diagram of an image with a mask in an embodiment of the present invention, and fig. 4 is a schematic diagram of a mask in an embodiment of the present invention.
First, the CelebA-HQ dataset was divided into a training set and a test set as a whole. The training set includes 27000 original images, and the testing set includes 3000 original images, and the original images are shown in fig. 2. The original image from the CelebA-HQ test set is then randomly covered by a rectangular mask, as shown in FIG. 4, so that the defect image to be repaired is obtained, as shown in FIG. 3, and the white squares in FIG. 4 can have random positions.
2. Construction of network model
Fig. 5 is a schematic structural diagram of a relative countermeasure generation network in the embodiment of the present invention.
As shown in FIG. 5, the inputs in the network are a defect image and a mask image of the same size. In the soft attention mechanism, the soft attention module receives the feature output from the feature extraction sub-network encoding the same corresponding layer and the feature delivered by the previous layer to restore or repair with weighting.
FIG. 6 is a schematic diagram of a soft attention mechanism according to an embodiment of the present invention.
As shown in fig. 6, two features are tiled with a size of 3 × 3, and two sets of tiles are rearranged along the direction of the channel to obtain two sets of feature tiles. Meanwhile, the masks are divided into two groups of mask squares according to the same principle and the sizes of the receptive fields of 3 multiplied by 3 and 5 multiplied by 5. The first set of feature blocks from the encoder is used to calculate similarity, and the second set of feature blocks from the previous layer is the object being replaced. Two sets of mask blocks are used to determine the similarity to be used. Finally, the weighted features are output after passing through four parallel but interconnected convolutional layers.
The relative discrimination network will input the repaired complete image and the corresponding original image in sequence, as shown in fig. 5. The discriminator is operated by five layers of convolution, the convolution kernel size of each layer is 5 x 5, and the step size of the shift is 2. The relative discrimination network measures whether the input complete image is more true or false according to a certain size area.
3. Optimization of network models
And inputting the original images in the training set of CelebA-HQ into a relative confrontation network model, and respectively performing network optimization training on the generated network and the relative discrimination network. The purpose of the training is to make the relative discrimination network judge that the original image is more true and the restored complete image is more false, and to make the generation network generate the restored complete image which makes the relative discrimination network more true. Specifically, the learning rates of the generator and the discriminator in this embodiment are set to 0.0001, the epoch of the loop iteration is 50, and the trained network model is obtained after the optimization is completed.
4. Screening of network models
And (3) acting the trained network model on the incomplete image of the CelebA-HQ test set, judging whether the trained model is over-fitted through four indexes of PSNR, MAE, MSE and SSIM, and if the over-fitting phenomenon is judged, correspondingly reducing the value of epoch. And if judging that no overfitting occurs, keeping the generation network after training is completed.
In this embodiment, it is found that for the CelebA-HQ dataset, when the epoch is 50, a good repairing effect can be obtained, and no overfitting phenomenon occurs.
5. Repairing a defective image to be repaired
Fig. 7 is a diagram of an image restoration result in an embodiment of the present invention.
The incomplete image to be repaired and the corresponding mask are input into the screened generation network, so that a repaired complete image is obtained, as shown in fig. 7.
Through the 5 steps, a repaired complete image is finally obtained, and therefore the repairing method capable of well repairing the incomplete image is obtained.
Effects and effects of the embodiments
Compared with other methods, the repairing method of the embodiment can solve the problem of discontinuous content of the boundary of the repairing area, and meanwhile, the integrally repaired picture can be more vivid and reliable.
According to the image repairing method for improving the edge incoherence phenomenon of the repairing area, the advantages of the confrontation network in deep learning are combined, the image repairing is completed through the iterative network model by adopting the multi-resolution information and the generation network and the relative judgment network of the soft attention mechanism, the repairing accuracy is high, and a better repairing result is obtained.
Therefore, the image inpainting method for improving the edge incoherence of the inpainting area of the embodiment can automatically and effectively repair the defect-removed image. Compared with the prior method, the repaired image texture has more continuity, the picture content has more detail, and the overall visual effect is more vivid.
The above embodiments are preferred examples of the present invention, and are not intended to limit the scope of the present invention.

Claims (8)

1. An image repairing method for improving the edge incoherence phenomenon of a repairing area is used for repairing a defective image and is characterized by comprising the following steps:
step 1, acquiring an original image data set, dividing the original image data set into a test set and a training set, and then preprocessing the original images in the test set to obtain a preprocessed test set with the defective images;
step 2, constructing a relative countermeasure network model, wherein the relative countermeasure network model comprises a generating network and a relative judging network which utilize image multi-resolution information and a soft attention mechanism, the generating network is used for repairing the incomplete image to obtain a complete image, and the judging network is used for judging the complete image to obtain a probability feature vector for describing the truth and falseness of the complete image;
step 3, inputting the original images of the training set into the relative countermeasure network model, then respectively performing network optimization training on the generation network and the relative judgment network, respectively calculating relative countermeasure loss, deeply supervised spatial loss and regional spatial loss, optimizing network parameters of the generation network and the relative judgment network based on a back propagation algorithm and an Adam optimizer, then performing iterative updating of the network parameters according to a set learning rate, and alternately updating the generation network and the relative judgment network, and when the complete images repaired by the generation network are close to the images of the original images without defects, completing the network training;
step 4, inputting the preprocessed incomplete images of the test set into the trained relative confrontation network model, judging whether the relative confrontation generation network is over-fitted through evaluation indexes, returning to the step 3 if over-fitting is judged, and reserving the generation network after training as an image restoration network if over-fitting is not judged;
and 5, inputting the incomplete image to be repaired into the image repairing network to obtain a repaired complete image.
2. The image inpainting method for improving the edge incoherence of the inpainting area according to claim 1, wherein:
wherein the pretreatment in the step 1 specifically comprises: and generating a rectangular shielding surface to randomly shield any position picture of the original image in the test set so as to obtain the incomplete image.
3. The image inpainting method for improving the edge incoherence of the inpainting area according to claim 1, wherein:
wherein the generation network in step 2 has a multi-resolution encoder and a dual channel decoder.
4. The image inpainting method for improving the edge incoherence of the inpainting area according to claim 3, wherein:
wherein the multi-resolution encoder has 5 sub-networks of feature extraction, and each sub-network of feature extraction contains 4 convolutional layers, the convolutional operations in the convolutional layers all use convolutional kernels of size, and the convolutional layers all use ELU function as activation function,
in the same layer of feature extraction sub-networks, the first convolution layer uses convolution kernels with the moving step length of 2, the next three convolution layers use convolution kernels with the moving step length of 1, the first three convolution layers of the previous layer of feature extraction sub-network and the first three convolution layers of the next layer of feature extraction sub-network can perform one-to-one corresponding feature transmission, meanwhile, the second convolution layer and the third convolution layer can perform one-time reduction to half of the original size in the transmission process, and the fourth convolution layer of the previous layer of feature extraction sub-network can transmit output features to the first convolution layer of the next layer of feature extraction sub-network.
5. The image inpainting method for improving the edge incoherence of the inpainting area according to claim 3, wherein:
wherein the dual-channel decoder has two feature reduction sub-networks, a first of the feature reduction sub-networks is formed by connecting four soft attention mechanism network layers and one deconvolution network layer in series, a second of the feature reduction sub-networks is formed by connecting five deconvolution network layers in series,
the first restoring sub-network is in hop connection with the multi-resolution decoder, and the corresponding network layer between the first restoring sub-network and the second restoring sub-network has a feature transfer relationship.
6. The image inpainting method for improving the edge incoherence of the inpainting area according to claim 1, wherein:
wherein, the relative discriminant network in the step 2 divides the complete image into regions, and measures the truth of each region through a relative confrontation loss function,
and the relative discrimination network adopts a LeakyReLU activation function to discriminate.
7. The image inpainting method for improving the edge incoherence of the inpainting area according to claim 1, wherein:
wherein the learning rate in the step 3 is 0.0001,
the relative confrontation loss is calculated by a relative confrontation loss function, and the formula of the relative confrontation loss function is as follows:
Figure FDA0002730480320000041
Figure FDA0002730480320000042
the deeply supervised spatial loss is calculated from a spatial loss function, and the spatial loss function is acted on the characteristics of each layer resolution output of the dual-channel decoder, and the formula is as follows:
Figure FDA0002730480320000043
the regional spatial loss is calculated by a regional spatial loss function, wherein the regional spatial loss function is as follows:
Llocal=||(xr-xf)⊙(1-M)||1
in the above-mentioned formula, the compound of formula,
Figure FDA0002730480320000044
and
Figure FDA0002730480320000045
respectively are uncookedLoss of the network and relative discriminant network loss, xrAnd xfRespectively an original image and a complete image restored with a generating network,
Figure FDA0002730480320000048
and
Figure FDA0002730480320000046
are respectively based on xrAnd xfAverage value of (D)RaFor relative discriminant networks, x is the number of missed portionsrAnd xfThe picture contents in (1) are consistent,
Figure FDA0002730480320000047
for the original image of the corresponding nth layer resolution size, pn(fn) Representing the result of compressing the multi-channel signature from the nth layer into a three-channel signature,
indicates multiplication of corresponding elements of the matrix, M is such that the original image xrAnd M is a single-channel tensor, the numerical value corresponding to the missing area is 0, and the numerical value corresponding to the non-missing area is 1.
8. The image inpainting method for improving the edge incoherence of the inpainting area according to claim 1, wherein:
wherein, the evaluation indexes in the step 4 are peak signal-to-noise ratio PSNR, average absolute error MAE, mean square error MSE and structure similarity SSIM.
CN202011116700.9A 2020-10-19 2020-10-19 Image repairing method for improving edge incoherence phenomenon of repairing area Pending CN114387170A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011116700.9A CN114387170A (en) 2020-10-19 2020-10-19 Image repairing method for improving edge incoherence phenomenon of repairing area

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011116700.9A CN114387170A (en) 2020-10-19 2020-10-19 Image repairing method for improving edge incoherence phenomenon of repairing area

Publications (1)

Publication Number Publication Date
CN114387170A true CN114387170A (en) 2022-04-22

Family

ID=81194209

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011116700.9A Pending CN114387170A (en) 2020-10-19 2020-10-19 Image repairing method for improving edge incoherence phenomenon of repairing area

Country Status (1)

Country Link
CN (1) CN114387170A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117078564A (en) * 2023-10-16 2023-11-17 北京网动网络科技股份有限公司 Intelligent generation method and system for video conference picture

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109559287A (en) * 2018-11-20 2019-04-02 北京工业大学 A kind of semantic image restorative procedure generating confrontation network based on DenseNet
CN110222628A (en) * 2019-06-03 2019-09-10 电子科技大学 A kind of face restorative procedure based on production confrontation network
US20190295227A1 (en) * 2018-03-26 2019-09-26 Adobe Inc. Deep patch feature prediction for image inpainting

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190295227A1 (en) * 2018-03-26 2019-09-26 Adobe Inc. Deep patch feature prediction for image inpainting
CN109559287A (en) * 2018-11-20 2019-04-02 北京工业大学 A kind of semantic image restorative procedure generating confrontation network based on DenseNet
CN110222628A (en) * 2019-06-03 2019-09-10 电子科技大学 A kind of face restorative procedure based on production confrontation network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
陈飞宇: "EIGAN: ENHANCED INPAINTING GENERATIVE ADVERSARIAL NETWORK", ICME, 10 July 2020 (2020-07-10), pages 1 - 6, XP033808187, DOI: 10.1109/ICME46284.2020.9102794 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117078564A (en) * 2023-10-16 2023-11-17 北京网动网络科技股份有限公司 Intelligent generation method and system for video conference picture
CN117078564B (en) * 2023-10-16 2024-01-12 北京网动网络科技股份有限公司 Intelligent generation method and system for video conference picture

Similar Documents

Publication Publication Date Title
Zhang et al. Multi-level fusion and attention-guided CNN for image dehazing
Huang et al. Self-filtering image dehazing with self-supporting module
CN115601549A (en) River and lake remote sensing image segmentation method based on deformable convolution and self-attention model
WO2024040973A1 (en) Multi-scale fused dehazing method based on stacked hourglass network
CN111832484A (en) Loop detection method based on convolution perception hash algorithm
Hong et al. Uncertainty-driven dehazing network
CN111723693A (en) Crowd counting method based on small sample learning
CN113160085B (en) Water bloom shielding image data collection method based on generation countermeasure network
CN114021704B (en) AI neural network model training method and related device
CN113971764B (en) Remote sensing image small target detection method based on improvement YOLOv3
CN113378775B (en) Video shadow detection and elimination method based on deep learning
CN115439442A (en) Industrial product surface defect detection and positioning method and system based on commonality and difference
CN116703885A (en) Swin transducer-based surface defect detection method and system
CN114757862B (en) Image enhancement progressive fusion method for infrared light field device
CN113159158A (en) License plate correction and reconstruction method and system based on generation countermeasure network
CN113111945B (en) Antagonistic sample defense method based on transformation self-encoder
CN117237994B (en) Method, device and system for counting personnel and detecting behaviors in oil and gas operation area
CN114387170A (en) Image repairing method for improving edge incoherence phenomenon of repairing area
CN116912675B (en) Underwater target detection method and system based on feature migration
CN117994573A (en) Infrared dim target detection method based on superpixel and deformable convolution
CN111275751B (en) Unsupervised absolute scale calculation method and system
CN117689617A (en) Insulator detection method based on defogging constraint network and series connection multi-scale attention
CN117391920A (en) High-capacity steganography method and system based on RGB channel differential plane
CN117351194A (en) Graffiti type weak supervision significance target detection method based on complementary graph inference network
CN117037052A (en) Water surface dead fish detection method based on multi-scale feature fusion and attention mechanism

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination