Nothing Special   »   [go: up one dir, main page]

CN111861886B - Image super-resolution reconstruction method based on multi-scale feedback network - Google Patents

Image super-resolution reconstruction method based on multi-scale feedback network Download PDF

Info

Publication number
CN111861886B
CN111861886B CN202010682515.XA CN202010682515A CN111861886B CN 111861886 B CN111861886 B CN 111861886B CN 202010682515 A CN202010682515 A CN 202010682515A CN 111861886 B CN111861886 B CN 111861886B
Authority
CN
China
Prior art keywords
image
resolution
feature
scale
conv
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010682515.XA
Other languages
Chinese (zh)
Other versions
CN111861886A (en
Inventor
陈晓
孙超文
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Information Science and Technology
Original Assignee
Nanjing University of Information Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Information Science and Technology filed Critical Nanjing University of Information Science and Technology
Priority to CN202010682515.XA priority Critical patent/CN111861886B/en
Publication of CN111861886A publication Critical patent/CN111861886A/en
Application granted granted Critical
Publication of CN111861886B publication Critical patent/CN111861886B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4007Scaling of whole images or parts thereof, e.g. expanding or contracting based on interpolation, e.g. bilinear interpolation
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to an image super-resolution reconstruction method based on a multi-scale feedback network, which comprises the following steps: (1) creating an image dataset; (2) Extracting features of an input image, recursively realizing low-resolution and high-resolution feature mapping by using a multi-scale upper projection unit and a multi-scale lower projection unit to obtain high-resolution feature images with different depths, performing convolution calculation on the high-resolution feature images to obtain residual images, and finally interpolating the low-resolution images and adding the residual images to obtain output images; (3) Training a multi-scale feedback network by using the data set, and generating a trained network model; (4) And inputting the low-resolution image to be processed into a trained network to obtain an output high-resolution image. The method can train networks with different depths and expand to other amplification factors through small parameter adjustment, saves training cost, can realize amplification with larger amplification factors, and improves peak signal-to-noise ratio and structural similarity of reconstructed images.

Description

Image super-resolution reconstruction method based on multi-scale feedback network
Technical Field
The invention relates to an image super-resolution reconstruction method based on a multi-scale feedback network, and belongs to the field of computer vision and the field of deep learning.
Background
The Super-resolution (SR) reconstruction technology is an important image processing technology in the field of computer vision, and is widely applied to the fields of medical imaging, security monitoring, remote sensing image quality improvement, image compression and target detection. Image super-Resolution reconstruction aims at building a suitable model to convert a Low Resolution (LR) image into a corresponding High Resolution (HR) image. Since a given LR image input corresponds to multiple possible HR images, the SR reconstruction problem is a challenging, pathologic inverse problem.
Currently, the proposed SR reconstruction methods are mainly divided into three categories, namely an interpolation-based method, a reconstruction-based method and a learning-based method. Among them, the SR method based on deep learning has received attention in recent years with its superior reconstruction performance. SRCNN is taken as an mountain-opening operation in the SR field of deep learning technology, and fully demonstrates the superiority of the convolutional neural network. Therefore, many networks have proposed a series of SR methods based on convolutional neural networks based on the srcan architecture. Depth as an important factor can provide a larger receptive field and more contextual information to the network, however increasing depth is extremely prone to two problems: gradient vanishing/exploding and a large number of network parameters.
In order to solve the gradient problem, researchers have proposed residual learning and successfully trained deeper networks, and in addition, some networks introduce dense connections to alleviate the gradient vanishing problem and encourage feature reuse; to reduce the parameters, researchers have proposed recursive learning to aid in weight sharing. Thanks to these mechanisms, many networks tend to construct deeper and more complex network structures to obtain higher evaluation indexes, however, many networks have the following problems:
the first and many SR methods, although realizing high performance of the deep network, neglect the training difficulty of the network, so that a huge training set is required to be spent, and more training skills and time are input.
Second, most SR methods learn hierarchical feature representations directly from LR inputs in a feed-forward manner and map to output space, this one-way mapping being dependent on limited features in the LR image. And many feedforward networks requiring preprocessing operations are only suitable for a single magnification, and the transition to other magnification requires cumbersome operations and extremely inflexible.
Disclosure of Invention
The invention provides an image super-resolution reconstruction method based on a multi-scale feedback network in order to solve the problems in the prior art. The method is characterized by comprising the following steps of:
step one, establishing a data set by using an image degradation model;
step two, constructing a multi-scale feedback network, wherein the multi-scale feedback network comprises an image feature extraction module, an image feature mapping module and a high-resolution image calculation module;
step 2.1, extracting image features;
network-input LR image I LR Input feature extraction module f 0 Generating an initial LR profile L 0
L 0 =f 0 (I LR )
Let conv (f, n) represent the convolution layer, f is the convolution kernel size, n is the channel number; f in the above 0 Consists of 2 convolutional layers conv (3, n 0 ) And conv (1, n), n0 represents the number of channels of the initial low resolution feature extraction layer, n represents the number of input channels in the feature mapping module; firstly using conv (3, n) 0 ) Generating shallow features L with low resolution image information from an input 0 Then the conv (1, n) is used to change the channel number from n 0 Reducing to n;
step 2.2, mapping image features;
low resolution feature map L g-1 Input recursive feedback module for generating high resolution feature map H g
Wherein G represents the number of multi-scale projection groups, i.e., the number of recursions;representing the feature mapping process for the multi-scale projection group in the g-th recursion. When g is equal to 1, the initial characteristic diagram L is represented 0 As input to the first multiscale projection group, when g is greater than 1, the LR signature L to be generated by the previous multiscale projection group is represented g-1 As a current input;
step 2.3, calculating a high-resolution image;
computing a residual image from the plurality of HR feature map depth cascades using:
I Res =f RM ([H 1 ,H 2 ,…,H g ])
wherein, [ H ] 1 ,H 2 ,…,H g ]Depth concatenation representing multiple HR feature maps, f RM Representing conv (3, 3) operation, I Res Is a residual image.
Image obtained by interpolating LR image and residual image I Res Adding to obtain reconstructed high resolution image I SR
I SR =I Res +f US (I LR )
Wherein f US Representing an interpolation operation.
Training a multi-scale feedback network;
and step four, reconstructing an image.
The technical scheme is further designed as follows: the process of establishing the data set by using the image degradation model in the first step is that I is given LR Representing LR images, I HR Representing the corresponding HR image, the degradation process is expressed as:
I LR =D(I HR ;δ)
modeling a degradation map that generates an LR image from HR images and modeling degradation as a single downsampling operation:
wherein ∈ s Indicating that the amplification factor s is downsampled and delta is a scale factor.
The interpolation algorithm is a bilinear interpolation algorithm or a bicubic interpolation algorithm.
The loss function of the training multi-scale feedback network in the third step is as follows:
wherein x is a set of weight parameters and bias parameters, i represents a serial number of iterative training in the whole training process, and m represents the number of training images.
The beneficial effects of the invention are as follows:
the modularized end-to-end system structure not only can flexibly train networks with different depths through small parameter adjustment and be arbitrarily expanded to other amplification factors, greatly saves training cost, but also can successfully realize amplification of a larger multiple (8 times), and improves peak signal-to-noise ratio and structural similarity of reconstructed images. The method can also relieve the influence of ringing effect and chessboard artifact based on the convolutional neural network method, predicts more high-frequency details and suppresses smooth components, so that the reconstructed image has clearer and sharper edge characteristics and is closer to a real high-resolution image.
Drawings
FIG. 1 is a flow chart of the method of the present invention;
FIG. 2 is a block diagram of a multi-scale feedback network;
FIG. 3 is a block diagram of a projection unit on multiple scales in a network;
fig. 4 is a block diagram of a multi-scale projection unit in a network.
Detailed Description
The invention will now be described in detail with reference to the accompanying drawings and specific examples.
Examples
As shown in fig. 1, the image super-resolution reconstruction method based on the multi-scale feedback network of the present embodiment includes the following steps:
step 1, establishing a data set by using an image degradation model;
set I LR Representing LR images, I HR Representing the corresponding HR image, the degradation process is expressed as:
I LR =D(I HR ;δ) (1)
modeling a degradation map that generates an LR image from HR images and modeling degradation as a single downsampling operation:
wherein ∈ s Indicating that the amplification factor s is downsampled and delta is a scale factor.
In this embodiment, bicubic interpolation with antialiasing is used as a downsampling operation, and m training images in DIV2K are acquired as a training set. Set5, set14, urban100, BSD100, and Manga109 were chosen as standard test sets and downsampled 2-fold, 3-fold, 4-fold, and 8-fold, respectively, using bicubic interpolation algorithms.
Step 2, constructing a multi-scale feedback network; the network structure is shown in fig. 2, and comprises the following steps:
step 2.1, extracting image features;
will initial LR image I LR Input feature extraction module f 0 Generating an initial LR profile L 0
L 0 =f 0 (I LR ) (3)
Let conv (f, n) denote the convolution layer, f be the convolution kernel size, n be the number of channels. Wherein f 0 Consists of 2 convolutional layers conv (3, n 0 ) And conv (1, n), n0 represents the number of channels of the initial LR image feature extraction layer, and n represents the number of input channels in the feature mapping module. Firstly using conv (3, n) 0 ) Generating shallow features L with LR image information from input 0 Then the conv (1, n) is used to change the channel number from n 0 Down to n.
Step 2.2, mapping image features;
and recursively realizing low-resolution and high-resolution feature mapping by using the multi-scale upper projection unit and the multi-scale lower projection unit to form a projection group, so as to obtain high-resolution feature graphs with different depths. Low resolution feature map L g-1 Input recursive feedback module for generating high resolution feature map H g
Wherein G represents multiple scalesThe number of degree projection sets, i.e., the number of recursions;representing the feature mapping process for the multi-scale projection group in the g-th recursion. When g is equal to 1, the initial characteristic diagram L is represented 0 As input to the first multiscale projection group, when g is greater than 1, the LR signature L to be generated by the previous multiscale projection group is represented g-1 As the current input.
The operations include two operations of mapping LR features to HR features and mapping HR features to LR features, the structure of which is shown in fig. 3 and 4.
The multiscale projection unit maps LR features to HR features (structure shown in fig. 3) by six steps:
(1): LR profile L calculated from previous cycle g-1 As input, deconvolution with different kernel sizes is used, respectivelyAnd->Up-sampling is performed on both branches to obtain two HR profile +.>And->
And->Respectively Deconv1 (k) 1 N) and Deconv2 (k) 2 ,n),k 1 And k 2 The size of the deconvolution kernel is represented, and n represents the number of channels.
(2): HR feature mapAnd->Concatenating, using convolutions of different kernel sizes, respectively>And->Performing downsampling operations on both branches and generating two LR profiles>And->
And->Respectively Conv1 (k) 1 2 n) and Conv2 (k) 2 2 n), the number of channels per branch is changed from n to 2n.
(3): will LR characteristic diagramAnd->Cascade, pooling and dimension reduction operations using 1 x 1 convolution, +.>And->Mapping to an LR profile +.>
C u Conv (1, n) is represented, and the number of channels per branch is changed from 2n to n. And all 1 x 1 convolutions add nonlinear excitation to the learned representation of the previous layer.
(4): calculate the LR feature map L of the input g-1 And reconstructed LR feature mapsResidual error between->
(5): deconvolution with different kernel sizesAnd->Respectively->Performing up-sampling operation, wherein residual error in LR characteristic is mapped into HR characteristic, thereby generating new HR residual error characteristic +.>And->
And->Respectively represent deconvolution layer Deconv1 (k) 1 N) and Deconv2 (k) 2 N), the number of channels per branch remains n.
(6): residual HR featureAnd->Cascading and overlapping with the HR features cascaded in the step (2), outputting a final HR feature map H of the upper projection unit through 1X 1 convolution g
C h Conv (1, n) is shown, the total number of channels after addition is 2n, and the number of output channels is reduced to n by Conv (1, n) to keep the same as the number of input channels.
The multiscale downscaling unit maps HR features to LR features (structure shown in fig. 4) by six steps:
step (1): HR characteristic diagram H outputted by the projection unit on multiple scales of previous cycle g As input, convolutions of different kernel sizes are used, respectivelyAnd->Performing downsampling operations on the two branches to obtain two LR profiles +.>And
and->Respectively Conv1 (k) 1 N) and Conv2 (k) 2 ,n)。
Step (2): will LR characteristic diagramAnd->Cascade, respectively using deconvolution +.>And->Performing up-sampling operations on both branches and generating two HR profile +.>And->
And->Respectively Deconv1 (k) 1 2 n) and Deconv2 (k) 2 2 n), the number of channels per branch is changed from n to 2n.
Step (3): HR feature mapAnd->Cascading and obtaining HR profile +.1 by 1X 1 convolution>
C d Conv (1, n) is represented, and the number of channels per branch is changed from 2n to n.
Step (4): computing an input HR feature map H g And reconstructed HR feature mapsResidual error between->
Step (5): convolution with different kernel sizesAnd->Respectively->Downsampling, wherein residuals in the HR feature are mapped into the LR feature to generate a new LR residual feature +.>And->
And->Respectively represent the convolution layers Conv1 (k) 1 N) and Conv2 (k) 2 N), the number of channels per branch remains n.
Step (6): residual LR characterizationAnd->Cascading and overlapping with the LR features cascaded in the step 2, and outputting a final LR feature map L of the lower projection unit through 1X 1 convolution g
C l Conv (1, n) is shown, the total number of channels after addition is 2n, and the number of output channels is reduced to n by Conv (1, n) to keep the same as the number of input channels.
Step 2.3, calculating a high-resolution image;
the depth cascade of a plurality of high-resolution feature images is used for calculating residual images;
I Res =f RM ([H 1 ,H 2 ,…,H g ]) (23)
wherein, [ H ] 1 ,H 2 ,…,H g ]Depth concatenation representing multiple HR feature maps, f RM Representing conv (3, 3) operation, a series to be cascadedColumn HR feature map input conv (3, 3) generates residual image I Res
Image obtained by interpolating low resolution image and residual image I Res Adding to generate a reconstructed high resolution image I SR
I SR =I Res +f US (I LR ) (24)
Wherein f US Indicating an interpolation up-sampling operation, a bilinear interpolation algorithm, a bicubic interpolation algorithm, or other interpolation algorithm may be used.
Step 3, training a multi-scale feedback network;
the batch of networks is set to 16 and data enhancement is performed using rotation and flipping. LR images with different sizes and corresponding HR images are input according to the magnification coefficients. The network parameters were optimized using Adam with a momentum factor of 0.9 and a weight decay of 0.0001. The initial value of the learning rate was set to 0.0001, and the learning rate decayed to half of the original value 200 times per iteration.
Different kernel sizes and fills are designed in each branch of the multi-scale projection unit and the kernel sizes and step sizes are adjusted according to the corresponding magnification. Both input and output use RGB channels of a color image. The PReLU is used as an activation function after all convolution and deconvolution layers except for the reconstruction layer at the end of the network. Training the network according to the process of the step 2 by using the image data set of the step 1 until the cost loss is reduced to a set value and the training reaches the maximum number of iterations. By L 1 The function is taken as a loss function and expressed as follows:
wherein x is a set of weight parameters and bias parameters, i represents a serial number of iterative training in the whole training process, and m represents the number of training images.
Step 4, reconstructing an image;
and inputting the low-resolution image to be processed into a trained network to obtain an output high-resolution image.
The peak signal-to-noise ratio and the structural similarity are used as evaluation indexes to evaluate the model performance in 5 standard test sets of Set5, set14, urban100, BSD100 and Manga109, and all tests use y channels.
To verify the effectiveness and reliability of the method, it is compared with existing reconstruction methods at different magnifications. In the low magnification (×2, ×3, ×4), the present method was compared with the 21 advanced methods currently existing. Since many models are not suitable for high magnification (×8), the method is compared with 12 advanced methods. For x 2 amplification, the method achieves the best peak signal-to-noise results in the five reference data sets. However, for x 3, x 4 and x 8 amplifications, the peak signal to noise ratio and structural similarity of the method are superior to all other models. The advantages are relatively more pronounced with increasing amplification factor, especially for x 8, demonstrating the effectiveness of the present method in handling high magnification. In the five data sets, the method has higher objective evaluation indexes in terms of peak signal-to-noise ratio and structural similarity. It has been demonstrated that the present method not only tends to construct regular artificial patterns, but also is good at reconstructing irregular natural patterns. The method has advantages in adapting to various scene features and has surprising super-resolution reconstruction results for images with different features.
The multi-scale feedback network method designed in this embodiment uses only m (m=800) training images from DIV2K, and can still obtain superior reconstruction performance in 8-fold amplification over other existing methods through a relatively small training set. The multi-scale convolution is combined with a feedback mechanism, so that the method not only can learn rich hierarchical feature representations on multiple context scales and capture image features of different scales, but also can refine low-level representations by utilizing high-level features, and better characterize the interrelationship between HR and LR images. In addition to combining the high-level information and the low-level information, the local information and the global information are combined through global residual learning and local residual feedback fusion, so that the quality of the reconstructed image is better improved. In addition, the modularized end-to-end architecture enables the method to train the network with different depths flexibly and to be arbitrarily expanded to other magnification factors through small parameter adjustment. Compared with many advanced methods at present, the method has excellent reconstruction performance, and particularly has more obvious advantages in high magnification that many methods are not good.
The foregoing is merely a preferred embodiment of the present invention, and it should be noted that it will be apparent to those skilled in the art that modifications and variations can be made without departing from the technical principles of the present invention, and these modifications and variations should also be regarded as the scope of the invention, including but not limited to using the present method and its modifications and variations for other image processing aspects, such as image classification, detection, denoising, enhancement, etc.

Claims (4)

1. The image super-resolution reconstruction method based on the multi-scale feedback network is characterized by comprising the following steps of:
step one, an image data set is established by utilizing an image degradation model;
step two, constructing a multi-scale feedback network, wherein the multi-scale feedback network comprises an image feature extraction module, an image feature mapping module and a high-resolution image calculation module;
step 2.1, extracting image features;
inputting a network into a low resolution image I LR Input feature extraction module f 0 Generating an initial low resolution feature map L 0
L 00 (I LR )
Let conv (f, n) represent the convolution layer, f is the convolution kernel size, n is the channel number; f in the above 0 Consists of 2 convolutional layers conv (3, n 0 ) And conv (1, n), n 0 The number of channels representing the initial low resolution feature extraction layer, n representing the number of input channels in the feature mapping module; firstly using conv (3, n) 0 ) Generating shallow features L with low resolution image information from an input 0 Then the conv (1, n) is used to change the channel number from n 0 Reducing to n;
step 2.2, mapping image features;
the multi-scale upper projection unit and the multi-scale lower projection unit form a projection group to recursively realize low-resolution and high-resolution feature mapping, so as to obtain high-resolution feature graphs with different depths; low resolution feature map L g-1 Input recursive feedback module for generating high resolution feature map H g
Wherein G represents the number of multi-scale projection groups, i.e., the number of recursions;representing a feature mapping process of the multi-scale projection group in the g-th recursion; when g is equal to 1, the initial characteristic diagram L is represented 0 As input to the first multiscale projection group, when g is greater than 1, the LR signature L to be generated by the previous multiscale projection group is represented g-1 As a current input; />The operations include two operations of mapping LR features to HR features and mapping HR features to LR features;
the mapping of LR features to HR features proceeds as follows:
(1): LR profile L calculated from previous cycle g-1 As input, deconvolution with different kernel sizes is used, respectivelyAndup-sampling is performed on both branches to obtain two HR profile +.>And->
And->Respectively Deconv1 (k) 1 N) and Deconv2 (k) 2 ,n),k 1 And k 2 Indicating the size of the deconvolution kernel, n indicating the number of channels;
(2): HR feature mapAnd->Concatenating, using convolutions of different kernel sizes, respectively>And->Performing downsampling operations on both branches and generating two LR profiles>And->
And->Respectively Conv1 (k) 1 2 n) and Conv2 (k) 2 2 n), the number of channels of each branch is changed from n to 2n;
(3): will LR characteristic diagramAnd->Cascade, pooling and dimension reduction operations using 1 x 1 convolution, +.>And->Mapping to an LR profile +.>
C u Representing Conv (1, n), the general state of each branchThe number of tracks is changed from 2n to n; and all 1 x 1 convolutions add nonlinear excitation to the learning representation of the previous layer;
(4): calculate the LR feature map L of the input g-1 And reconstructed LR feature mapsResidual error between->
(5): deconvolution with different kernel sizesAnd->Respectively->Performing up-sampling operation, wherein residual error in LR characteristic is mapped into HR characteristic, thereby generating new HR residual error characteristic +.>And->
And->Respectively represent deconvolution layer Deconv1 (k) 1 N) and Deconv2 (k) 2 N), the number of channels of each branch is still n;
(6): residual HR featureAnd->Cascading and overlapping with the HR features cascaded in the step (2), outputting a final HR feature map H of the upper projection unit through 1X 1 convolution g
Ch represents Conv (1, n), the total channel number after addition is 2n, the output channel number is reduced to n by Conv (1, n), and the output channel number is consistent with the input channel number;
the mapping HR features to LR features proceeds as follows:
(1): HR characteristic diagram H outputted by the projection unit on multiple scales of previous cycle g As input, convolutions of different kernel sizes are used, respectivelyAnd->Performing downsampling operations on the two branches to obtain two LR profiles +.>And->
And->Respectively Conv1 (k) 1 N) and Conv2 (k) 2 ,n);
(2): will LR characteristic diagramAnd->Cascade, respectively using deconvolution +.>And->Performing up-sampling operations on both branches and generating two HR profile +.>And->
And->Respectively Deconv1 (k) 1 2 n) and Deconv2 (k) 2 2 n), the number of channels of each branch is changed from n to 2n;
step (3): HR feature mapAnd->Cascading and obtaining HR profile +.1 by 1X 1 convolution>
C d Representing Conv (1, n), the number of channels per branch is changed from 2n to n;
step (4): computing an input HR feature map H g And reconstructed HR feature mapsResidual error between->
Step (5): convolution with different kernel sizesAnd->Respectively->Downsampling, wherein residuals in the HR feature are mapped into the LR feature to generate a new LR residual feature +.>And->
And->Respectively are provided withRepresenting the convolutional layer Conv1 (k) 1 N) and Conv2 (k) 2 N), the number of channels of each branch is still n;
(6): residual LR characterizationAnd->Cascading and overlapping with the LR characteristic cascaded in the step (2), outputting a final LR characteristic diagram L of the lower projection unit through 1X 1 convolution g
C l Conv (1, n) is represented, the total channel number after addition is 2n, and the output channel number is reduced to n through Conv (1, n) and is consistent with the input channel number;
step 2.3, calculating a high-resolution image;
the depth cascade of a plurality of high-resolution feature images is used for calculating residual images;
l Res =f RM ([H 1 ,H 2 ,…,H g ])
wherein, [ H ] 1 ,H 2 ,…,H g ]Depth concatenation representing multiple high resolution feature maps, f RM Representing conv (3, 3) operation, I Res Is a residual image;
image obtained by interpolating low resolution image and residual image I Res Adding to generate a reconstructed high resolution image I SR
I SR =I Res +f US (I LR )
Wherein f US Representing an interpolation operation;
training a multi-scale feedback network;
step four, reconstructing an image;
and inputting the low-resolution image to be processed into a trained network to obtain an output high-resolution image.
2. The image super-resolution reconstruction method based on a multi-scale feedback network according to claim 1, wherein: the process of establishing the data set by using the image degradation model in the first step is that,
given I LR Representing low resolution images, I HR Representing the corresponding high resolution image, the degradation process is represented as:
I LR =D(I HR ;δ)
modeling a degradation map that generates a low resolution image from a high resolution image and modeling degradation as a single downsampling operation:
wherein ∈ s Indicating that the amplification factor s is downsampled and delta is a scale factor.
3. The image super-resolution reconstruction method based on a multi-scale feedback network according to claim 1, wherein: the interpolation algorithm is a bilinear interpolation algorithm or a bicubic interpolation algorithm.
4. The image super-resolution reconstruction method based on a multi-scale feedback network according to claim 1, wherein: the loss function of the training multi-scale feedback network in the third step is as follows:
wherein x is a set of weight parameters and bias parameters, i represents a serial number of iterative training in the whole training process, and m represents the number of training images.
CN202010682515.XA 2020-07-15 2020-07-15 Image super-resolution reconstruction method based on multi-scale feedback network Active CN111861886B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010682515.XA CN111861886B (en) 2020-07-15 2020-07-15 Image super-resolution reconstruction method based on multi-scale feedback network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010682515.XA CN111861886B (en) 2020-07-15 2020-07-15 Image super-resolution reconstruction method based on multi-scale feedback network

Publications (2)

Publication Number Publication Date
CN111861886A CN111861886A (en) 2020-10-30
CN111861886B true CN111861886B (en) 2023-08-08

Family

ID=72983037

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010682515.XA Active CN111861886B (en) 2020-07-15 2020-07-15 Image super-resolution reconstruction method based on multi-scale feedback network

Country Status (1)

Country Link
CN (1) CN111861886B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112418418B (en) * 2020-11-11 2024-08-06 无锡禹空间智能科技有限公司 Data processing method and device based on neural network, storage medium and server
CN112767427A (en) * 2021-01-19 2021-05-07 西安邮电大学 Low-resolution image recognition algorithm for compensating edge information
CN112927159B (en) * 2021-03-11 2022-08-02 清华大学深圳国际研究生院 True image denoising method based on multi-scale selection feedback network
CN113191949B (en) * 2021-04-28 2023-06-20 中南大学 Multi-scale super-resolution pathology image digitizing method, system and storage medium

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108550115A (en) * 2018-04-25 2018-09-18 中国矿业大学 A kind of image super-resolution rebuilding method
CN108734660A (en) * 2018-05-25 2018-11-02 上海通途半导体科技有限公司 A kind of image super-resolution rebuilding method and device based on deep learning
CN108921789A (en) * 2018-06-20 2018-11-30 华北电力大学 Super-resolution image reconstruction method based on recurrence residual error network
CN109035163A (en) * 2018-07-09 2018-12-18 南京信息工程大学 A kind of adaptive denoising method based on deep learning
CN109741260A (en) * 2018-12-29 2019-05-10 天津大学 A kind of efficient super-resolution method based on depth back projection network
CN110197468A (en) * 2019-06-06 2019-09-03 天津工业大学 A kind of single image Super-resolution Reconstruction algorithm based on multiple dimensioned residual error learning network
CN110276721A (en) * 2019-04-28 2019-09-24 天津大学 Image super-resolution rebuilding method based on cascade residual error convolutional neural networks
CN111192200A (en) * 2020-01-02 2020-05-22 南京邮电大学 Image super-resolution reconstruction method based on fusion attention mechanism residual error network

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9369621B2 (en) * 2010-05-03 2016-06-14 Invisage Technologies, Inc. Devices and methods for high-resolution image and video capture

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108550115A (en) * 2018-04-25 2018-09-18 中国矿业大学 A kind of image super-resolution rebuilding method
CN108734660A (en) * 2018-05-25 2018-11-02 上海通途半导体科技有限公司 A kind of image super-resolution rebuilding method and device based on deep learning
CN108921789A (en) * 2018-06-20 2018-11-30 华北电力大学 Super-resolution image reconstruction method based on recurrence residual error network
CN109035163A (en) * 2018-07-09 2018-12-18 南京信息工程大学 A kind of adaptive denoising method based on deep learning
CN109741260A (en) * 2018-12-29 2019-05-10 天津大学 A kind of efficient super-resolution method based on depth back projection network
CN110276721A (en) * 2019-04-28 2019-09-24 天津大学 Image super-resolution rebuilding method based on cascade residual error convolutional neural networks
CN110197468A (en) * 2019-06-06 2019-09-03 天津工业大学 A kind of single image Super-resolution Reconstruction algorithm based on multiple dimensioned residual error learning network
CN111192200A (en) * 2020-01-02 2020-05-22 南京邮电大学 Image super-resolution reconstruction method based on fusion attention mechanism residual error network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
黄思炜.基于深度学习的超分辨率图像重建算法研究.《中国优秀硕士学位论文全文数据库信息科技辑》.2018,(第11期),第I138-528页. *

Also Published As

Publication number Publication date
CN111861886A (en) 2020-10-30

Similar Documents

Publication Publication Date Title
CN112215755B (en) Image super-resolution reconstruction method based on back projection attention network
Yang et al. DRFN: Deep recurrent fusion network for single-image super-resolution with large factors
CN111861886B (en) Image super-resolution reconstruction method based on multi-scale feedback network
CN109903228B (en) Image super-resolution reconstruction method based on convolutional neural network
CN109389552B (en) Image super-resolution algorithm based on context-dependent multitask deep learning
CN110136062B (en) Super-resolution reconstruction method combining semantic segmentation
CN106228512A (en) Based on learning rate adaptive convolutional neural networks image super-resolution rebuilding method
CN106204447A (en) The super resolution ratio reconstruction method with convolutional neural networks is divided based on total variance
CN109636721B (en) Video super-resolution method based on countermeasure learning and attention mechanism
CN112837224A (en) Super-resolution image reconstruction method based on convolutional neural network
CN105631807A (en) Single-frame image super resolution reconstruction method based on sparse domain selection
Guo et al. Multiscale semilocal interpolation with antialiasing
Cao et al. New architecture of deep recursive convolution networks for super-resolution
CN113538246A (en) Remote sensing image super-resolution reconstruction method based on unsupervised multi-stage fusion network
CN104899835A (en) Super-resolution processing method for image based on blind fuzzy estimation and anchoring space mapping
CN116188272B (en) Two-stage depth network image super-resolution reconstruction method suitable for multiple fuzzy cores
CN111353938A (en) Image super-resolution learning method based on network feedback
CN116797456A (en) Image super-resolution reconstruction method, system, device and storage medium
CN109993701B (en) Depth map super-resolution reconstruction method based on pyramid structure
CN118447031A (en) Skin mirror image segmentation method based on diffusion fuzzy learning
CN116611995B (en) Super-resolution reconstruction method of handwritten text image based on depth expansion network
CN116188273A (en) Uncertainty-oriented bimodal separable image super-resolution method
CN110211059A (en) A kind of image rebuilding method based on deep learning
CN113538231B (en) Single image super-resolution reconstruction system and method based on pixel distribution estimation
CN114140317A (en) Image animation method based on cascade generation confrontation network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant