Nothing Special   »   [go: up one dir, main page]

CN115984406B - SS-OCT compression imaging method for deep learning and spectral domain airspace combined sub-sampling - Google Patents

SS-OCT compression imaging method for deep learning and spectral domain airspace combined sub-sampling Download PDF

Info

Publication number
CN115984406B
CN115984406B CN202310264537.8A CN202310264537A CN115984406B CN 115984406 B CN115984406 B CN 115984406B CN 202310264537 A CN202310264537 A CN 202310264537A CN 115984406 B CN115984406 B CN 115984406B
Authority
CN
China
Prior art keywords
neural network
image
mask
stripe
sampling
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310264537.8A
Other languages
Chinese (zh)
Other versions
CN115984406A (en
Inventor
凌玉烨
董振兴
张�杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Always Wuxi Medical Technology Co ltd
Original Assignee
Always Wuxi Medical Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Always Wuxi Medical Technology Co ltd filed Critical Always Wuxi Medical Technology Co ltd
Priority to CN202310264537.8A priority Critical patent/CN115984406B/en
Publication of CN115984406A publication Critical patent/CN115984406A/en
Application granted granted Critical
Publication of CN115984406B publication Critical patent/CN115984406B/en
Priority to PCT/CN2024/088631 priority patent/WO2024193719A1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • General Physics & Mathematics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Pathology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to the technical field of image processing, and discloses an SS-OCT compression imaging method for deep learning and spectrum domain airspace combined sub-sampling, which comprises the following steps: acquiring complete sample interference fringes and complete target detection interference fringes; step two: constructing and generating an image sampling system; step three: and inputting the complete target detection interference fringes into the image sampling system to obtain a target image. According to the invention, the mask neural network, the enhancement neural network and the stripe complement neural network are synchronously subjected to gradient descent update of parameters, so that the combined optimization of the two-dimensional mask and OCT image reconstruction is achieved, the problem of poor OCT image reconstruction effect caused by the adoption of an empirical fixed two-dimensional mask in the current OCT image sampling system is solved, and the OCT image reconstruction quality of the image sampling system is optimal under the same data compression ratio.

Description

SS-OCT compression imaging method for deep learning and spectral domain airspace combined sub-sampling
Technical Field
The invention relates to the technical field of image processing, in particular to an SS-OCT compression imaging method for deep learning and spectrum domain airspace combined sub-sampling.
Background
Swept-frequency optical coherence tomography (SS-OCT) is a non-invasive volumetric imaging modality, widely used in the biomedical field. Thanks to advances in laser technology, SS-OCT imaging rates (typically quantified by the a-line rate of the scanning light source) are increasing all the way from a few hertz to a few megahertz. For a typical 200 kHz SS-OCT, the data bandwidth may exceed 800 MB/s if 2048 spectral sample points are measured in 12 bits. Researchers have proposed various strategies to mitigate data bandwidth. However, existing approaches mostly achieve reduced data bandwidth by employing a "sub-sampling and reconstruction" paradigm: the object interferograms are first spectrally or spatially sub-sampled and then OCT images are reconstructed by signal processing techniques.
In the method, a fixed sub-sampling mask is needed to reduce the data bandwidth, and the quality of the reconstructed OCT image is low because the fixed sub-sampling mask cannot be optimized together with the OCT image reconstruction.
For the problems in the related art, no effective solution has been proposed at present.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides the SS-OCT image compression method for the deep learning and spectrum domain airspace combined sub-sampling, which has the advantages of neural network combined synchronous optimization and the like, and solves the problems that a fixed sub-sampling mask cannot be optimized together with OCT image reconstruction, so that the reconstructed OCT image is low in quality.
In order to solve the technical problem that the quality of the reconstructed OCT image is low because the fixed sub-sampling mask cannot be optimized together with the OCT image reconstruction, the invention provides the following technical scheme:
an SS-OCT image compression method combining deep learning and spectral domain spatial domain joint sub-sampling comprises the following steps:
step one, constructing and training each neural network;
constructing a mask neural network:
the input of the mask neural network is random noise, two sets of feature images are output through a U-Net architecture of four layers of up-sampling and four layers of down-sampling, wherein each layer of up-sampling or down-sampling comprises two layers of convolution layers, and the two last sets of feature images generate a two-dimensional mask through GumbeSoftmax activation functions;
wherein GumbeSoftmax activation function is:
Figure SMS_1
wherein D and
Figure SMS_2
the output of the mask neural network and Gumbel Softmax activation function are respectively; random selection->
Figure SMS_3
As a two-dimensional mask; g is random noise of independent and identical samples in Gumbel (0, 1) distribution, τ is a value controlling Gumbel distribution density, p and c are pixel index and channel index, respectively, # and (x) are the values of the Gumbel distribution density>
Figure SMS_4
The output of the gummel Softmax activation function, represented as channel c, pixel p; />
Figure SMS_5
The output of the mask neural network is represented as channel c and pixel p;
Figure SMS_6
the output of the mask neural network is represented by the channel k and the pixel p; />
Figure SMS_7
Output represented by random noise under channel k and pixel p, < >>
Figure SMS_8
An output representing random noise with channel c and pixel p;
the method for constructing the stripe complement neural network comprises the following steps:
the input of the stripe complement neural network is undersampled stripe, and the complement sample stripe is obtained through the output of the U-Net architecture of four layers of downsampling and four layers of upsampling, wherein each layer of upsampling or downsampling comprises two layers of convolution layers;
the method for constructing the enhanced neural network comprises the following steps:
the input of the enhanced neural network is a preliminarily reconstructed image, and the enhanced OCT image is obtained through the output of a U-Net architecture of four layers of downsampling and four layers of upsampling, wherein each layer of upsampling or downsampling comprises two layers of convolution layers;
training the mask neural network, the stripe complement neural network and the enhancement neural network until convergence to obtain a trained mask neural network, a trained stripe complement neural network and a trained enhancement neural network;
the convergence time is the time when the similarity between the enhanced image and the true image reaches the standard;
the enhanced image is obtained by the following steps: inputting random noise into the mask neural network after the current iteration, and outputting and generating a two-dimensional mask; multiplying the two-dimensional mask with the corresponding point of the complete sample interference fringe to obtain an undersampled fringe; inputting the undersampled stripe into a stripe complement neural network after the current iteration to obtain a complement sample stripe; obtaining a preliminarily reconstructed image by carrying out IDFT on the completed sample stripes; inputting the preliminarily reconstructed image into an enhanced neural network after the current iteration, and outputting to obtain an enhanced image;
generating an image sampling system, wherein the image sampling system comprises a two-dimensional mask generated by outputting a mask neural network after training, a stripe complement neural network after training, an enhancement neural network after training and IDFT;
step three: and inputting the complete target detection interference fringes into the image sampling system to obtain a target image.
Preferably, the method for acquiring the complete sample interference fringes and the complete target detection interference fringes comprises the following steps: and acquiring the sample object and the object to be detected through the OCT system to obtain complete sample interference fringes and complete target detection interference fringes.
Preferably, the image truth value is obtained by: and obtaining a preliminary OCT image by the integral sample interference fringes and IDFT, and obtaining an image truth value by calculating the preliminary OCT image through a classical post-processing algorithm.
Preferably, the method of training the masking neural network, enhancing the neural network and stripe complement neural network comprises:
in the training framework, a batch size of 1 was used, an initial learning rate of 0.001, a network was trained for 30 cycles using an AdamW optimizer with momentum (0.9, 0.999), and then a cosine decay strategy was used to reduce the learning rate.
Preferably, the method for obtaining the preliminary reconstructed image by the IDFT of the completed sample fringes and the method for obtaining the preliminary OCT image by the IDFT of the complete sample interference fringes are respectively as follows: taking the completed sample stripes as input of IDFT (x) commands in matlab or python programs, and outputting the input after operation to obtain a preliminarily reconstructed image;
and taking the complete sample interference fringes as input of an IDFT (x) command in a matlab or python program, and obtaining a preliminary OCT image after operation.
Preferably, the method for calculating the image true value of the preliminary OCT image through a classical post-processing algorithm comprises the following steps:
the preliminary OCT image firstly takes a logarithmic function, and then obtains an image true value through thresholding and normalization.
Preferably, three groups of loss functions are constructed to evaluate the mask neural network after the current iteration, the enhancement neural network after the current iteration and the stripe completion neural network after the current iteration respectively, wherein the three groups of loss functions are specifically as follows:
Figure SMS_9
wherein
Figure SMS_10
Representing a two-dimensional mask>
Figure SMS_11
Represents +.>
Figure SMS_12
Line and->
Figure SMS_13
A column;
the above formula represents a statistical two-dimensional mask
Figure SMS_14
The number of the middle pixel points is 1;
Figure SMS_15
wherein
Figure SMS_16
Representing interference fringes, & lt & gt>
Figure SMS_17
Representing interference fringes->
Figure SMS_18
The%>
Figure SMS_19
Line and->
Figure SMS_20
A column;
Figure SMS_21
representing interference fringe truth value, +.>
Figure SMS_22
Interference fringe true value +.>
Figure SMS_23
Line and->
Figure SMS_24
A column;
the above formula is used to calculate the interference fringe after completion
Figure SMS_25
And interference fringe truth->
Figure SMS_26
Root mean square error of (a);
Figure SMS_27
wherein
Figure SMS_28
Representing an image after image enhancement, < >>
Figure SMS_29
Representing an enhanced image +.>
Figure SMS_30
The%>
Figure SMS_31
Line and->
Figure SMS_32
A column;
Figure SMS_33
representing image truth value->
Figure SMS_34
Representing the +.>
Figure SMS_35
Line and->
Figure SMS_36
A column;
the above formula is used to calculate an image after image enhancement
Figure SMS_37
And image truth->
Figure SMS_38
Average absolute error of (a);
wherein ,L mask film The data compression ratio is used for calculating the two-dimensional mask;
L stripe pattern The method is used for evaluating the effect of the stripe completion network after the current iteration;
L image processing apparatus The method is used for evaluating the effect of the image enhancement network after the current iteration;
and in the training stage of the mask neural network, the enhancement neural network and the stripe completion neural network, carrying out weighted summation on the three groups of loss functions, then counter-propagating gradients of the three groups of loss functions to carry out iterative solution on the mask neural network after the next iteration, the enhancement neural network after the next iteration and the stripe completion neural network after the next iteration until the similarity between the enhanced image and the true image is judged to reach the standard when the weighted summation value of the three groups of loss functions is no longer reduced, and obtaining a two-dimensional mask after the training is completed, the enhancement neural network after the training is completed and the stripe completion neural network after the training is completed, otherwise, continuing to train the mask neural network, the enhancement neural network and the stripe completion neural network.
The SS-OCT compression imaging system comprises a subsampled mask generating module, an interference fringe repairing module and an image enhancement module, wherein the subsampled mask generating module is used for sampling complete interference fringes acquired by the SS-OCT system to obtain undersampled interference fringes, and then the interference fringe repairing module and the image enhancement module are used for obtaining a target.
Compared with the prior art, the invention provides the SS-OCT compression imaging method for the deep learning and spectrum domain airspace combined sub-sampling, which has the following beneficial effects:
1. according to the invention, the mask neural network, the enhancement neural network and the stripe complement neural network are synchronously subjected to gradient descent update of parameters, so that the combined optimization of the two-dimensional mask and OCT image reconstruction is achieved, the problem of poor OCT image reconstruction effect caused by the adoption of an empirical fixed two-dimensional mask in the current OCT image sampling system is solved, and the OCT image reconstruction quality of the image sampling system is optimal under the same data compression ratio.
2. The invention counter propagates gradient by means of chained derivation, and simultaneously carries out gradient descent optimization updating on parameters of a mask neural network, an enhanced neural network and a stripe complement neural network, and the gradient descent optimization updating is carried out on the parameters of the mask neural network, the enhanced neural network and the stripe complement neural networkL Mask filmL Stripe pattern AndL image processing apparatus After the three groups of loss functions are weighted and summed and converged, the best balance between the image reconstruction quality and the data compression rate can be obtained, and the problems that fixed empirical sub-sampling masks, such as a center mask and a random mask, cannot be optimized together with OCT image reconstruction, and the reconstructed OCT image quality is low are solved.
Drawings
FIG. 1 is a flow chart of the image sampling system construction of the present invention;
FIG. 2 is a schematic diagram showing the effect of reconstructing human cardiac muscle according to the present invention;
FIG. 3 is a schematic diagram of the reconstruction effect of a human finger according to the present invention;
FIG. 4 is a schematic diagram showing the onion skin reconstruction effect of the present invention;
FIG. 5 is a schematic diagram of the reconstruction effect of the comparative random mask and center mask of the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
As described in the background art, the present application provides an SS-OCT image compression method combining deep learning and spectral domain spatial domain joint sub-sampling in order to solve the above technical problems.
Embodiment one:
an SS-OCT compression imaging method for deep learning and spectral domain spatial domain joint sub-sampling, comprising:
step one: acquiring complete sample interference fringes and complete target detection interference fringes;
the method for acquiring the complete sample interference fringes and the complete target detection interference fringes comprises the following steps: the OCT system is used for collecting a sample object and an object to be detected to obtain a complete sample interference fringe and a complete target detection interference fringe, wherein the OCT system is used for converting coherent light into an electric signal through a photoelectric detector in the OCT system, and the electric signal is used for sampling through a data acquisition card to obtain the interference fringe.
Preferably, the sample interference fringes are preferably obtained by a dataset imaged by a commercial OCT system (Thorlabs Ganymede, newton, NJ), the dataset containing 3784 OCT images from 17 samples and corresponding interference fringes.
Step two: referring to fig. 1, an image sampling system is constructed and generated;
the construction and generation method of the image sampling system comprises the following steps:
s1: constructing a mask neural network, a stripe complement neural network and an enhancement neural network;
the mask neural network construction method comprises the following steps: the input of the mask neural network is random noise, and two groups of characteristic diagrams are output through four layers of up-sampling and four layers of down-sampling U-Net architectures, wherein each layer of up-sampling or down-sampling U-Net architecture comprises two layers of convolution layers;
the construction method of the stripe complement neural network comprises the following steps: the input of the stripe complement neural network is an undersampled interference stripe, and the complement sample stripe is obtained through the output of a four-layer downsampling and four-layer upsampling U-Net framework, wherein each layer of upsampling or downsampling U-Net framework comprises two layers of convolution layers;
the construction method of the enhanced neural network comprises the following steps: the input of the enhanced neural network is a preliminarily reconstructed image, and the enhanced OCT image is output through a four-layer downsampling and four-layer upsampling U-Net framework, wherein each layer of upsampling or downsampling framework comprises two layers of convolution layers.
S2: training a mask neural network, a stripe complement neural network and an enhancement neural network to respectively obtain a trained mask neural network, a trained stripe complement neural network and a trained enhancement neural network;
the method for training the mask neural network, the enhancement neural network and the stripe complement neural network comprises the following steps:
in the training framework, a batch size of 1 was used, an initial learning rate of 0.001, a network was trained for 30 cycles using an AdamW optimizer with momentum (0.9, 0.999), and then a cosine decay strategy was used to reduce the learning rate.
The concrete operation of the cosine decay strategy for reducing the learning rate is as follows: using the cosineAnneanlingLR function in Torch, the learning rate drops as a cosine function.
S21: inputting random noise into the mask neural network after the current nth iteration, and outputting and generating a two-dimensional mask; n is an integer, and the minimum value is 0;
when n=0, the corresponding mask neural network, stripe complement neural network and enhancement neural network are respectively the mask neural network, stripe complement neural network and enhancement neural network constructed in the S1;
generating a two-dimensional mask from the two sets of feature maps in the step S1 through GumbeSoftmax activation functions, wherein the GumbeSoftmax activation functions are as follows:
Figure SMS_39
wherein D and
Figure SMS_40
the output of the mask neural network and gummel Softmax activation function, respectively. We randomly select +.>
Figure SMS_41
G is random noise of independent and identical samples in Gumbel (0, 1) distribution, τ is a value controlling Gumbel distribution density, p and c are pixel index and channel index, respectively,/">
Figure SMS_42
The output of the gummel Softmax activation function, represented as channel c, pixel p; />
Figure SMS_43
The output of the mask neural network is represented as channel c and pixel p; />
Figure SMS_44
The output of the mask neural network is represented by the channel k and the pixel p; />
Figure SMS_45
Output represented by random noise under channel k and pixel p, < >>
Figure SMS_46
Represented as the output of random noise with channel c and pixel p.
S22: and multiplying the two-dimensional mask with the corresponding point of the complete sample interference fringe to obtain an undersampled fringe.
S23: and inputting the undersampled stripe into a stripe complement neural network after the current nth iteration to obtain a complement sample stripe.
S24: performing Inverse Discrete Fourier Transform (IDFT) on the completed sample stripes to obtain a primarily reconstructed image;
the method for obtaining the preliminarily reconstructed image by the completed sample stripes through IDFT (inverse discrete Fourier transform) comprises the following steps: and taking the completed sample stripes as input of IDFT (x) commands in matlab or python programs, and obtaining a preliminarily reconstructed image after operation.
S25: and inputting the preliminarily reconstructed image into an enhanced neural network after the current nth iteration, and outputting the enhanced image.
S26: the complete sample interference fringes and the IDFT (inverse discrete Fourier transform) are used for obtaining a preliminary OCT image, and the preliminary OCT image is calculated through a classical post-processing algorithm to obtain an image truth value;
the method for obtaining the preliminary OCT image by the complete sample interference fringes through IDFT (inverse discrete Fourier transform) comprises the following steps of: taking the complete sample interference fringes as input of IDFT (x) commands in matlab or python programs, and obtaining a preliminary OCT image after operation;
the method for calculating the image truth value of the preliminary OCT image through the classical post-processing algorithm comprises the following steps:
the preliminary OCT image firstly takes a logarithmic function, and then obtains an image true value through thresholding and normalization, namely: the preliminary OCT image is taken as input of a log10 (X) function in Matlab, then the value (marked as X) output by the logarithmic function is input into the Matlab, and the true value of the image is output according to the (X-3)/3 function.
S27: judging whether the similarity between the enhanced image and the true image meets the standard, if so, training the mask neural network after the current nth iteration, the stripe complement neural network after the current nth iteration and the enhanced neural network after the current nth iteration to obtain a trained mask neural network, a trained stripe complement neural network and a trained enhanced neural network, and completing the construction of an image sampling system, and entering step S10;
if the parameter does not reach the standard, updating parameters of the mask neural network after the current nth iteration, the stripe complement neural network after the current nth iteration and the enhancement neural network after the current nth iteration to obtain the mask neural network after the current n+1th iteration, the stripe complement neural network after the current n+1th iteration and the enhancement neural network after the current n+1th iteration, and returning to the step S2;
the method for updating the parameters of the enhanced neural network after the current nth iteration comprises the following steps: random gradient drop, namely: the network weights and bias terms are iteratively updated by computing gradients of the loss function over small batches of data.
The method for judging whether the similarity between the enhanced image and the true image meets the standard comprises the following steps:
three groups of loss functions are constructed to evaluate the mask neural network after the current nth iteration, the enhancement neural network after the current nth iteration and the stripe completion neural network after the current nth iteration respectively, wherein the three groups of loss functions are specifically as follows:
Figure SMS_47
wherein
Figure SMS_48
Representing a two-dimensional mask>
Figure SMS_49
Represents +.>
Figure SMS_50
Line and->
Figure SMS_51
A column;
the above formula represents a statistical two-dimensional mask
Figure SMS_52
The number of the middle pixel points is 1;
Figure SMS_53
wherein
Figure SMS_54
Representing interference fringes, & lt & gt>
Figure SMS_55
Representing interference fringes->
Figure SMS_56
The%>
Figure SMS_57
Line and->
Figure SMS_58
A column;
Figure SMS_59
representing interference fringe truth value, +.>
Figure SMS_60
Interference fringe true value +.>
Figure SMS_61
Line and->
Figure SMS_62
A column;
the above formula is used to calculate the interference fringe after completion
Figure SMS_63
And interference fringe truth->
Figure SMS_64
Root mean square error of (a);
Figure SMS_65
wherein
Figure SMS_66
Representing an image after image enhancement, < >>
Figure SMS_67
Representing an enhanced image +.>
Figure SMS_68
The%>
Figure SMS_69
Line and->
Figure SMS_70
A column;
Figure SMS_71
representing image truth value->
Figure SMS_72
Representing the +.>
Figure SMS_73
Line and->
Figure SMS_74
A column;
the above formula is used to calculate an image after image enhancement
Figure SMS_75
And image truth->
Figure SMS_76
Average absolute error of (a);
wherein ,L mask film The data compression ratio is used for calculating the two-dimensional mask;
L stripe pattern The method is used for evaluating the effect of the stripe completion network after the current nth iteration;
L image processing apparatus The method comprises the steps of evaluating the effect of an image enhancement network after the current nth iteration;
and in the training stage of the mask neural network, the enhancement neural network and the stripe completion neural network, carrying out weighted summation on the three groups of loss functions, and then counter-propagating gradients of the three groups of loss functions to carry out iterative solution on the mask neural network after the n+1st iteration, the enhancement neural network after the n+1st iteration and the stripe completion neural network after the n+1st iteration until the similarity of the enhanced image and the image true value reaches the standard when the weighted summation value of the three groups of loss functions is not reduced any more, so as to obtain a two-dimensional mask after the training is completed, the enhancement neural network after the training is completed and the stripe completion neural network after the training is completed, otherwise, continuing training the mask neural network, the enhancement neural network and the stripe completion neural network. The calculation mode and structure of the loss function are the prior art, and the invention only performs transfer.
S3: an image sampling system is generated that includes the current two-dimensional mask, the trained stripe completion neural network, the trained enhancement neural network, and the IDFT.
Step three: and inputting the complete target detection interference fringes into the image sampling system to obtain a target image.
The invention counter propagates gradient by means of chained derivation, and simultaneously carries out gradient descent optimization updating on parameters of a mask neural network, an enhanced neural network and a stripe complement neural network, and the gradient descent optimization updating is carried out on the parameters of the mask neural network, the enhanced neural network and the stripe complement neural networkL Mask filmL Stripe pattern AndL image processing apparatus After the three groups of loss functions are weighted and summed and converged, the best balance between the image reconstruction quality and the data compression rate can be obtained, and the problems that fixed empirical sub-sampling masks, such as a center mask and a random mask, cannot be optimized together with OCT image reconstruction, and the reconstructed OCT image quality is low are solved.
Embodiment two:
referring to fig. 2-5, an SS-OCT compression imaging system with deep learning and spectrum domain spatial domain combined sub-sampling uses the SS-OCT image compression method, which includes a sub-sampling mask generation module, an interference fringe repairing module and an image enhancement module, and the sub-sampling mask module samples the complete interference fringe acquired by the SS-OCT system to obtain an undersampled fringe, and then the interference fringe repairing module and the image enhancement module obtain a target image.
The sub-sampling mask generation module generates a sub-sampling mask through input noise, and the sub-sampling mask samples data acquired by the SS-OCT system to form undersampled stripes; then, repairing the undersampled stripe by an interference stripe repairing module to obtain a completed sample stripe; and the image enhancement module performs image enhancement operation on the completed sample stripes to obtain a target image.
Further, the sub-sampling mask generation module comprises a U-Net architecture with 1 channel input and 2 channels output and a Gumbel Softmax activation function. The U-Net architecture comprises a four-layer downsampling module, a four-layer upsampling module and a jump connection. The Gumbel Softmax activation function encodes the output of the U-Net architecture, generating a (0, 1) sub-sampling mask.
The interference fringe repairing module comprises: the input is 1 channel, the output is 1 channel U-Net architecture and IDFT (inverse discrete Fourier transform). The U-Net architecture comprises a four-layer downsampling module, a four-layer upsampling module and a jump connection. The undersampled streaks are first complemented by the U-Net architecture to obtain complemented sample streaks, and then the complemented sample streaks are converted into a preliminarily reconstructed image by IDFT.
The image enhancement module comprises: the input is a U-Net architecture with 2 channels and 1 channel output. And carrying out image enhancement on the preliminarily reconstructed image through a U-Net architecture to obtain an enhanced image (namely, a target image).
The image compression system adopts an L1 norm of a difference between a calculated target image and a true value image, an L2 norm of a difference between a repair stripe (namely a completed sample stripe) and an original stripe (namely a complete sample interference stripe) and a data compression ratio (namely a total 1 ratio in a formed two-dimensional mask), and specifically comprises the following steps:
Figure SMS_77
wherein
Figure SMS_80
For the total loss function +.>
Figure SMS_81
An L1 norm that is the difference between the target image and the truth image;
Figure SMS_83
is->
Figure SMS_78
Weights of (2); />
Figure SMS_82
An L2 norm for repairing the difference between the stripe and the original stripe; />
Figure SMS_84
Is that
Figure SMS_85
Weights of (2); />
Figure SMS_79
Is the total number of 1's in the two-dimensional mask;
further, a sub-sampling mask generating module is adopted to sample the complete stripe acquired by the SS-OCT through a sub-sampling mask generated by input noise, and then an enhanced image (namely a target image) is obtained through a stripe complementing module and an image enhancing module, specifically:
a1, for the complete sample interference fringes acquired by SS-OCT, adjusting the interference fringes to an image with 2048 x 992 resolution through zero padding or clipping;
a2, inputting noise into a generator in the sub-sampling mask generation module, and generating a sub-sampling mask (namely a two-dimensional mask) with only 0 and 1 through a Gumbel Softmax activation function;
a3, multiplying the sub-sampling mask generated in the step a2 with the corresponding points of the complete sample interference fringes to obtain undersampled fringes;
a4, obtaining a completed sample stripe by a stripe repair module for the undersampled stripe obtained in the step a 3;
a5, obtaining a preliminarily reconstructed image (namely an preliminarily reconstructed OCT image) of the completed sample stripe obtained in the step a4 through IDFT operation;
a6, obtaining an enhanced image (namely a target image) of the primarily reconstructed image obtained in the step a5 through an image enhancement module;
the subsampling mask generating module, the interference fringe repairing module and the image enhancement module are all U-Net architecture networks, all the architecture networks are provided with four downsampling (and corresponding upsampling) blocks, and the initial characteristics of the input layer are set to be 32 characteristics. The downsampling block uses a leaklylrelu (negative slope 0.2) activation function and the upsampling block uses a ReLU activation function. All blocks were normalized using a batch, with two convolutional layers for each block. The sub-sampling mask generation module has 1 input channel and 2 output channels. The interference fringe repairing module is provided with 1 input channel and 1 output channel. The image enhancement module has 2 input channels and 1 output channel.
In the training framework, the batch size used was 1, the initial learning rate was 0.001, the network was trained for 30 cycles using an AdamW optimizer with momentum (0.9,0.999), and then the cosine decay strategy was used to reduce the learning rate. All experiments were trained and tested using NVIDIA GeForce RTX 3090 GPU card.
Further, two human coronary arteries, two human fingers and one onion collected by the same device were used as test sets to further illustrate the generalization of the proposed technique.
Case one:
the performance of the present invention was evaluated using peak signal to noise ratio (PSNR) and Structural Similarity Index (SSIM), and experiments were performed on human coronary arteries, human fingers and onions, respectively, and the experimental results are shown in table 1.
TABLE 1
Figure SMS_86
Obviously, the image quality steadily decreases with the increase of the compression rate; the same trend was observed when the test set was deviated from the coronary artery sample to the finger and onion. However, even if a large amount of original data is lost, the original image can be restored; when only 1.6% of the data is used, a PSNR of 24.2 dB can be achieved. Three exemplary reconstructed OCT images (one coronary artery, one finger, and one onion) obtained by using 1.6%, 3.6%, 10%, 25%, and 50% of the raw data and the corresponding truth images are shown in fig. 2, fig. 3, fig. 4, respectively. Most texture information is well preserved. Wherein DCR is the data compression rate.
Case two:
in this case, compared to case two, a comparison study is performed on the human coronary dataset by replacing the learnable mask with other empirically fixed mask patterns, such as random spectrum sub-sampling (referred to as "random mask") and center truncated spectrum (referred to as "center mask"). It should be noted that the entire network is retrained and trimmed for each different mask pattern. Fig. 5 shows the corresponding results for different DCRs.
The proposed learner mask performs best in all mask patterns for all different DCRs. In particular for very high DCR (1%), only the proposed method is able to reconstruct images with a PSNR exceeding 21 dB. In contrast, the center mask performs relatively well at low DCR (50%), with the corresponding PSNR almost 1.8 dB higher than the random mask, 2.2 dB lower than the present invention. However, as DCR increases (e.g., 1% and 5%), its performance begins to perform poorly, streak artifacts appear; on the other hand, the reconstruction result obtained by using the random mask is only degraded with the increase of DCR.
Although embodiments of the present invention have been shown and described, it will be understood by those skilled in the art that various changes, modifications, substitutions and alterations can be made hereto without departing from the spirit and principles of the present invention.

Claims (8)

1. The SS-OCT image compression method for the deep learning and spectral domain airspace combined sub-sampling is characterized by comprising the following steps of:
the method comprises the following steps:
step one, constructing and training each neural network;
constructing a mask neural network:
the input of the mask neural network is random noise, two sets of feature graphs are output through a U-Net architecture of four layers of up-sampling and four layers of down-sampling, wherein each layer of up-sampling or down-sampling comprises two layers of convolution layers, and the two last sets of feature graphs generate a two-dimensional mask through Gumbel Softmax activation functions;
wherein Gumbel Softmax activation function is:
Figure QLYQS_1
wherein D and
Figure QLYQS_2
the output of the mask neural network and Gumbel Softmax activation function are respectively; random selection->
Figure QLYQS_3
As a two-dimensional mask; g is random noise of independent and identical samples in Gumbel (0, 1) distribution, τ is a value controlling Gumbel distribution density, p and c are pixel index and channel index, respectively, # and (x) are the values of the Gumbel distribution density>
Figure QLYQS_4
The output of the gummel Softmax activation function, represented as channel c, pixel p; />
Figure QLYQS_5
The output of the mask neural network is represented as channel c and pixel p;
Figure QLYQS_6
the output of the mask neural network is represented by the channel k and the pixel p; />
Figure QLYQS_7
Represented by random noise under channel k and pixel pOutput (I)>
Figure QLYQS_8
An output representing random noise with channel c and pixel p;
the method for constructing the stripe complement neural network comprises the following steps:
the input of the stripe complement neural network is undersampled stripe, and the complement sample stripe is obtained through the output of the U-Net architecture of four layers of downsampling and four layers of upsampling, wherein each layer of upsampling or downsampling comprises two layers of convolution layers;
the method for constructing the enhanced neural network comprises the following steps:
the input of the enhanced neural network is a preliminarily reconstructed image, and the enhanced OCT image is obtained through the output of a U-Net architecture of four layers of downsampling and four layers of upsampling, wherein each layer of upsampling or downsampling comprises two layers of convolution layers;
training the mask neural network, the stripe complement neural network and the enhancement neural network until convergence to obtain a trained mask neural network, a trained stripe complement neural network and a trained enhancement neural network;
the convergence time is the time when the similarity between the enhanced image and the true image reaches the standard;
the enhanced image is obtained by the following steps: inputting random noise into the mask neural network after the current iteration, and outputting and generating a two-dimensional mask; multiplying the two-dimensional mask with the corresponding point of the complete sample interference fringe to obtain an undersampled fringe; inputting the undersampled stripe into a stripe complement neural network after the current iteration to obtain a complement sample stripe; obtaining a preliminarily reconstructed image by carrying out IDFT on the completed sample stripes; inputting the preliminarily reconstructed image into an enhanced neural network after the current iteration, and outputting to obtain an enhanced image;
generating an image sampling system, wherein the image sampling system comprises a two-dimensional mask generated by outputting a mask neural network after training, a stripe complement neural network after training, an enhancement neural network after training and IDFT;
step three: and inputting the complete target detection interference fringes into the image sampling system to obtain a target image.
2. The SS-OCT image compression method of joint sub-sampling of deep learning and spectral domain spatial domain of claim 1, wherein: the method for acquiring the complete sample interference fringes and the complete target detection interference fringes comprises the following steps: and acquiring the sample object and the object to be detected through the OCT system to obtain complete sample interference fringes and complete target detection interference fringes.
3. The SS-OCT image compression method of joint sub-sampling of deep learning and spectral domain spatial domain of claim 1, wherein: the image truth value is obtained by the following steps: and obtaining a preliminary OCT image by the integral sample interference fringes and IDFT, and obtaining an image truth value by calculating the preliminary OCT image through a classical post-processing algorithm.
4. The SS-OCT image compression method of joint sub-sampling of deep learning and spectral domain spatial domain of claim 1, wherein: the method for training the mask neural network, the enhanced neural network and the stripe complement neural network comprises the following steps:
in the training framework, a batch size of 1 was used, an initial learning rate of 0.001, a network was trained for 30 cycles using an AdamW optimizer with momentum (0.9, 0.999), and then a cosine decay strategy was used to reduce the learning rate.
5. The SS-OCT image compression method of joint sub-sampling of deep learning and spectral domain spatial domain of claim 1, wherein: the method for obtaining the preliminarily reconstructed image by the completed sample fringes through IDFT and the method for obtaining the preliminarily OCT image by the complete sample interference fringes through IDFT are respectively as follows: taking the completed sample stripes as input of IDFT (x) commands in matlab or python programs, and outputting the input after operation to obtain a preliminarily reconstructed image;
and taking the complete sample interference fringes as input of an IDFT (x) command in a matlab or python program, and obtaining a preliminary OCT image after operation.
6. The SS-OCT image compression method of joint sub-sampling of deep learning and spectral domain spatial domain of claim 1, wherein: the method for calculating the image truth value of the preliminary OCT image through the classical post-processing algorithm comprises the following steps:
the preliminary OCT image firstly takes a logarithmic function, and then obtains an image true value through thresholding and normalization.
7. The SS-OCT image compression method of claim 6, wherein the SS-OCT image is compressed using a combination of deep learning and spectral domain spatial sub-sampling, wherein: three groups of loss functions are constructed to evaluate the mask neural network after the current iteration, the enhancement neural network after the current iteration and the stripe completion neural network after the current iteration respectively, wherein the three groups of loss functions are specifically as follows:
Figure QLYQS_9
wherein
Figure QLYQS_10
Representing a two-dimensional mask>
Figure QLYQS_11
Representing a two-dimensional mask->
Figure QLYQS_12
The%>
Figure QLYQS_13
Line and->
Figure QLYQS_14
A column;
the above formula represents a statistical two-dimensional mask
Figure QLYQS_15
The number of the middle pixel points is 1;
Figure QLYQS_16
wherein
Figure QLYQS_17
Representing interference fringes, & lt & gt>
Figure QLYQS_18
Representing interference fringes->
Figure QLYQS_19
The%>
Figure QLYQS_20
Line and->
Figure QLYQS_21
A column;
Figure QLYQS_22
representing interference fringe truth value, +.>
Figure QLYQS_23
Interference fringe true value +.>
Figure QLYQS_24
Line and->
Figure QLYQS_25
A column;
the above formula is used to calculate the interference fringe after completion
Figure QLYQS_26
And interference fringe truth->
Figure QLYQS_27
Root mean square error of (a);
Figure QLYQS_28
wherein
Figure QLYQS_29
Representing an image after image enhancement, < >>
Figure QLYQS_30
Representing an enhanced image +.>
Figure QLYQS_31
The%>
Figure QLYQS_32
Line and->
Figure QLYQS_33
A column;
Figure QLYQS_34
representing image truth value->
Figure QLYQS_35
Representing the +.>
Figure QLYQS_36
Line and->
Figure QLYQS_37
A column;
the above formula is used to calculate an image after image enhancement
Figure QLYQS_38
And image truth->
Figure QLYQS_39
Average absolute error of (a);
wherein ,L mask film The data compression ratio is used for calculating the two-dimensional mask;
L stripe pattern The method is used for evaluating the effect of the stripe completion network after the current iteration;
L image processing apparatus The method is used for evaluating the effect of the image enhancement network after the current iteration;
and in the training stage of the mask neural network, the enhancement neural network and the stripe completion neural network, carrying out weighted summation on the three groups of loss functions, then counter-propagating gradients of the three groups of loss functions to carry out iterative solution on the mask neural network after the next iteration, the enhancement neural network after the next iteration and the stripe completion neural network after the next iteration until the similarity between the enhanced image and the true image is judged to reach the standard when the weighted summation value of the three groups of loss functions is no longer reduced, and obtaining a two-dimensional mask after the training is completed, the enhancement neural network after the training is completed and the stripe completion neural network after the training is completed, otherwise, continuing to train the mask neural network, the enhancement neural network and the stripe completion neural network.
8. An SS-OCT compression imaging system with deep learning and spectral domain spatial domain joint sub-sampling, using the SS-OCT image compression method of any one of claims 1-7, characterized in that: the method comprises a sub-sampling mask generation module, an interference fringe repair module and an image enhancement module, wherein the sub-sampling mask module is used for sampling complete interference fringes acquired by an SS-OCT system to obtain undersampled fringes, and then the interference fringe repair module and the image enhancement module are used for obtaining a target image.
CN202310264537.8A 2023-03-20 2023-03-20 SS-OCT compression imaging method for deep learning and spectral domain airspace combined sub-sampling Active CN115984406B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202310264537.8A CN115984406B (en) 2023-03-20 2023-03-20 SS-OCT compression imaging method for deep learning and spectral domain airspace combined sub-sampling
PCT/CN2024/088631 WO2024193719A1 (en) 2023-03-20 2024-04-18 Ss-oct compressive imaging method based on deep learning and spectral domain and spatial domain joint sub-sampling

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310264537.8A CN115984406B (en) 2023-03-20 2023-03-20 SS-OCT compression imaging method for deep learning and spectral domain airspace combined sub-sampling

Publications (2)

Publication Number Publication Date
CN115984406A CN115984406A (en) 2023-04-18
CN115984406B true CN115984406B (en) 2023-06-20

Family

ID=85963478

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310264537.8A Active CN115984406B (en) 2023-03-20 2023-03-20 SS-OCT compression imaging method for deep learning and spectral domain airspace combined sub-sampling

Country Status (2)

Country Link
CN (1) CN115984406B (en)
WO (1) WO2024193719A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115984406B (en) * 2023-03-20 2023-06-20 始终(无锡)医疗科技有限公司 SS-OCT compression imaging method for deep learning and spectral domain airspace combined sub-sampling

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113674376A (en) * 2021-08-23 2021-11-19 合肥工业大学 Ultrasonic plane wave composite imaging method based on deep learning
CN114331911A (en) * 2022-01-05 2022-04-12 北京理工大学重庆创新中心 Fourier laminated microscopic image denoising method based on convolutional neural network
CN114820849A (en) * 2022-04-19 2022-07-29 浙江大学 Magnetic resonance CEST image reconstruction method, device and equipment based on deep learning

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4389032B2 (en) * 2007-01-18 2009-12-24 国立大学法人 筑波大学 Optical coherence tomography image processing device
US20210369195A1 (en) * 2018-04-26 2021-12-02 Voxeleron, LLC Method and system for disease analysis and interpretation
CN109829880A (en) * 2018-12-07 2019-05-31 清影医疗科技(深圳)有限公司 A kind of CT image detecting method based on deep learning, device and control equipment
US20220058803A1 (en) * 2019-02-14 2022-02-24 Carl Zeiss Meditec Ag System for oct image translation, ophthalmic image denoising, and neural network therefor
JP2023523246A (en) * 2020-04-29 2023-06-02 ノバルティス アーゲー Computer-implemented system and method for assessing disease or condition activity level in a patient's eye
WO2021220008A1 (en) * 2020-04-29 2021-11-04 Deep Render Ltd Image compression and decoding, video compression and decoding: methods and systems
WO2022008986A1 (en) * 2020-07-09 2022-01-13 Nidek Technologies S.R.L. A computer-implemented method for executing semantic segmentation of rgb digital images
CN111798535B (en) * 2020-09-09 2020-12-01 南京安科医疗科技有限公司 CT image enhancement display method and computer readable storage medium
CN113658115B (en) * 2021-07-30 2023-02-14 华南理工大学 Image anomaly detection method based on depth convolution generation countermeasure network
CN114723840A (en) * 2022-03-09 2022-07-08 天津大学 Slice-adaptive-oriented active undersampling method for magnetic resonance imaging
CN115797231A (en) * 2022-12-05 2023-03-14 上海交通大学 Real-time hologram generation method based on neural network of Fourier inspiration
CN115984406B (en) * 2023-03-20 2023-06-20 始终(无锡)医疗科技有限公司 SS-OCT compression imaging method for deep learning and spectral domain airspace combined sub-sampling

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113674376A (en) * 2021-08-23 2021-11-19 合肥工业大学 Ultrasonic plane wave composite imaging method based on deep learning
CN114331911A (en) * 2022-01-05 2022-04-12 北京理工大学重庆创新中心 Fourier laminated microscopic image denoising method based on convolutional neural network
CN114820849A (en) * 2022-04-19 2022-07-29 浙江大学 Magnetic resonance CEST image reconstruction method, device and equipment based on deep learning

Also Published As

Publication number Publication date
CN115984406A (en) 2023-04-18
WO2024193719A1 (en) 2024-09-26

Similar Documents

Publication Publication Date Title
CN109035142B (en) Satellite image super-resolution method combining countermeasure network with aerial image prior
CN111127320B (en) Deep learning-based photoacoustic image super-resolution reconstruction method and device
CN113344799A (en) System and method for reducing colored noise in medical images using deep neural networks
Deora et al. Structure preserving compressive sensing MRI reconstruction using generative adversarial networks
CN110880196B (en) Tumor photoacoustic image rapid reconstruction method and device based on deep learning
CN113376600B (en) Pedestrian radar echo denoising method based on RSDNet
CN114648048B (en) Electrocardiosignal noise reduction method based on variational self-coding and PixelCNN model
CN115984406B (en) SS-OCT compression imaging method for deep learning and spectral domain airspace combined sub-sampling
CN116563110A (en) Blind image super-resolution reconstruction method based on Bicubic downsampling image space alignment
CN115496659A (en) Three-dimensional CT image reconstruction method and device based on single projection data
CN106023116B (en) A kind of compressed sensing image rebuilding method and device based on block Weighted Constraint
CN115375785A (en) Magnetic resonance image reconstruction method and device based on artificial neural network
CN117094897B (en) Super-resolution phase gradient estimation method for phase contrast optical coherence elastography
CN111815692A (en) Method, system and storage medium for generating artifact-free data and artifact-containing data
CN114414050B (en) Self-adaptive Fourier calculation correlation imaging method and system
CN117115452A (en) Controllable medical ultrasonic image denoising method, system and computer storage medium
CN112634385B (en) Rapid magnetic resonance imaging method based on deep Laplace network
CN117974468A (en) Multi-mode medical image fusion method for global and local feature interaction parallelism
CN113744132B (en) MR image depth network super-resolution method based on multiple optimization
CN116630154A (en) Deconvolution super-resolution reconstruction method and device for optical coherence tomography
US20230135995A1 (en) Multi-slice mri data processing using deep learning techniques
CN115526792A (en) Point spread function prior-based coding imaging reconstruction method
He et al. Learning priors in high-frequency domain for inverse imaging reconstruction
CN115100069A (en) Ultrasonic image reconstruction method, system, device and medium
CN118014016B (en) OCT image reconstruction method based on antagonism convolution neural network and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant