CN116327103B - Large-visual-angle laryngoscope based on deep learning - Google Patents
Large-visual-angle laryngoscope based on deep learning Download PDFInfo
- Publication number
- CN116327103B CN116327103B CN202310618436.6A CN202310618436A CN116327103B CN 116327103 B CN116327103 B CN 116327103B CN 202310618436 A CN202310618436 A CN 202310618436A CN 116327103 B CN116327103 B CN 116327103B
- Authority
- CN
- China
- Prior art keywords
- image
- micro
- training
- images
- lens imaging
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000013135 deep learning Methods 0.000 title claims abstract description 17
- 238000003384 imaging method Methods 0.000 claims abstract description 42
- 230000005540 biological transmission Effects 0.000 claims abstract description 30
- 230000004927 fusion Effects 0.000 claims abstract description 24
- 238000013528 artificial neural network Methods 0.000 claims abstract description 23
- 230000000007 visual effect Effects 0.000 claims abstract description 13
- 238000000034 method Methods 0.000 claims description 21
- 230000003042 antagnostic effect Effects 0.000 claims description 10
- 230000006870 function Effects 0.000 claims description 9
- 230000008569 process Effects 0.000 claims description 7
- 239000011159 matrix material Substances 0.000 claims description 3
- 230000003068 static effect Effects 0.000 claims description 3
- 230000009466 transformation Effects 0.000 claims description 3
- 238000005516 engineering process Methods 0.000 abstract description 7
- 238000001514 detection method Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 2
- 238000003062 neural network model Methods 0.000 description 2
- 230000036961 partial effect Effects 0.000 description 2
- 230000002829 reductive effect Effects 0.000 description 2
- 238000002627 tracheal intubation Methods 0.000 description 2
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 1
- 210000003484 anatomy Anatomy 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 230000004907 flux Effects 0.000 description 1
- 239000003365 glass fiber Substances 0.000 description 1
- 210000004704 glottis Anatomy 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000000338 in vitro Methods 0.000 description 1
- 238000001727 in vivo Methods 0.000 description 1
- 238000002576 laryngoscopy Methods 0.000 description 1
- 230000000670 limiting effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 239000005304 optical glass Substances 0.000 description 1
- 238000007500 overflow downdraw method Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/267—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor for the respiratory tract, e.g. laryngoscopes, bronchoscopes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00004—Operational features of endoscopes characterised by electronic signal processing
- A61B1/00009—Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00163—Optical arrangements
- A61B1/00174—Optical arrangements characterised by the viewing angles
- A61B1/00183—Optical arrangements characterised by the viewing angles for variable viewing angles
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/04—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
- A61B1/05—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances characterised by the image sensor, e.g. camera, being in the distal end portion
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Surgery (AREA)
- Engineering & Computer Science (AREA)
- Radiology & Medical Imaging (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Optics & Photonics (AREA)
- Pathology (AREA)
- Veterinary Medicine (AREA)
- Public Health (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Animal Behavior & Ethology (AREA)
- Otolaryngology (AREA)
- Physiology (AREA)
- Pulmonology (AREA)
- Signal Processing (AREA)
- Endoscopes (AREA)
Abstract
The invention discloses a large-view laryngoscope based on deep learning, which comprises a micro-lens imaging component, a micro-lens imaging component and a camera, wherein the micro-lens imaging component is used for acquiring image data; the micro-lens imaging assembly comprises at least two groups of small-caliber micro-lenses with non-overlapping visual fields; the flexible image transmission assembly is connected with the micro-lens imaging assembly and used for transmitting image data; the image processing module is connected with the other end of the flexible image transmission component, and is used for training the anti-neural network according to the image data obtained by the micro-lens imaging component to obtain an image fusion model, and the image fusion model is used for re-splicing the training output result into a complete image with gaps between non-overlapping view field images eliminated; the display assembly is used for displaying the complete image. The technical scheme of the invention is based on the flexible image transmission and artificial intelligent image fusion technology, and the capability of carrying out high-resolution imaging on a scene with an ultra-large visual angle in single shooting is realized, and the ultra-large visual field image reconstructed by an intelligent image fusion algorithm is observed through a display screen.
Description
Technical Field
The invention relates to the technical field of laryngoscopes, in particular to a large-viewing-angle laryngoscope based on deep learning.
Background
Examination of the throat is a major problem in the field of medical diagnostics. Because of the deep and complex structure of the throat, direct visual detection is generally difficult, and high dynamic, high spatial resolution imaging and viewing is generally required with video detection equipment. Video laryngoscopes are a widely used representative of a class of laryngeal examination devices. For the difficult tracheal intubation process, the video laryngoscope collects laryngeal images through the image sensor and displays the laryngeal images on a display screen in real time through the video transmission line. On one hand, the visual field range of human eyes in a common laryngoscope is exceeded, and on the other hand, the lens of the video laryngoscope does not need to move the laryngeal structure, so that the pain on a patient is lighter, and meanwhile, the difficulty of tracheal intubation is reduced because the lens accords with the anatomical structure of the human laryngeal, and the device has the advantages of being fast in operation, convenient to collect, high in success rate, improved in glottis exposure condition and the like.
The imaging module in the common video laryngoscope structure is generally carried out in a cooperative manner of a camera handle and a host, namely, the camera handle comprises a set of lenses and an image sensor unit for high-speed imaging of a field of view corresponding to the lenses, and a light source for illumination in a darker environment in the throat. In the method, the observation effect of the video laryngoscope mainly depends on factors such as the size of the field of view of the adopted lens and sensor, the spatial resolution, the imaging signal to noise ratio and the like, and meanwhile, the method has certain operation limitation. On the one hand, since the angle of view of the camera system is fixed, when the field of view needs to be changed or moved, the camera system needs to be mechanically moved to find a desired structure. This increases both the difficulty of operation and the time cost of imaging. On the other hand, due to the thin throat structure, the adopted lens and image sensor are generally small in size, low in spatial resolution and incapable of providing higher imaging precision for throat observation with finer dimensions. For this reason, a method of implementing endoscopic observation by using a plurality of lenses has been proposed, but a conventional image fusion method needs to reserve a part of overlapping area between adjacent fields of view, and cannot maximize the angle of the imaging field of view.
Therefore, the current common video laryngoscope technology only adopts the traditional imaging technology, and the endoscopic method realized by combining a plurality of lenses does not use the advantages of the artificial intelligent method such as deep learning and the like in the aspects of learning and generating new features, so that the method has the advantages of small volume, high real-time performance, convenient operation and the like, and meanwhile, mechanical movement is not required in single shooting, and large-scale laryngeal images with ultra-large visual angles and high spatial resolution are acquired to the maximum extent.
Disclosure of Invention
The invention aims to provide a deep learning-based large-view laryngoscope, which automatically guides partial images shot by a micro-lens sequence into a computer and performs fusion display, does not need any expensive additional equipment, and simultaneously maintains all the inherent advantages of a general video laryngoscope.
The invention aims to achieve the aim, and is specifically realized by the following technical scheme:
a deep learning-based large viewing angle laryngoscope comprising:
the micro-lens imaging assembly is used for acquiring image data; the micro-lens imaging assembly comprises at least two groups of small-caliber micro-lenses with non-overlapping visual fields;
the flexible image transmission assembly is connected with the micro-lens imaging assembly and used for transmitting image data; the flexible image transmission beam is a flexible image transmission beam with the diameter of the outer ring of the tail end smaller than 1.6mm and the number of 1-6;
the image processing module is connected with the other end of the flexible image transmission component, and is used for training the anti-neural network according to the image data obtained by the micro-lens imaging component to obtain an image fusion model, and the image fusion model is used for re-splicing the training output result into a complete image with gaps between non-overlapping view field images eliminated;
the display component is connected with the image processing module and is used for displaying the complete image re-spliced by the image processing module;
the generating antagonistic neural network comprises a generator G and a discriminator D, and the training process is as follows: s1, acquiring training data; s2, building a generator neural network and a discriminator neural network; and S3, performing countermeasure training by the generator and the discriminator to obtain an image fusion model.
Further, in step S1, the step of obtaining the training data includes:
s11, aiming at throat static samples under the same conditions, a micro-lens imaging assembly moves along a lens distribution plane, at least 10 series of images containing gaps are continuously collected, and the overlapping area of adjacent series of images is not less than 50% of the image area of Yu Shanzhang;
s12, cutting a series of images containing gaps into 256 multiplied by 256 pixel image small blocks without overlapping by taking one series of images as a reference, adopting scale-invariant feature transformation for each image small block, respectively carrying out feature matching with other images in the series of images, and acquiring an image small block area matched with each image small block;
s13, analyzing the image small block and the image small block area matched with the image small block: discarding if both contain gaps; if no gap exists, any one of the two is used as the input of the condition x for generating the antagonistic neural network, and the other is used as the true value y; otherwise, inputting the image small blocks containing the gaps as a condition x, and inputting the image small blocks without the gaps as a true value y; thus obtaining pairs of image block inputs as training data.
Further, the criterion for judging whether the small image block contains a gap is as follows: an image patch having more than 50 points with luminance values of 0 is a slit-containing image; otherwise, judging the image to be the image without the gap.
Further, in step S2, the generator G adopts a U-shaped symmetric network structure with an encoder/decoder structure, the size of each layer of feature map of the encoder portion is halved, and the size of each layer of the decoder is doubled; the number of filters per layer is doubled layer by layer in the encoder section and halved in the decoder section.
Further, the number of first layer filters of the generator G is 64, and each of the encoder and the decoder has 7 layers.
Further, in step S2, the arbiter D adopts a full convolution form, and has 4 layers in total, and the size of each layer of feature images is halved, the number is doubled, and finally, the matrix is averaged, and the obtained value represents a probability judgment on whether the image result is a true value.
Further, in the training process of step S3, for the generator G and the corresponding discriminator D, the loss function against the neural network is expressed as:
(1)
wherein the method comprises the steps ofAnd->Representing the mean value under the whole data, wherein z is the noise input of the generator G, x is the condition input of the countermeasure network, and y is the true value input of the countermeasure network training;
the loss function expression for the L1 norm error between the predicted value and the true value is:
(2)
the objective function formula of the countermeasure network training is as follows:
(3)
in the formula (3), λ is the weight of two losses, and λ is set to 100.
According to the technical scheme, based on a flexible image transmission and artificial intelligent image fusion technology, an external image sensor is connected to the tail end of a flexible image guide device to relay partial images acquired by a micro-lens sequence, conventional laryngoscopy is performed, the capability of high-resolution imaging is performed on a scene with an ultra-large visual angle in single shooting, and the ultra-large visual field image reconstructed through an intelligent image fusion algorithm is observed through a display screen, so that the detection range and imaging precision of a video laryngoscope are improved, and the method has important significance in the field of video laryngoscope diagnosis.
Advantages of the present technology also include: (1) The technology only needs to refit the traditional video laryngoscope with low cost, and keeps the advantages of convenient operation, light weight and the like;
(2) On one hand, the throat image obtained by the technology has obviously improved visual field range and spatial resolution, and on the other hand, the throat image with multiple visual angles can be synchronously recorded in different areas due to the adoption of the in-vitro image sensor with high spatial resolution and pixel number, so that the method is suitable for real-time video imaging; on the other hand, unlike the traditional image fusion, which needs to reserve the overlapping area between adjacent fields of view, the method is processed by a rapid deep learning algorithm without reserving the overlapping area, thereby further effectively improving the imaging flux of the system and effectively avoiding the sacrifice of time or space resolution;
(3) The invention can be applied to other in-vivo detection fields except laryngoscopes, and can be applied to wide clinical scenes only by adjusting the adopted micro lens sequence, the image sensor and the artificial intelligent algorithm parameters according to actual conditions.
Drawings
FIG. 1 is a schematic diagram of the overall structure of the present invention;
FIG. 2 is a flow chart of the generation of an image fusion model in the present invention;
FIG. 3 is a schematic diagram of an antagonistic neural network generated in accordance with the present invention.
In the figure, a 1-micro lens imaging assembly; 2-a flexible image transmission assembly; 3-an image processing module; 4-display assembly.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and examples.
As shown in fig. 1, the large-view laryngoscope based on deep learning of the invention comprises a micro-lens imaging component 1 for acquiring image data; the micro-lens imaging assembly comprises at least two groups of small-caliber micro-lenses with non-overlapping visual fields;
the flexible image transmission assembly 2 is connected with the micro-lens imaging assembly 1 and is used for transmitting image data; the flexible image transmission beam is a flexible image transmission beam with the diameter of the outer ring of the tail end smaller than 1.6mm and the number of 1-6;
the image processing module 3 is connected to the other end of the flexible image transmission component 2, the image processing module 3 performs training by using a generated antagonistic neural network according to the image data obtained by the micro-lens imaging component 1 to obtain an image fusion model, and the image fusion model re-splices the training output result into a complete image with gaps between non-overlapping view field images eliminated;
and the display component 4 is connected with the image processing module 3 and is used for displaying the complete image which is spliced again by the image processing module 3.
The invention expands the imaging field angle by replacing a single lens inside the lens in a conventional video laryngoscope with a micro lens group having a different orientation. And a plurality of lenses which are symmetrical front and back by taking the original direction as a symmetry axis are additionally arranged outside the direction which keeps the same orientation with the original lenses, the included angles of the lenses and the direction of the original lenses are distributed in an equi-differential array, and the field-of-view ranges of the adjacent lenses are ensured not to overlap at the same time, so that the field-of-view range is enlarged as much as possible, and the relative positions of the lenses are fixed.
As shown in FIG. 1, in a preferred embodiment, the micro-lens imaging unit 1 employs three groups of small-caliber micro-lenses, and the small-caliber lenses employ small-caliber lens modules 1/18 SEL120-015 manufactured by Sumito optical glass company (Sumita) of Japan, which are suitable for use in medical endoscopes, and have a horizontal angle of view of about 62.8. When the fields of view of adjacent lenses are not overlapped, the equivalent imaging field angle generated by the 3 adjacent lenses is at least 188.4 degrees. And (5) covering the front end of the laryngoscope lens with a corresponding transparent lens. The number and the position relation of the lenses in the step can be increased or decreased according to the actual requirement, but the size of the lenses should be selected mainly by considering that the size of the front end of the laryngoscope is not too large. One possible embodiment is that for patients with a mouth opening of more than 10mm, it is not preferable to select micro-lenses with a length exceeding 3 mm.
The arrangement of the micro lens groups can be adjusted according to practical situations, including whether the micro lens groups are overlapped, symmetry and other requirements can be adjusted. The purpose of the micro lens group is to generate a larger angle of view, and the model of the micro lens group can be adjusted according to the requirement. For example, the requirement on the quality of the lens can be reduced by adopting an image fusion algorithm with better performance; if a larger-scale lens is used, larger edge distortion may occur, and at this time, adjacent areas may be properly imaged with a certain overlap, so as to obtain better image fusion performance.
For the flexible image transmission component 2, the invention leads out the imaging result of the micro lens group to the outside of the body by adopting the bendable, light and long-service-life glass fiber bundle, thereby avoiding the difficulty of integrating a large-scale sensor to the front end of the laryngoscope. A specific embodiment corresponding to the above embodiment is that 3 flexible image transmission beams with an outer diameter of 1.2mm manufactured by the german schottky (schottky) company are used and are respectively connected with 3 micro lenses of the micro lens imaging unit 1. The symmetry of the 3 micro lenses ensures that the image transmission distances of the image transmission beams are in theoretical agreement. The 3 flexible image transmission bundles are tightly fixed and integrated into the laryngoscope to be led out of the laryngoscope, the other sides of the image transmission bundles are respectively fixed in different areas of the high-resolution sensor, and the images of the sensor are displayed on a screen in real time through matched software after subsequent processing. The number of lenses is consistent with the number of the adopted flexible image transmission beams, and the upper limit of the lenses is mainly considered that the size of the front end of the laryngoscope is not excessively large. One possible example is that for patients with a flare of more than 10mm, at most 5-6 image-transmitting bundles are used with a corresponding number of micro-lenses, since the proximal ferrule diameter of the flexible image-transmitting bundle with an outer diameter of 1.2mm is about 1.6 mm.
The image transmission device adopted by the flexible image transmission assembly 2 can be theoretically replaced by any other image transmission device such as an optical fiber panel, and the flexible image transmission assembly has the necessary properties of similar size, flexibility, easy connection with a lens, low imaging loss and the like.
In the large-view-angle image fusion part based on deep learning, the image processing method of the invention firstly places the acquired sub-images of each non-overlapping area at different positions of a blank image according to the spatial position relation of the micro lens group, so that the image theoretically comprises a plurality of gaps generated by non-overlapping images. The slit area does not belong to any sub-image, and the luminance value is 0. In this case, the elimination of the slit is completed, that is, the fusion of the images is realized. In particular, the present invention refers to the adoption of the characteristics of generating a challenge network (GAN) itself, learning and high robustness to the target signal, thus proposing a deep neural network based on a conditional GAN. The algorithm adopts the thought of supervised learning, and training of the deep neural network model is firstly carried out before execution.
In the large-view-angle image fusion part, the adopted artificial intelligent algorithm can be replaced by other neural network structures in theory, and parameters can be adjusted according to actual conditions.
As shown in fig. 2 and 3, the generating antagonistic neural network in the present invention includes a generator G and a discriminator D, and the training process is as follows: s1, acquiring training data; s2, building a generator neural network and a discriminator neural network; and S3, performing countermeasure training by the generator and the discriminator to obtain an image fusion model.
In a specific embodiment, in step S1 of generating a training process for an antagonistic neural network, the step of acquiring training data includes:
s11, aiming at throat static samples under the same conditions, the micro-lens imaging assembly 1 moves along a lens distribution plane, at least 10 series of images containing gaps are continuously collected, and the overlapping area of adjacent series of images is not less than 50% of the image area of Yu Shanzhang; the number of images shot by the 3 micro-lenses of the flexible image transmission assembly 1 is one, and as the fields of view of the adjacent lenses are not overlapped, a gap is necessarily included in one image; the overlapping area of the adjacent series of images is not smaller than 50% of the image area of Yu Shanzhang, so that when the images are matched with small blocks in image processing, enough paired training data can be fully obtained through matching; to ensure adequate training data, the series of images containing the slit is more preferably at least 20;
s12, cutting a series of images containing gaps into 256 multiplied by 256 pixel image small blocks without overlapping by taking one series of images as a reference, adopting scale-invariant feature transformation for each image small block, respectively carrying out feature matching with other images in the series of images, and acquiring an image small block area matched with each image small block;
s13, analyzing the image small block and the image small block area matched with the image small block: discarding if both contain gaps; if no gap exists, any one of the two is used as the input of the condition x for generating the antagonistic neural network, and the other is used as the true value y; otherwise, inputting the image small blocks containing the gaps as a condition x, and inputting the image small blocks without the gaps as a true value y; thus obtaining pairs of image block inputs as training data.
Preferably, the criterion for judging whether the small image block contains a gap is as follows: an image patch having more than 50 points with luminance values of 0 is a slit-containing image; otherwise, judging the image to be the image without the gap.
Preferably, in step S2, the generator G adopts a U-shaped symmetrical network structure with an encoder/decoder structure, the size of each layer of the feature map of the encoder section is halved, and the size of each layer of the decoder is doubled; the number of filters per layer is doubled layer by layer in the encoder section and halved in the decoder section.
Preferably, the number of first layer filters of generator G is 64, and the encoder and decoder each have 7 layers.
Preferably, in step S2, the arbiter D adopts a full convolution form, and there are 4 layers in total, and the size of each layer of feature map is halved, the number is doubled, and finally the matrix is averaged, and the obtained value represents a probability judgment on whether the image result is a true value.
Preferably, in the training process of step S3, for the generator G and the corresponding arbiter D, the loss function against the neural network is expressed as:
(1)
wherein the method comprises the steps ofAnd->Representing the mean value under the whole data, wherein z is the noise input of the generator G, x is the condition input of the countermeasure network, and y is the true value input of the countermeasure network training;
the loss function expression for the L1 norm error between the predicted value and the true value is:
(2)
the objective function formula of the countermeasure network training is as follows:
(3)
in the formula (3), λ is the weight of two losses, and λ is set to 100.
The lambda parameter serves to regulate the magnitude of both losses, set to 100 in the present invention. The deep learning procedure of the invention is implemented by Tensorflow, and an Adam optimizer (momentum of 0.5) is adopted, and the initial learning rate is set to 0.0002. For each sample, the training set size is 10000. The training model was obtained after 400 training on a single NVIDIA GeForce RTX 2080 Ti GPU.
In the embodiment, the real-time image acquired by the micro lens group of the micro lens imaging assembly 1 is divided into 256×256 pixel image small blocks without overlapping, the image small blocks are respectively input into the trained deep neural network model proposed by the method, the result G (x, z) output by the generator (G) is spliced again into a complete image, and the complete image is displayed on a screen. The execution time of the fusion algorithm is less than 1 second, and the real-time requirement is met.
The specific embodiments of the present invention are intended to be illustrative, rather than limiting, of the invention, and modifications thereof will be suggested to persons skilled in the art to which the present invention pertains without inventive contribution, as desired, after having read the present specification, but are to be protected by the patent law within the scope of the appended claims.
Claims (6)
1. A deep learning-based large viewing angle laryngoscope comprising:
a micro-lens imaging component (1) for acquiring image data; the micro-lens imaging assembly comprises at least two groups of small-caliber micro-lenses with non-overlapping visual fields;
the flexible image transmission assembly (2) is connected with the micro-lens imaging assembly and is used for transmitting image data; the flexible image transmission beam is a flexible image transmission beam with the diameter of the outer ring of the tail end smaller than 1.6mm and the number of 1-6;
the image processing module (3) is connected with the other end of the flexible image transmission component, and is used for training by using the generated antagonistic neural network according to the image data obtained by the micro-lens imaging component so as to obtain an image fusion model, and the image fusion model is used for re-splicing the training output result into a complete image with gaps between non-overlapping view field images eliminated;
the display component (4) is connected with the image processing module and is used for displaying the complete image re-spliced by the image processing module;
the generating antagonistic neural network comprises a generator G and a discriminator D, and the training process is as follows: s1, acquiring training data; s2, building a generator neural network and a discriminator neural network; s3, performing countermeasure training by the generator and the discriminator to obtain an image fusion model;
in step S1, the training data obtaining step includes:
s11, aiming at throat static samples under the same conditions, the micro-lens imaging assembly (1) moves along a lens distribution plane, at least 10 series of images containing gaps are continuously collected, and the overlapping area of adjacent series of images is not less than 50% of the image area of Yu Shanzhang;
s12, cutting a series of images containing gaps into 256 multiplied by 256 pixel image small blocks without overlapping by taking one series of images as a reference, adopting scale-invariant feature transformation for each image small block, respectively carrying out feature matching with other images in the series of images, and acquiring an image small block area matched with each image small block;
s13, analyzing the image small block and the image small block area matched with the image small block: discarding if both contain gaps; if no gap exists, any one of the two is used as the input of the condition x for generating the antagonistic neural network, and the other is used as the true value y; otherwise, inputting the image small blocks containing the gaps as a condition x, and inputting the image small blocks without the gaps as a true value y; thus obtaining pairs of image block inputs as training data.
2. The deep learning based large viewing angle laryngoscope according to claim 1, wherein the criterion for whether the image patch contains a slit is: an image patch having more than 50 points with luminance values of 0 is a slit-containing image; otherwise, judging the image to be the image without the gap.
3. The deep learning based large viewing laryngoscope according to claim 1, wherein in step S2, the generator G adopts a U-shaped symmetric network structure with encoder/decoder structure, the size of each layer of feature map of the encoder section is halved, and each layer of the decoder is doubled in size; the number of filters per layer is doubled layer by layer in the encoder section and halved in the decoder section.
4. A deep learning based large view laryngoscope according to claim 3 wherein the number of first layer filters of generator G is 64 and the encoder and decoder each have 7 layers.
5. The deep learning-based large viewing angle laryngoscope according to claim 1, wherein in step S2, the discriminator D adopts a full convolution form, and there are 4 layers, each layer of feature images is halved in size and doubled in number, and finally the matrix is averaged, and the obtained value represents a probability determination as to whether the image result is true.
6. The deep learning based large viewing laryngoscope according to claim 1, wherein during the training of step S3, for generator G and corresponding discriminator D, the loss function against neural network is expressed as:
(1)
wherein the method comprises the steps ofAnd->Representing the mean value under the whole data, wherein z is the noise input of the generator G, x is the condition input of the countermeasure network, and y is the true value input of the countermeasure network training;
the loss function expression for the L1 norm error between the predicted value and the true value is:
(2)
the objective function formula of the countermeasure network training is as follows:
(3)
in the formula (3), λ is the weight of two losses, and λ is set to 100.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310618436.6A CN116327103B (en) | 2023-05-30 | 2023-05-30 | Large-visual-angle laryngoscope based on deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310618436.6A CN116327103B (en) | 2023-05-30 | 2023-05-30 | Large-visual-angle laryngoscope based on deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116327103A CN116327103A (en) | 2023-06-27 |
CN116327103B true CN116327103B (en) | 2023-07-21 |
Family
ID=86879089
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310618436.6A Active CN116327103B (en) | 2023-05-30 | 2023-05-30 | Large-visual-angle laryngoscope based on deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116327103B (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5547455A (en) * | 1994-03-30 | 1996-08-20 | Medical Media Systems | Electronically steerable endoscope |
WO2022002399A1 (en) * | 2020-07-02 | 2022-01-06 | Ecole Polytechnique Federale De Lausanne (Epfl) | Multicore fiber endoscope for phase imaging based on intensity recording using deep neural networks |
KR20220157833A (en) * | 2021-05-21 | 2022-11-29 | (주)엔도아이 | Method and apparatus for learning image for detecting lesions |
CN115720506A (en) * | 2020-10-02 | 2023-02-28 | 豪雅株式会社 | Program, information processing method, and endoscope system |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CA3074106A1 (en) * | 2016-05-19 | 2017-11-23 | Psip, Llc | Methods for polyp detection |
US11011275B2 (en) * | 2018-02-12 | 2021-05-18 | Ai.Skopy, Inc. | System and method for diagnosing gastrointestinal neoplasm |
US11563929B2 (en) * | 2019-06-24 | 2023-01-24 | Align Technology, Inc. | Intraoral 3D scanner employing multiple miniature cameras and multiple miniature pattern projectors |
US20220287545A1 (en) * | 2021-03-11 | 2022-09-15 | Medtronic Advanced Energy Llc | Enhanced vision in electrosurgery |
CN118284907A (en) * | 2021-09-16 | 2024-07-02 | 史赛克公司 | Method and system for generating intraoperative imaging data of an object |
-
2023
- 2023-05-30 CN CN202310618436.6A patent/CN116327103B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5547455A (en) * | 1994-03-30 | 1996-08-20 | Medical Media Systems | Electronically steerable endoscope |
WO2022002399A1 (en) * | 2020-07-02 | 2022-01-06 | Ecole Polytechnique Federale De Lausanne (Epfl) | Multicore fiber endoscope for phase imaging based on intensity recording using deep neural networks |
CN115720506A (en) * | 2020-10-02 | 2023-02-28 | 豪雅株式会社 | Program, information processing method, and endoscope system |
KR20220157833A (en) * | 2021-05-21 | 2022-11-29 | (주)엔도아이 | Method and apparatus for learning image for detecting lesions |
Non-Patent Citations (2)
Title |
---|
Mitigating fluorescence spectral overlap in wide-field endoscopic imaging;Chenying Yang Vivian Hou Leonard Y. Nelson Eric J. Seibel;Journal of Biomedical Optics;第18卷(第8期);全文 * |
基于StyleGAN的心脏表面三维重建方法研究;徐思源;中国优秀硕士学位论文全文数据库 医药卫生科技辑(第1期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN116327103A (en) | 2023-06-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10203493B2 (en) | Optical systems for multi-sensor endoscopes | |
JP4982358B2 (en) | An imaging device that looks in all and forward directions | |
KR100852310B1 (en) | Endoscope and endoscope system | |
CN111308690B (en) | Optical field electronic endoscopic device and imaging method thereof | |
CN103460685B (en) | Image pickup apparatus and image pickup method | |
CN110463174A (en) | For the optical system of surgical probe, the system and method for forming it, and the method for executing surgical operation | |
JP2020006176A (en) | Camera scope electronic variable prism | |
CN109068035B (en) | Intelligent micro-camera array endoscopic imaging system | |
JPH03165732A (en) | Detecting method for insertion direction of endoscope | |
CN116327103B (en) | Large-visual-angle laryngoscope based on deep learning | |
WO2009150653A1 (en) | Optical system for use in an endoscope | |
WO2012103767A1 (en) | New 3d electronic choledochoscopy system and method for use thereof | |
CN210902962U (en) | Laparoscope external view mirror device capable of scanning inside of abdominal cavity | |
Roulet et al. | 360 endoscopy using panomorph lens technology | |
CN110680264A (en) | 3D optical endoscope system based on dual-optical-path design | |
RU2337606C1 (en) | Optical system of endoscope | |
CN210962341U (en) | Robot outer sight glass with confocal laser scanning function | |
JP2014054499A (en) | Full field small-sized imaging device using cylinder type image pickup element | |
CN205411128U (en) | Micro - endoscope system | |
CN216526523U (en) | Multipurpose hard endoscope with large visual field and large field depth | |
WO2024122323A1 (en) | Imaging device, operating method of imaging device, and program | |
CN107040745B (en) | method for improving frame frequency of probe type confocal micro endoscope | |
CN218074940U (en) | Compatible three-dimensional hard tube endoscope | |
CN106444004A (en) | Front and rear field-of-view electronic endoscope | |
CN119279465A (en) | 8K stereoscopic display endoscope |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |