Nothing Special   »   [go: up one dir, main page]

CN117391973A - Image de-motion blur method based on multi-scale improved residual block CNN - Google Patents

Image de-motion blur method based on multi-scale improved residual block CNN Download PDF

Info

Publication number
CN117391973A
CN117391973A CN202311444732.5A CN202311444732A CN117391973A CN 117391973 A CN117391973 A CN 117391973A CN 202311444732 A CN202311444732 A CN 202311444732A CN 117391973 A CN117391973 A CN 117391973A
Authority
CN
China
Prior art keywords
image
scale
convolution
layer
improved residual
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311444732.5A
Other languages
Chinese (zh)
Inventor
陈新
谢非
杨继全
夏光圣
张策
张弘毅
王凯琳
陈雅婷
蔺莹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Normal University
Original Assignee
Nanjing Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Normal University filed Critical Nanjing Normal University
Priority to CN202311444732.5A priority Critical patent/CN117391973A/en
Publication of CN117391973A publication Critical patent/CN117391973A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • G06T7/0008Industrial image inspection checking presence/absence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • G06N3/0455Auto-encoder networks; Encoder-decoder networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection
    • G06T2207/30164Workpiece; Machine component
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y04INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
    • Y04SSYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
    • Y04S10/00Systems supporting electrical power generation, transmission or distribution
    • Y04S10/50Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an image motion blur removing method based on a multi-scale improved residual block CNN, which comprises the following steps: collecting and storing electric power inspection video data of the transmission line inspection robot, preparing a data set, and marking the data set as a source data set; clipping and compressing the source data set to obtain a clear and fuzzy image pair, and recording the clear and fuzzy image pair as a power inspection deblurring data set; constructing a multi-scale improved residual block convolutional neural network, and training the multi-scale improved residual block convolutional neural network by using a power inspection deblurring data set to obtain a trained multi-scale improved residual block convolutional neural network model; and deblurring the power inspection blurred image data by using a trained multi-scale improved residual block convolutional neural network model. The invention improves the deblurring effect and accuracy of the power inspection image, is suitable for processing the problem of the image shooting blur of the high-voltage power transmission inspection robot, and has the advantages of high operation speed, good deblurring effect and strong environment interference resistance.

Description

基于多尺度改进残差块CNN的图像去运动模糊方法Image de-motion blur method based on multi-scale improved residual block CNN

技术领域Technical field

本发明属于电力巡检数据处理领域,涉及电力巡检模糊图像数据采集及深度学习技术,具体涉及一种基于多尺度改进残差块CNN的图像去运动模糊方法。The invention belongs to the field of power inspection data processing, relates to power inspection fuzzy image data collection and deep learning technology, and specifically relates to an image de-motion blur method based on multi-scale improved residual block CNN.

背景技术Background technique

目前,我国主要有人工巡检,无人机巡检,巡检机器人三种巡检方式,来对架空输电线路进行定期巡检。机器人巡检是使用可攀爬线缆的机器人,以一定速度沿输电线缆进行行走,并且能够跨越一些线缆上的障碍物,利用自身搭载的相机和检测设备对线缆进行检查架空输电线路的方法。巡检机器人能在输电线路上自主行走,通过携带的可见光和红外摄像机等巡检设备,可以近距离的对线路通道和本体进行巡检,而且不存在巡检盲区,不受空域管制,巡视效率高、效果好,与其它巡检方式相比,在山川、湖泊、森林等地区的优势更为突出。该方式不仅降低了巡检的人力成本,而且提升了工作效率和巡检精度。At present, my country mainly has three inspection methods: manual inspection, drone inspection, and inspection robot to conduct regular inspections of overhead transmission lines. Robot inspection uses a cable-climbing robot to walk along transmission cables at a certain speed, and can cross some obstacles on the cables. It uses its own camera and detection equipment to inspect the cables. Overhead transmission lines Methods. The inspection robot can walk autonomously on the transmission line. Through the inspection equipment such as visible light and infrared cameras it carries, it can inspect the line channel and body at close range. There are no inspection blind spots, and it is not subject to airspace control, which improves inspection efficiency. High efficiency and good effect. Compared with other inspection methods, its advantages are more prominent in areas such as mountains, rivers, lakes, and forests. This method not only reduces the labor cost of inspection, but also improves work efficiency and inspection accuracy.

然而,在实际的工程现场中,巡检机器人由于拍照的焦距远近,大风天气下的机器人本体和云台相机的抖动,拍摄物体和成像设备之间的相对运动,光线强弱等原因,拍摄图像经常出现模糊现象,图像模糊对后续技术工作获取清晰图像信息造成很大影响,甚至中断后续一些基于图像的智能化技术工作,比如图像识别、目标检测、图像分割。如何使得模糊图像变得清晰,以此增强图像数据的可靠性和可读性,为后续技术工作提供前提,成为亟待解决的问题。However, in the actual engineering site, the inspection robot has difficulty in capturing images due to the focal length of the photo, the jitter of the robot body and the gimbal camera in windy weather, the relative movement between the photographed object and the imaging device, the intensity of the light, etc. Blurring often occurs, and image blurring has a great impact on subsequent technical work in obtaining clear image information, and even interrupts some subsequent image-based intelligent technical work, such as image recognition, target detection, and image segmentation. How to make blurry images clear, thereby enhancing the reliability and readability of image data and providing a prerequisite for subsequent technical work, has become an urgent problem to be solved.

发明内容Contents of the invention

发明目的:为了克服现有技术中存在的不足,提供一种基于多尺度改进残差块CNN的图像去运动模糊方法,具有运算速度快、去运动模糊效果好、抗环境干扰能力强的优点。Purpose of the invention: In order to overcome the deficiencies in the existing technology, an image de-motion blur method based on multi-scale improved residual block CNN is provided, which has the advantages of fast operation speed, good motion blur removal effect, and strong anti-environmental interference ability.

技术方案:为实现上述目的,本发明提供一种基于多尺度改进残差块CNN的图像去运动模糊方法,包括如下步骤:Technical solution: In order to achieve the above objectives, the present invention provides an image de-motion blur method based on multi-scale improved residual block CNN, which includes the following steps:

S1:采集并保存输电线路巡检机器人的电力巡检视频数据,对所采集的电力巡检视频数据进行分帧处理,然后进行图像样本筛选,图像预处理,得到清晰和模糊图像对,制成数据集,并记为源数据集;S1: Collect and save the power inspection video data of the transmission line inspection robot, perform frame processing on the collected power inspection video data, and then perform image sample screening and image preprocessing to obtain clear and blurred image pairs, which are made Data set, and recorded as the source data set;

S2:完成对源数据集的图像再处理,将源数据集进行裁剪和压缩,得到133对清晰和模糊图像对,并记为电力巡检去模糊数据集,该数据集是本发明独有的,目前电力巡检方向还未出现去模糊数据集,与传统公开数据集GoPro制作方法不同,GoPro数据集是采用十几张连续帧的清晰图像融合成一张模糊图像的方法,电力巡检去模糊数据集是采用单张模糊图像的压缩裁剪方法形成新的模糊图像;S2: Complete the image reprocessing of the source data set, crop and compress the source data set, and obtain 133 pairs of clear and blurred images, which are recorded as the power inspection deblurred data set. This data set is unique to the present invention. , At present, there is no deblurred data set in the direction of power inspection. Different from the traditional public data set GoPro production method, the GoPro data set uses a method of merging clear images of more than a dozen consecutive frames into one blurred image. Deblurring of power inspection The data set uses the compression and cropping method of a single blurred image to form a new blurred image;

S3:搭建多尺度改进残差块卷积神经网络,并使用电力巡检去模糊数据集对多尺度改进残差块卷积神经网络进行充分训练,得到训练好的多尺度改进残差块卷积神经网络模型,该模型算法与传统多尺度卷积神经网络不同之处在于网络残差结构的改进和损失函数的构造;S3: Build a multi-scale improved residual block convolutional neural network, and use the power inspection deblurred data set to fully train the multi-scale improved residual block convolutional neural network, and obtain the trained multi-scale improved residual block convolution Neural network model. The difference between this model algorithm and the traditional multi-scale convolutional neural network lies in the improvement of the network residual structure and the construction of the loss function;

S4:使用训练好的多尺度改进残差块卷积神经网络模型对电力巡检模糊图像数据进行去模糊,再对去模糊图像进行目标检测和识别,判断电力巡检目标是否为缺陷状态,并且在后台和云端记录。S4: Use the trained multi-scale improved residual block convolutional neural network model to deblur the power inspection blurred image data, and then perform target detection and recognition on the deblurred image to determine whether the power inspection target is in a defective state, and Record in the background and in the cloud.

进一步地,所述步骤S1具体为:Further, the step S1 is specifically:

A1:在输电线路巡检机器人晃动状态下,对巡检机器人的电力巡检视频进行数据采集,得到并保存输电线路巡检机器人晃动状态下电力巡检视频数据,电力巡检视频包括同一场景下的清晰图像和模糊图像;A1: When the transmission line inspection robot is shaking, collect data from the power inspection video of the inspection robot, and obtain and save the power inspection video data when the transmission line inspection robot is shaking. The power inspection video includes the same scene. clear images and blurred images;

A2:采用视频分帧处理方法对步骤A1采集得到的巡检视频进行分帧处理操作,得到分帧处理后的图像,然后进行图像样本筛选,得到符合要求的图像;A2: Use the video frame processing method to perform frame processing on the inspection video collected in step A1 to obtain the framed image, and then screen the image samples to obtain images that meet the requirements;

A3:采用直方图均衡化方法对步骤A2采集得到的图像进行预处理操作,得到预处理后的图像数据;A3: Use the histogram equalization method to preprocess the image collected in step A2 to obtain the preprocessed image data;

A4:对步骤A3得到的预处理后的数据进行分类和配对操作,得到一一对应的清晰和模糊图像对,并记为源数据集。A4: Classify and pair the preprocessed data obtained in step A3 to obtain one-to-one corresponding pairs of clear and blurred images, which are recorded as source data sets.

进一步地,所述步骤A1包括:Further, the step A1 includes:

在输电线路巡检机器人晃动状态下,对巡检机器人的电力巡检视频进行数据采集,由巡检机器人在不同巡检线路位置,不同焦距下,不同场景下拍摄,得到一系列帧率为16帧/S的短视频,并保存。When the transmission line inspection robot is shaking, the data of the power inspection video of the inspection robot is collected. The inspection robot shoots at different inspection line positions, different focal lengths, and different scenes to obtain a series of frame rates of 16 Frame/S short video and save it.

进一步地,所述步骤A2具体为:Further, the step A2 is specifically:

由步骤A1采集到的视频,逐个进行分帧处理,得到16*t张图像,其中t为视频的时间长度;The videos collected in step A1 are processed into frames one by one to obtain 16*t images, where t is the time length of the video;

将分帧处理后的图像,进行样本筛选,选择包含电力巡检目标的图像,丢弃与电力巡检目标无关的图像。Perform sample screening on the frame-processed images, select images that contain the power inspection target, and discard images that have nothing to do with the power inspection target.

进一步地,所述步骤A3具体为:Further, the step A3 is specifically:

采用直方图均衡化方法,对步骤A2所得图像进行预处理,用于增强图像局部的对比度而不影响整体的对比度;图像直方图定义:一个灰度级在范围[0,L-1]的数字图像,该图像的直方图是一个离散函数:The histogram equalization method is used to preprocess the image obtained in step A2 to enhance the local contrast of the image without affecting the overall contrast; image histogram definition: a number with a gray level in the range [0, L-1] An image whose histogram is a discrete function:

p(rk)=nk/n (1)p(r k )=n k /n (1)

其中p为概率函数,rk是第k个灰度级,k=0,1,2,…,L-1,L为灰度级总数,nk是图像中第k个灰度级的像素总数,n是图像的像素总数,继而得到预处理后的图像。where p is the probability function, r k is the k-th gray level, k = 0,1,2,...,L-1, L is the total number of gray levels, n k is the pixel of the k-th gray level in the image The total number, n is the total number of pixels in the image, and then the preprocessed image is obtained.

进一步地,所述步骤A4包括:Further, the step A4 includes:

将预处理后的图像进行分类,分成清晰图像和模糊图像两大类;Classify the preprocessed images into two categories: clear images and blurred images;

对得到的清晰和模糊图像,以相同背景和相同巡检目标位置为条件,一一对应进行配对,得到源数据集A1The obtained clear and blurred images are paired in one-to-one correspondence on the condition of the same background and the same inspection target position, and the source data set A 1 is obtained.

进一步地,所述步骤S2具体为:Further, the step S2 is specifically:

B1:将源数据集A1中的图像进行裁剪,原图像分辨率为2688*1520,采用固定大小相框移动的方法,裁剪包含电力巡检目标的区域,所得图像的分辨率,长度要大于256像素,宽度要大于256像素,清晰和模糊图像对的分辨率大小要完全相等;B1: Crop the image in the source data set A1 . The original image resolution is 2688*1520. Use the method of moving the fixed size photo frame to crop the area containing the power inspection target. The resolution and length of the resulting image must be greater than 256 Pixels, the width must be greater than 256 pixels, and the resolution sizes of clear and blurry image pairs must be exactly equal;

B2:将步骤B1所得的图像,进行图像压缩,保证图像长度要大于256像素,宽度要大于256像素,清晰和模糊图像对的分辨率大小要完全相等,由此得到133对清晰和模糊图像对,并记为电力巡检去模糊数据集A2B2: Carry out image compression on the image obtained in step B1 to ensure that the image length is greater than 256 pixels and the width is greater than 256 pixels. The resolutions of the clear and blurry image pairs must be completely equal, thus obtaining 133 pairs of clear and blurry image pairs. , and recorded as the power inspection deblurred data set A 2 .

进一步地,所述步骤S3具体为:Further, the step S3 is specifically:

C1:搭建多尺度改进残差块卷积神经网络,并确定网络结构;C1: Build a multi-scale improved residual block convolutional neural network and determine the network structure;

C2:选择网络优化器并使用公开数据集GoPro与电力巡检去模糊数据集A2对多尺度改进残差块卷积神经网络进行训练,得到训练好的多尺度改进残差块卷积神经网络模型。C2: Select the network optimizer and use the public data set GoPro and the power inspection deblurred data set A2 to train the multi-scale improved residual block convolutional neural network to obtain the trained multi-scale improved residual block convolutional neural network. Model.

进一步地,所述步骤C1中多尺度改进残差块卷积神经网络的结构由3个不同大小尺度的神经网络结构组成,在每个尺度下生成一个清晰的图像,作为去模糊任务的子问题,该子问题将一张模糊图像和初始去模糊的结果(来自上一级尺度的下采样)作为输入,并推测在这一尺度下,这张清晰的图像为:Further, the structure of the multi-scale improved residual block convolutional neural network in step C1 is composed of three neural network structures of different size scales, and a clear image is generated at each scale as a sub-problem of the deblurring task , this sub-problem takes a blurred image and the initial deblurred result (downsampling from the previous scale) as input, and infers that at this scale, this clear image is:

Im,hm=Net(Bm,Im+1,hm+1;θ) (2)I m , h m =Net(B m ,I m+1 ,h m+1 ; θ) (2)

其中,Im为本级尺度下该层网络输出的去模糊图像,hm为本级尺度下的隐藏层特征,Net为本发明提出的多尺度改进残差块卷积神经网络,Bm为本级尺度下该层网络输入的模糊图像,Im+1为上一级尺度输出的去模糊图像,hm+1为上一级尺度的隐藏层特征,θ为待训练参数,m为本级尺度,m+1为上一级尺度。Among them, I m is the deblurred image output by the layer network at this level of scale, h m is the hidden layer feature at this level of scale, Net is the multi-scale improved residual block convolutional neural network proposed by the present invention, and B m is The blurred image input by the network of this layer at this level of scale, I m+1 is the deblurred image output by the previous level of scale, h m+1 is the hidden layer feature of the previous level of scale, θ is the parameter to be trained, and m is the original Level scale, m+1 is the upper level scale.

进一步地,所述步骤C1中任一尺度下的神经网络结构相同,具体搭建方法如下:Further, the neural network structure at any scale in step C1 is the same, and the specific construction method is as follows:

第一部分是输入模块,包括一个卷积层和三个改进残差块;其中的任意一个改进残差块,与传统意义的改进残差块结构不同,本发明利用拼接操作代替原有改进残差块的逐元素相加操作,包含两个卷积层和一个线性整流函数和一个跳层拼接;拼接操作相比于原有改进残差块的逐元素相加操作有以下三个优点:1.信息融合:通过将两个特征图在通道维度上连接,可以实现信息的融合和组合,使网络能够利用更多的上下文信息。2.表示能力增强:连接操作扩展了特征图的通道数,增加了网络的表示能力,有助于学习更复杂的特征表示。3.多尺度特征融合:通过连接具有不同空间分辨率的特征图可以实现多尺度的特征融合,提升模型对不同尺度目标的感知能力。第一部分的输入来自于公开数据集GoPro与电力巡检去模糊数据集A2,数据集里面的每张图像,经过OpenCV程序处理,得到256*256大小的局部图像,进而送到输入模块,输入第一个卷积层,该层选用了64个3*3的卷积核,步长为1,填充数量为1,卷积层的输出尺寸计算公式如下列公式所示:The first part is the input module, including a convolution layer and three improved residual blocks; any one of the improved residual blocks is different from the traditional improved residual block structure. This invention uses splicing operations to replace the original improved residual blocks. The element-by-element addition operation of the block includes two convolutional layers, a linear rectification function and a skip layer splicing; the splicing operation has the following three advantages compared to the original element-wise addition operation of the improved residual block: 1. Information fusion: By connecting two feature maps in the channel dimension, the fusion and combination of information can be achieved, allowing the network to utilize more contextual information. 2. Representation capability enhancement: The connection operation expands the number of channels of the feature map, increases the representation capability of the network, and helps to learn more complex feature representations. 3. Multi-scale feature fusion: Multi-scale feature fusion can be achieved by connecting feature maps with different spatial resolutions, improving the model's perception of targets at different scales. The input of the first part comes from the public data set GoPro and the power inspection deblurred data set A2 . Each image in the data set is processed by the OpenCV program to obtain a partial image of 256*256 size, and then sent to the input module. The first convolution layer uses 64 3*3 convolution kernels, the step size is 1, and the number of padding is 1. The output size calculation formula of the convolution layer is as follows:

其中,Z是卷积输出数据的长度,W是卷积输入数据的长度,P是填充数量,F是卷积核的长度,S表示步长;对于第一部分中的第一个编码块,第一个卷积层的输出尺寸由计算公式(3)计算得到第一个卷积层的输出大小是256*256*64;由于第一部分的卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第一部分的卷积层和三个改进残差块的最终输出大小也是256*256*64;Among them, Z is the length of the convolution output data, W is the length of the convolution input data, P is the number of padding, F is the length of the convolution kernel, and S represents the step size; for the first coding block in the first part, the The output size of a convolution layer is calculated by formula (3) and the output size of the first convolution layer is 256*256*64; since the convolution kernel size and format used in the first part of the convolution layer are the same, the parameters Same, and so on, the final output size of the first part of the convolutional layer and the three improved residual blocks is also 256*256*64;

第二部分由两个编码块组成的编码结构,其中任意一个编码块包含一个卷积层,一个线性整流函数和三个改进残差块;对于第二部分中的第一个编码块,该层选用了128个3*3的卷积核,步长为2,填充数量为0.5,第一个卷积层的输出尺寸计算公式(3),计算得到第一个卷积层的输出大小是128*128*128;在编码块的第一个卷积层后使用线性整流函数作为激活函数,将经过激活函数的数据送入编码块的第二个卷积层,第二个卷积层采用64个3*3卷积核,步长为2,填充数量为0.5,则根据卷积层的输出尺寸计算公式(3),第二部分第二个卷积层的输出大小也是128*128*128;由于第二部分的卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第二部分的第一个编码块最终输出大小是128*128*128;同理,第二部分的第二个编码块,该层选用了256个3*3的卷积核,步长为2,填充数量为0.5,第一个卷积层的输出尺寸由计算公式(3)计算得到第一个卷积层的输出大小是64*64*256;由于第二部分的卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第二部分的第二个编码块最终输出大小是64*64*256;The second part is a coding structure composed of two coding blocks, where any coding block contains a convolutional layer, a linear rectification function and three improved residual blocks; for the first coding block in the second part, this layer 128 3*3 convolution kernels are selected, the step size is 2, and the padding amount is 0.5. The output size of the first convolution layer is calculated using the formula (3). The calculated output size of the first convolution layer is 128 *128*128; After the first convolutional layer of the coding block, a linear rectification function is used as the activation function, and the data after the activation function is sent to the second convolutional layer of the coding block. The second convolutional layer uses 64 A 3*3 convolution kernel, the step size is 2, and the padding amount is 0.5. According to the output size calculation formula (3) of the convolution layer, the output size of the second convolution layer in the second part is also 128*128*128 ; Since the convolutional layer in the second part uses the same convolution kernel size and format, the same parameters, and so on, the final output size of the first coding block in the second part is 128*128*128; similarly, the second For the second coding block of the part, this layer uses 256 3*3 convolution kernels, the step size is 2, and the number of padding is 0.5. The output size of the first convolution layer is calculated by the calculation formula (3). The output size of a convolutional layer is 64*64*256; since the second part of the convolutional layer uses the same convolution kernel size and format, the same parameters, and so on, the second coding block of the second part finally The output size is 64*64*256;

第三部分由两个解码块组成,其中任意一个解码块由三个改进残差块,一个反卷积层和一个线性整流函数组成;反卷积层的输出尺寸计算公式如下列公式所示:The third part consists of two decoding blocks, any of which is composed of three improved residual blocks, a deconvolution layer and a linear rectification function; the output size calculation formula of the deconvolution layer is as follows:

Z1=(W1-1)×S1-2×P1+F1 (4)Z 1 =(W 1 -1)×S 1 -2×P 1 +F 1 (4)

其中,Z1是卷积输出数据的长度,W1是卷积输入数据的长度,P1是填充数量,F1是卷积核的长度,S1表示步长;根据反卷积层的输出尺寸计算公式(4),该层选用了256个3*3的卷积核,步长为2,填充数量为0.5,第三部分第一个反卷积层的输出大小是128*128*256;由于第三部分的反卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第三部分的第一个解码块最终输出大小是128*128*256;同理,第三部分的第二个解码块,该层选用了128个3*3的卷积核,步长为2,填充数量为0.5,第一个反卷积层的输出尺寸由计算公式(4)计算得到第一个反卷积层的输出大小是256*256*128;由于第二个解码块的反卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第三部分的第二个解码块最终输出大小是256*256*128;Among them, Z 1 is the length of the convolution output data, W 1 is the length of the convolution input data, P 1 is the number of padding, F 1 is the length of the convolution kernel, and S 1 represents the step size; according to the output of the deconvolution layer Size calculation formula (4), this layer uses 256 3*3 convolution kernels, the step size is 2, and the number of padding is 0.5. The output size of the first deconvolution layer in the third part is 128*128*256 ; Since the deconvolution layer in the third part uses the same convolution kernel size and format, the same parameters, and so on, the final output size of the first decoding block in the third part is 128*128*256; similarly, the The second decoding block of the three parts, this layer uses 128 3*3 convolution kernels, the stride is 2, and the number of padding is 0.5. The output size of the first deconvolution layer is calculated by formula (4) The output size of the first deconvolution layer is 256*256*128; since the deconvolution layer of the second decoding block uses the same convolution kernel size and format, the same parameters, and so on, the third part The final output size of the second decoding block is 256*256*128;

第四部分是输出模块,包括三个改进残差块和一个卷积层;该层选用了64个3*3的卷积核,步长为1,填充数量为1,第四部分第一个卷积层的输出大小是256*256*64;由于第四部分的卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第四部分的解码块最终输出大小是256*256*64。The fourth part is the output module, including three improved residual blocks and a convolution layer; this layer uses 64 3*3 convolution kernels, the step size is 1, and the number of padding is 1. The first part of the fourth part The output size of the convolution layer is 256*256*64; since the convolution kernel size and format used in the fourth part of the convolution layer are the same, the parameters are the same, and so on, the final output size of the decoding block in the fourth part is 256* 256*64.

进一步地,所述步骤C2具体为:Further, the step C2 is specifically:

将公开数据集GoPro与电力巡检去模糊数据集A2按照4:1的比例划分为训练集和测试集,在训练的过程中,损失函数使用四损失权重平衡函数,四损失权重平衡函数定义:The public data set GoPro and the power inspection deblurred data set A2 are divided into training sets and test sets according to the ratio of 4:1. During the training process, the loss function uses a four-loss weight balance function, and the four-loss weight balance function is defined :

L=Lp+k1L1+k2L2+k3L3 (5)L=L p +k 1 L 1 +k 2 L 2 +k 3 L 3 (5)

其中,L为四损失权重平衡函数,k1,k2与k3为融合系数,本发明中k1,k2与k3都取0.3数值大小,Lp为感知损失函数,L1为L1范数损失函数,L2为L2范数损失函数,L3为多尺度频率重建损失函数,φi,j是VGG19网络中激活之后的第j个卷积在第i层最大池化之前所取得的特征映射,Wi,j是特征映射的宽度,Hi,j是特征映射的高度,S代表清晰图像,I代表生成的去模糊图像,x是特征映射宽度方向的坐标值,y是特征映射高度方向的坐标值,n为尺度的批次,In为本网络中第n尺度输出的去模糊图像,Sn为本网络中第n尺度输出的清晰图像,tk表示第k个尺度下的元素总数,||·||1是L1距离计算符,||·||2是L2距离计算符,F表示将图像信号传输到频域的快速傅立叶变换(FFT);Among them, L is the four-loss weight balance function, k 1 , k 2 and k 3 are fusion coefficients. In the present invention, k 1 , k 2 and k 3 all take a numerical value of 0.3, L p is the perceptual loss function, and L 1 is L1 Norm loss function, L 2 is the L2 norm loss function, L 3 is the multi-scale frequency reconstruction loss function, φ i,j is the j-th convolution after activation in the VGG19 network before the i-th layer maximum pooling. The feature map of Mapping the coordinate value in the height direction, n is the batch of scales, I n is the deblurred image output by the nth scale in this network, S n is the clear image output by the nth scale in this network, t k represents the kth scale The total number of elements under , ||·|| 1 is the L1 distance calculator, ||·|| 2 is the L2 distance calculator, and F represents the fast Fourier transform (FFT) that transmits the image signal to the frequency domain;

网络优化器采用亚当优化器,训练批次尺寸为24,学习率为对5×10-5。对网络进行充分训练,得到训练好的多尺度改进残差块卷积神经网络模型。The network optimizer adopts Adam optimizer, the training batch size is 24, and the learning rate is 5×10 -5 . The network is fully trained to obtain the trained multi-scale improved residual block convolutional neural network model.

进一步地,所述步骤S4包括:Further, the step S4 includes:

D1:实时采集电力巡检机器人巡检过程中拍摄的巡检图像数据,使用步骤S3得到的训练好的模型对实时采集的图像数据进行去模糊,得到巡检目标当前的状态情况;D1: Collect the inspection image data captured during the inspection process of the power inspection robot in real time, and use the trained model obtained in step S3 to deblur the image data collected in real time to obtain the current status of the inspection target;

D2:根据步骤D1得到的巡检目标去模糊图像,进行图像的目标检测和识别,若出现巡检目标出现缺陷和损坏,后台和云端发出警报并进行数据记录。D2: Deblur the image based on the inspection target obtained in step D1, and perform target detection and recognition of the image. If the inspection target is defective or damaged, the background and cloud will issue an alarm and record the data.

有益效果:本发明与现有技术相比,具备如下优点Beneficial effects: Compared with the existing technology, the present invention has the following advantages:

1、本发明提供了基于多尺度改进残差块CNN的图像去运动模糊方法,其充分考虑电力巡检机器人拍摄模糊图像的特点,设计了多尺度改进残差块卷积神经网络。利用清晰和模糊图像数据的比对来进行特征比对,解决了目前电力巡检机器人的巡检图像去运动模糊问题。并且具有很高的可移植性,可应用于嵌入式系统平台,应用前景广泛。本发明具有运算速度快、去运动模糊效果好、抗环境干扰能力强的优点。1. The present invention provides an image de-motion blur method based on multi-scale improved residual block CNN, which fully considers the characteristics of blurred images captured by power inspection robots and designs a multi-scale improved residual block convolutional neural network. The comparison of clear and blurred image data is used for feature comparison, which solves the problem of motion blur in inspection images of current power inspection robots. And it has high portability, can be applied to embedded system platforms, and has broad application prospects. The invention has the advantages of fast calculation speed, good motion blur removal effect and strong anti-environmental interference ability.

2、本发明针对电力巡检机器人因抖动而导致拍摄图像出现模糊的问题,制作了符合电力巡检场景的数据集,提出了基于多尺度改进残差块卷积神经网络去电力巡检图像运动模糊的方法,设计了多尺度改进残差块卷积神经网络模型与算法。本发明提升了电力巡检图像去模糊的效果与准确性,适用于处理高压输电巡检机器人的巡检图像拍摄模糊问题。2. In order to solve the problem of blurred images of the electric power inspection robot due to jitter, the present invention creates a data set that conforms to the electric power inspection scene, and proposes a multi-scale improved residual block convolutional neural network to remove the motion of the electric power inspection image. Using the fuzzy method, a multi-scale improved residual block convolutional neural network model and algorithm are designed. The invention improves the effect and accuracy of deblurring power inspection images, and is suitable for solving the blurring problem of inspection images taken by high-voltage power transmission inspection robots.

附图说明Description of the drawings

图1是本发明实施例提供的基于多尺度改进残差块CNN的图像去运动模糊方法的工作流程示意图;Figure 1 is a schematic workflow diagram of an image de-motion blur method based on multi-scale improved residual block CNN provided by an embodiment of the present invention;

图2是本发明实施例提供的一种基于多尺度改进残差块卷积神经网络的电力巡检图像去运动模糊的巡检系统总体框图;Figure 2 is an overall block diagram of an inspection system for removing motion blur from power inspection images based on a multi-scale improved residual block convolutional neural network provided by an embodiment of the present invention;

图3是本发明实施例提供的一种巡检机器人的坐标系统示意图;Figure 3 is a schematic diagram of the coordinate system of an inspection robot provided by an embodiment of the present invention;

图4是本发明实施例提供的一种多尺度改进残差块卷积神经网络模型图;Figure 4 is a multi-scale improved residual block convolutional neural network model diagram provided by an embodiment of the present invention;

图5是本发明实施例提供的一种基于多尺度改进残差块卷积神经网络的电力巡检图像去运动模糊网络总体结构图;Figure 5 is an overall structural diagram of a power inspection image motion blur removal network based on a multi-scale improved residual block convolutional neural network provided by an embodiment of the present invention;

图6是本发明实施例提供的电力巡检图像去运动模糊现场图像。Figure 6 is an on-site image of a power inspection image with motion blur removed according to an embodiment of the present invention.

具体实施方式Detailed ways

下面结合附图和具体实施例,进一步阐明本发明,应理解这些实施例仅用于说明本发明而不用于限制本发明的范围,在阅读了本发明之后,本领域技术人员对本发明的各种等价形式的修改均落于本申请所附权利要求所限定的范围。The present invention will be further clarified below in conjunction with the accompanying drawings and specific examples. It should be understood that these examples are only used to illustrate the present invention and are not intended to limit the scope of the present invention. After reading the present invention, those skilled in the art will be familiar with various aspects of the present invention. Modifications in the form of equivalents fall within the scope defined by the appended claims of this application.

为了解决高压输电巡检机器人的巡检图像拍摄模糊问题,如图1和图2所示,本发明提供一种基于多尺度改进残差块CNN的图像去运动模糊方法,包括如下步骤:In order to solve the blurring problem of inspection images captured by high-voltage power transmission inspection robots, as shown in Figures 1 and 2, the present invention provides an image de-motion blur method based on multi-scale improved residual block CNN, which includes the following steps:

S1:采集并保存输电线路巡检机器人的电力巡检视频数据,对所采集的电力巡检视频数据进行分帧处理,然后进行图像样本筛选,图像预处理,得到清晰和模糊图像对,制成数据集,并记为源数据集;S1: Collect and save the power inspection video data of the transmission line inspection robot, perform frame processing on the collected power inspection video data, and then perform image sample screening and image preprocessing to obtain clear and blurred image pairs, which are made Data set, and recorded as the source data set;

S2:完成对源数据集的图像再处理,将源数据集进行裁剪和压缩,得到133对清晰和模糊图像对,并记为电力巡检去模糊数据集,该数据集是本发明独有的,目前电力巡检方向还未出现去模糊数据集,与传统公开数据集GoPro制作方法不同,GoPro数据集是采用十几张连续帧的清晰图像融合成一张模糊图像的方法,电力巡检去模糊数据集是采用单张模糊图像的压缩裁剪方法形成新的模糊图像;S2: Complete the image reprocessing of the source data set, crop and compress the source data set, and obtain 133 pairs of clear and blurred images, which are recorded as the power inspection deblurred data set. This data set is unique to the present invention. , At present, there is no deblurred data set in the direction of power inspection. Different from the traditional public data set GoPro production method, the GoPro data set uses a method of merging clear images of more than a dozen consecutive frames into one blurred image. Deblurring of power inspection The data set uses the compression and cropping method of a single blurred image to form a new blurred image;

S3:搭建多尺度改进残差块卷积神经网络,并使用电力巡检去模糊数据集对多尺度改进残差块卷积神经网络进行充分训练,得到训练好的多尺度改进残差块卷积神经网络模型,该模型算法与传统多尺度卷积神经网络不同之处在于网络残差结构的改进和损失函数的构造;S3: Build a multi-scale improved residual block convolutional neural network, and use the power inspection deblurred data set to fully train the multi-scale improved residual block convolutional neural network, and obtain the trained multi-scale improved residual block convolution Neural network model. The difference between this model algorithm and the traditional multi-scale convolutional neural network lies in the improvement of the network residual structure and the construction of the loss function;

S4:使用训练好的多尺度改进残差块卷积神经网络模型对电力巡检模糊图像数据进行去模糊,再对去模糊图像进行目标检测和识别,判断电力巡检目标是否为缺陷状态,并且在后台和云端记录。S4: Use the trained multi-scale improved residual block convolutional neural network model to deblur the power inspection blurred image data, and then perform target detection and recognition on the deblurred image to determine whether the power inspection target is in a defective state, and Record in the background and in the cloud.

在本实施例中,电力巡检机器人的坐标系统如图3所示,参照图3,对电力巡检数据进行采集,图像预处理和数据集制作。In this embodiment, the coordinate system of the power inspection robot is shown in Figure 3. Referring to Figure 3, the power inspection data is collected, image preprocessed and data sets are produced.

本实施例中步骤S1具体为:Step S1 in this embodiment is specifically:

A1:在输电线路巡检机器人晃动状态下,对巡检机器人的电力巡检视频进行数据采集,得到并保存输电线路巡检机器人晃动状态下电力巡检视频数据,电力巡检视频包括同一场景下的清晰图像和模糊图像;A1: When the transmission line inspection robot is shaking, collect data from the power inspection video of the inspection robot, and obtain and save the power inspection video data when the transmission line inspection robot is shaking. The power inspection video includes the same scene. clear images and blurred images;

A2:采用视频分帧处理方法对步骤A1采集得到的巡检视频进行分帧处理操作,得到分帧处理后的图像,然后进行图像样本筛选,得到符合要求的图像;A2: Use the video frame processing method to perform frame processing on the inspection video collected in step A1 to obtain the framed image, and then screen the image samples to obtain images that meet the requirements;

A3:采用直方图均衡化方法对步骤A2采集得到的图像进行预处理操作,得到预处理后的图像数据;A3: Use the histogram equalization method to preprocess the image collected in step A2 to obtain the preprocessed image data;

A4:对步骤A3得到的预处理后的数据进行分类和配对操作,得到一一对应的清晰和模糊图像对,并记为源数据集。A4: Classify and pair the preprocessed data obtained in step A3 to obtain one-to-one corresponding pairs of clear and blurred images, which are recorded as source data sets.

本实施例中步骤A1包括:Step A1 in this embodiment includes:

在输电线路巡检机器人晃动状态下,对巡检机器人的电力巡检视频进行数据采集,由巡检机器人在不同巡检线路位置,不同焦距下,不同场景下拍摄,得到一系列帧率为16帧/S的短视频,并保存。When the transmission line inspection robot is shaking, the data of the power inspection video of the inspection robot is collected. The inspection robot shoots at different inspection line positions, different focal lengths, and different scenes to obtain a series of frame rates of 16 Frame/S short video and save it.

本实施例中步骤A2具体为:Step A2 in this embodiment is specifically:

由步骤A1采集到的视频,逐个进行分帧处理,得到16*t张图像,其中t为视频的时间长度;The videos collected in step A1 are processed into frames one by one to obtain 16*t images, where t is the time length of the video;

将分帧处理后的图像,进行样本筛选,选择包含电力巡检目标的图像,丢弃与电力巡检目标无关的图像。Perform sample screening on the frame-processed images, select images that contain the power inspection target, and discard images that have nothing to do with the power inspection target.

本实施例中步骤A3具体为:Step A3 in this embodiment is specifically:

采用直方图均衡化方法【可参考冈萨雷斯等.数字图像处理[M].电子工业出版社,2011:72-85.】,对步骤A2所得图像进行预处理,用于增强图像局部的对比度而不影响整体的对比度;图像直方图定义:一个灰度级在范围[0,L-1]的数字图像,该图像的直方图是一个离散函数:The histogram equalization method [please refer to Gonzalez et al. Digital Image Processing [M]. Electronic Industry Press, 2011: 72-85.] is used to preprocess the image obtained in step A2 to enhance the local image. Contrast without affecting the overall contrast; image histogram definition: a digital image with a gray level in the range [0, L-1], the histogram of the image is a discrete function:

p(rk)=nk/n (1)p(r k )=n k /n (1)

其中p为概率函数,rk是第k个灰度级,k=0,1,2,…,L-1,L为灰度级总数,nk是图像中第k个灰度级的像素总数,n是图像的像素总数,继而得到预处理后的图像。where p is the probability function, r k is the k-th gray level, k = 0,1,2,...,L-1, L is the total number of gray levels, n k is the pixel of the k-th gray level in the image The total number, n is the total number of pixels in the image, and then the preprocessed image is obtained.

本实施例中步骤A4包括:Step A4 in this embodiment includes:

将预处理后的图像进行分类,分成清晰图像和模糊图像两大类;Classify the preprocessed images into two categories: clear images and blurred images;

对得到的清晰和模糊图像,以相同背景和相同巡检目标位置为条件,一一对应进行配对,得到源数据集A1The obtained clear and blurred images are paired in one-to-one correspondence on the condition of the same background and the same inspection target position, and the source data set A 1 is obtained.

本实施例中步骤S2具体为:Step S2 in this embodiment is specifically:

B1:将源数据集A1中的图像进行裁剪,原图像分辨率为2688*1520,采用固定大小相框移动的方法,裁剪包含电力巡检目标的区域,所得图像的分辨率,长度要大于256像素,宽度要大于256像素,清晰和模糊图像对的分辨率大小要完全相等;B1: Crop the image in the source data set A1 . The original image resolution is 2688*1520. Use the method of moving the fixed size photo frame to crop the area containing the power inspection target. The resolution and length of the resulting image must be greater than 256 Pixels, the width must be greater than 256 pixels, and the resolution sizes of clear and blurry image pairs must be exactly equal;

B2:将步骤B1所得的图像,进行图像压缩,保证图像长度要大于256像素,宽度要大于256像素,清晰和模糊图像对的分辨率大小要完全相等,由此得到133对清晰和模糊图像对,并记为电力巡检去模糊数据集A2B2: Carry out image compression on the image obtained in step B1 to ensure that the image length is greater than 256 pixels and the width is greater than 256 pixels. The resolutions of the clear and blurry image pairs must be completely equal, thus obtaining 133 pairs of clear and blurry image pairs. , and recorded as the power inspection deblurred data set A 2 .

本实施例中步骤S3具体为:Step S3 in this embodiment is specifically:

C1:搭建多尺度改进残差块卷积神经网络,并确定网络结构;C1: Build a multi-scale improved residual block convolutional neural network and determine the network structure;

参照图4和图5,多尺度改进残差块卷积神经网络的结构由3个不同大小尺度的神经网络结构组成,在每个尺度下生成一个清晰的图像,作为去模糊任务的子问题,该子问题将一张模糊图像和初始去模糊的结果(来自上一级尺度的下采样)作为输入,并推测在这一尺度下,这张清晰的图像为:Referring to Figures 4 and 5, the structure of the multi-scale improved residual block convolutional neural network consists of three neural network structures of different size scales, and generates a clear image at each scale as a sub-problem of the deblurring task. This sub-problem takes a blurred image and the initial deblurred result (downsampling from the previous scale) as input, and infers that at this scale, this clear image is:

Im,hm=Net(Bm,Im+1,hm+1;θ) (2)I m , h m =Net(B m ,I m+1 ,h m+1 ; θ) (2)

其中,Im为本级尺度下该层网络输出的去模糊图像,hm为本级尺度下的隐藏层特征,Net为本发明提出的多尺度改进残差块卷积神经网络,Bm为本级尺度下该层网络输入的模糊图像,Im+1为上一级尺度输出的去模糊图像,hm+1为上一级尺度的隐藏层特征,θ为待训练参数,m为本级尺度,m+1为上一级尺度。Among them, I m is the deblurred image output by the layer network at this level of scale, h m is the hidden layer feature at this level of scale, Net is the multi-scale improved residual block convolutional neural network proposed by the present invention, and B m is The blurred image input by the network of this layer at this level of scale, I m+1 is the deblurred image output by the previous level of scale, h m+1 is the hidden layer feature of the previous level of scale, θ is the parameter to be trained, and m is the original Level scale, m+1 is the upper level scale.

任一尺度下的神经网络结构相同,具体搭建方法如下:The neural network structure at any scale is the same. The specific construction method is as follows:

第一部分是输入模块,包括一个卷积层和三个改进残差块;其中的任意一个改进残差块,与传统意义的改进残差块结构不同,本发明利用拼接操作代替原有改进残差块的逐元素相加操作,包含两个卷积层和一个线性整流函数和一个跳层拼接;拼接操作相比于原有改进残差块的逐元素相加操作有以下三个优点:1.信息融合:通过将两个特征图在通道维度上连接,可以实现信息的融合和组合,使网络能够利用更多的上下文信息。2.表示能力增强:连接操作扩展了特征图的通道数,增加了网络的表示能力,有助于学习更复杂的特征表示。3.多尺度特征融合:通过连接具有不同空间分辨率的特征图可以实现多尺度的特征融合,提升模型对不同尺度目标的感知能力。第一部分的输入来自于公开数据集GoPro与电力巡检去模糊数据集A2,数据集里面的每张图像,经过OpenCV程序处理,得到256*256大小的局部图像,进而送到输入模块,输入第一个卷积层,该层选用了64个3*3的卷积核,步长为1,填充数量为1,卷积层的输出尺寸计算公式如下列公式所示:The first part is the input module, including a convolution layer and three improved residual blocks; any one of the improved residual blocks is different from the traditional improved residual block structure. This invention uses splicing operations to replace the original improved residual blocks. The element-by-element addition operation of the block includes two convolutional layers, a linear rectification function and a skip layer splicing; the splicing operation has the following three advantages compared to the original element-wise addition operation of the improved residual block: 1. Information fusion: By connecting two feature maps in the channel dimension, the fusion and combination of information can be achieved, allowing the network to utilize more contextual information. 2. Representation capability enhancement: The connection operation expands the number of channels of the feature map, increases the representation capability of the network, and helps to learn more complex feature representations. 3. Multi-scale feature fusion: Multi-scale feature fusion can be achieved by connecting feature maps with different spatial resolutions, improving the model's perception of targets at different scales. The input of the first part comes from the public data set GoPro and the power inspection deblurred data set A2 . Each image in the data set is processed by the OpenCV program to obtain a partial image of 256*256 size, and then sent to the input module. The first convolution layer uses 64 3*3 convolution kernels, the step size is 1, and the number of padding is 1. The output size calculation formula of the convolution layer is as follows:

其中,Z是卷积输出数据的长度,W是卷积输入数据的长度,P是填充数量,F是卷积核的长度,S表示步长;对于第一部分中的第一个编码块,第一个卷积层的输出尺寸由计算公式(3)计算得到第一个卷积层的输出大小是256*256*64;由于第一部分的卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第一部分的卷积层和三个改进残差块的最终输出大小也是256*256*64;Among them, Z is the length of the convolution output data, W is the length of the convolution input data, P is the number of padding, F is the length of the convolution kernel, and S represents the step size; for the first coding block in the first part, the The output size of a convolution layer is calculated by formula (3) and the output size of the first convolution layer is 256*256*64; since the convolution kernel size and format used in the first part of the convolution layer are the same, the parameters Same, and so on, the final output size of the first part of the convolutional layer and the three improved residual blocks is also 256*256*64;

第二部分由两个编码块组成的编码结构,其中任意一个编码块包含一个卷积层,一个线性整流函数和三个改进残差块;对于第二部分中的第一个编码块,该层选用了128个3*3的卷积核,步长为2,填充数量为0.5,第一个卷积层的输出尺寸计算公式(3),计算得到第一个卷积层的输出大小是128*128*128;在编码块的第一个卷积层后使用线性整流函数作为激活函数,将经过激活函数的数据送入编码块的第二个卷积层,第二个卷积层采用64个3*3卷积核,步长为2,填充数量为0.5,则根据卷积层的输出尺寸计算公式(3),第二部分第二个卷积层的输出大小也是128*128*128;由于第二部分的卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第二部分的第一个编码块最终输出大小是128*128*128;同理,第二部分的第二个编码块,该层选用了256个3*3的卷积核,步长为2,填充数量为0.5,第一个卷积层的输出尺寸由计算公式(3)计算得到第一个卷积层的输出大小是64*64*256;由于第二部分的卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第二部分的第二个编码块最终输出大小是64*64*256;The second part is a coding structure composed of two coding blocks, where any coding block contains a convolutional layer, a linear rectification function and three improved residual blocks; for the first coding block in the second part, this layer 128 3*3 convolution kernels are selected, the step size is 2, and the padding amount is 0.5. The output size of the first convolution layer is calculated using the formula (3). The calculated output size of the first convolution layer is 128 *128*128; After the first convolutional layer of the coding block, a linear rectification function is used as the activation function, and the data after the activation function is sent to the second convolutional layer of the coding block. The second convolutional layer uses 64 A 3*3 convolution kernel, the step size is 2, and the padding amount is 0.5. According to the output size calculation formula (3) of the convolution layer, the output size of the second convolution layer in the second part is also 128*128*128 ; Since the convolutional layer in the second part uses the same convolution kernel size and format, the same parameters, and so on, the final output size of the first coding block in the second part is 128*128*128; similarly, the second For the second coding block of the part, this layer uses 256 3*3 convolution kernels, the step size is 2, and the number of padding is 0.5. The output size of the first convolution layer is calculated by the calculation formula (3). The output size of a convolutional layer is 64*64*256; since the second part of the convolutional layer uses the same convolution kernel size and format, the same parameters, and so on, the second coding block of the second part finally The output size is 64*64*256;

第三部分由两个解码块组成,其中任意一个解码块由三个改进残差块,一个反卷积层和一个线性整流函数组成;反卷积层的输出尺寸计算公式如下列公式所示:The third part consists of two decoding blocks, any of which is composed of three improved residual blocks, a deconvolution layer and a linear rectification function; the output size calculation formula of the deconvolution layer is as follows:

Z1=(W1-1)×S1-2×P1+F1 (4)Z 1 =(W 1 -1)×S 1 -2×P 1 +F 1 (4)

其中,Z1是卷积输出数据的长度,W1是卷积输入数据的长度,P1是填充数量,F1是卷积核的长度,S1表示步长;根据反卷积层的输出尺寸计算公式(4),该层选用了256个3*3的卷积核,步长为2,填充数量为0.5,第三部分第一个反卷积层的输出大小是128*128*256;由于第三部分的反卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第三部分的第一个解码块最终输出大小是128*128*256;同理,第三部分的第二个解码块,该层选用了128个3*3的卷积核,步长为2,填充数量为0.5,第一个反卷积层的输出尺寸由计算公式(4)计算得到第一个反卷积层的输出大小是256*256*128;由于第二个解码块的反卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第三部分的第二个解码块最终输出大小是256*256*128;Among them, Z 1 is the length of the convolution output data, W 1 is the length of the convolution input data, P 1 is the number of padding, F 1 is the length of the convolution kernel, and S 1 represents the step size; according to the output of the deconvolution layer Size calculation formula (4), this layer uses 256 3*3 convolution kernels, the step size is 2, and the number of padding is 0.5. The output size of the first deconvolution layer in the third part is 128*128*256 ; Since the deconvolution layer in the third part uses the same convolution kernel size and format, the same parameters, and so on, the final output size of the first decoding block in the third part is 128*128*256; similarly, the The second decoding block of the three parts, this layer uses 128 3*3 convolution kernels, the stride is 2, and the number of padding is 0.5. The output size of the first deconvolution layer is calculated by formula (4) The output size of the first deconvolution layer is 256*256*128; since the deconvolution layer of the second decoding block uses the same convolution kernel size and format, the same parameters, and so on, the third part The final output size of the second decoding block is 256*256*128;

第四部分是输出模块,包括三个改进残差块和一个卷积层;该层选用了64个3*3的卷积核,步长为1,填充数量为1,第四部分第一个卷积层的输出大小是256*256*64;由于第四部分的卷积层采用的卷积核大小和格式相同,参数相同,依此类推,第四部分的解码块最终输出大小是256*256*64。The fourth part is the output module, including three improved residual blocks and a convolution layer; this layer uses 64 3*3 convolution kernels, the step size is 1, and the number of padding is 1. The first part of the fourth part The output size of the convolution layer is 256*256*64; since the convolution kernel size and format used in the fourth part of the convolution layer are the same, the parameters are the same, and so on, the final output size of the decoding block in the fourth part is 256* 256*64.

C2:选择网络优化器并使用公开数据集GoPro与电力巡检去模糊数据集A2对多尺度改进残差块卷积神经网络进行训练,得到训练好的多尺度改进残差块卷积神经网络模型;C2: Select the network optimizer and use the public data set GoPro and the power inspection deblurred data set A2 to train the multi-scale improved residual block convolutional neural network to obtain the trained multi-scale improved residual block convolutional neural network. Model;

本实施例中步骤C2具体为:Step C2 in this embodiment is specifically:

将公开数据集GoPro与电力巡检去模糊数据集A2按照4:1的比例划分为训练集和测试集,在训练的过程中,损失函数使用四损失权重平衡函数,四损失权重平衡函数定义:The public data set GoPro and the power inspection deblurred data set A2 are divided into training sets and test sets according to the ratio of 4:1. During the training process, the loss function uses a four-loss weight balance function, and the four-loss weight balance function is defined :

L=Lp+k1L1+k2L2+k3L3 (5)L=L p +k 1 L 1 +k 2 L 2 +k 3 L 3 (5)

其中,L为四损失权重平衡函数,k1,k2与k3为融合系数,本发明中k1,k2与k3都取0.3数值大小,Lp为感知损失函数,L1为L1范数损失函数,L2为L2范数损失函数,L3为多尺度频率重建损失函数,φi,j是VGG19网络中激活之后的第j个卷积在第i层最大池化之前所取得的特征映射,Wi,j是特征映射的宽度,Hi,j是特征映射的高度,S代表清晰图像,I代表生成的去模糊图像,x是特征映射宽度方向的坐标值,y是特征映射高度方向的坐标值,n为尺度的批次,In为本网络中第n尺度输出的去模糊图像,Sn为本网络中第n尺度输出的清晰图像,tk表示第k个尺度下的元素总数,||·||1是L1距离计算符,||·||2是L2距离计算符,F表示将图像信号传输到频域的快速傅立叶变换(FFT);Among them, L is the four-loss weight balance function, k 1 , k 2 and k 3 are fusion coefficients. In the present invention, k 1 , k 2 and k 3 all take a numerical value of 0.3, L p is the perceptual loss function, and L 1 is L1 Norm loss function, L 2 is the L2 norm loss function, L 3 is the multi-scale frequency reconstruction loss function, φ i,j is the j-th convolution after activation in the VGG19 network before the i-th layer maximum pooling. The feature map of Mapping the coordinate value in the height direction, n is the batch of scales, I n is the deblurred image output by the nth scale in this network, S n is the clear image output by the nth scale in this network, t k represents the kth scale The total number of elements under , ||·|| 1 is the L1 distance calculator, ||·|| 2 is the L2 distance calculator, and F represents the fast Fourier transform (FFT) that transmits the image signal to the frequency domain;

本实施例中网络优化器采用亚当优化器【亚当优化器,英文名为AdamOptimizer,可参考杨观赐,杨静,李少波,胡建军.基于Dropout与ADAM优化器的改进CNN算法[J].华中科技大学学报(自然科学版),2018,46(07):122-127.】,训练批次尺寸为24,学习率为对5×10-5。对网络进行充分训练,得到训练好的多尺度改进残差块卷积神经网络模型。In this embodiment, the network optimizer uses the Adam optimizer [Adam optimizer, English name is AdamOptimizer, please refer to Yang Guanci, Yang Jing, Li Shaobo, Hu Jianjun. Improved CNN algorithm based on Dropout and ADAM optimizer [J]. Journal of Huazhong University of Science and Technology (Natural Science Edition), 2018,46(07):122-127.], the training batch size is 24, and the learning rate is 5×10 -5 . The network is fully trained to obtain the trained multi-scale improved residual block convolutional neural network model.

本实施例中步骤S4包括:Step S4 in this embodiment includes:

D1:实时采集电力巡检机器人巡检过程中拍摄的巡检图像数据,使用步骤S3得到的训练好的模型对实时采集的图像数据进行去模糊,得到巡检目标当前的状态情况;D1: Collect the inspection image data captured during the inspection process of the power inspection robot in real time, and use the trained model obtained in step S3 to deblur the image data collected in real time to obtain the current status of the inspection target;

D2:根据步骤D1得到的巡检目标去模糊图像,进行图像的目标检测和识别,若出现巡检目标出现缺陷和损坏,后台和云端发出警报并进行数据记录。D2: Deblur the image based on the inspection target obtained in step D1, and perform target detection and recognition of the image. If the inspection target is defective or damaged, the background and cloud will issue an alarm and record the data.

为了验证上述本发明方案的有效性,本实施例以天津国网巡检验收项目为例,如图6所示,机器人现场拍摄的模糊图像和去模糊图像之后的对比,发现红色框内区域的部分,去模糊效果尤为明显,该部分属于金具类关键部位,是电力巡检目标的重要部分。说明本发明提供的多尺度改进残差块卷积神经网络对电力巡检去运动模糊具有很强的适应性和优越性。In order to verify the effectiveness of the above-mentioned solution of the present invention, this embodiment takes the Tianjin State Grid inspection and acceptance project as an example. As shown in Figure 6, the comparison between the blurred image taken by the robot on site and the deblurred image shows that the area in the red box is Part, the deblurring effect is particularly obvious. This part is a key part of hardware and an important part of the power inspection target. It shows that the multi-scale improved residual block convolutional neural network provided by the present invention has strong adaptability and superiority for power inspection to remove motion blur.

通过上述技术方案的实施,可见本发明的优点可以总结为如下几点:Through the implementation of the above technical solutions, it can be seen that the advantages of the present invention can be summarized as follows:

(1)提供了电力巡检机器人的数据采集方法及数据预处理过程,包括:分帧处理、裁剪压缩、图形配对与直方图均衡化。(1) Provides the data collection method and data preprocessing process of the power inspection robot, including: frame processing, cropping and compression, graphics pairing and histogram equalization.

(2)提供了数据集的制作方法以及多尺度改进残差块卷积神经网络的搭建方法,利用清晰和模糊图像对之间的差异,来完成对电力巡检图像的去模糊。(2) Provides a method for creating a data set and a method for building a multi-scale improved residual block convolutional neural network, using the difference between pairs of clear and blurred images to complete the deblurring of power inspection images.

(3)提供了基于多尺度改进残差块CNN的图像去运动模糊方法,该方法有运算速度快、去运动模糊效果好、抗环境干扰能力强的优点。(3) An image de-motion blur method based on multi-scale improved residual block CNN is provided. This method has the advantages of fast operation speed, good motion blur removal effect, and strong resistance to environmental interference.

Claims (10)

1. An image motion blur removing method based on a multi-scale improved residual block CNN is characterized by comprising the following steps:
s1: collecting and storing electric power inspection video data of the electric power line inspection robot, framing the collected electric power inspection video data, then screening an image sample, preprocessing an image to obtain a clear and fuzzy image pair, preparing a data set, and marking the data set as a source data set;
s2: clipping and compressing the source data set to obtain a clear and fuzzy image pair, and recording the clear and fuzzy image pair as a power inspection deblurring data set;
s3: building a multi-scale improved residual block convolutional neural network, and fully training the multi-scale improved residual block convolutional neural network by using a power inspection deblurring data set to obtain a trained multi-scale improved residual block convolutional neural network model;
s4: and deblurring the power inspection blurred image data by using a trained multi-scale improved residual block convolutional neural network model.
2. The image motion blur removal method based on the multi-scale improved residual block CNN according to claim 1, wherein the step S1 is specifically:
a1: in the shaking state of the power transmission line inspection robot, acquiring data of an electric power inspection video of the inspection robot to obtain and store electric power inspection video data of the power transmission line inspection robot in the shaking state, wherein the electric power inspection video comprises a clear image and a blurred image in the same scene;
a2: carrying out framing processing operation on the inspection video acquired in the step A1 by adopting a video framing processing method to obtain a frame processed image, and then carrying out image sample screening to obtain an image meeting the requirements;
a3: performing preprocessing operation on the image acquired in the step A2 by adopting a histogram equalization method to obtain preprocessed image data;
a4: and (3) classifying and pairing the preprocessed data obtained in the step (A3) to obtain clear and fuzzy image pairs corresponding to each other one by one, and recording the clear and fuzzy image pairs as a source data set.
3. The image motion blur removal method based on the multi-scale improved residual block CNN according to claim 2, wherein the step A2 is specifically:
carrying out frame-dividing treatment on the videos acquired in the step A1 one by one to obtain 16 x t images, wherein t is the time length of the videos;
and carrying out sample screening on the image subjected to framing treatment, selecting an image containing the power inspection target, and discarding the image irrelevant to the power inspection target.
4. The image motion blur removing method based on the multi-scale improved residual block CNN according to claim 2, wherein the step A3 is specifically:
preprocessing the image obtained in the step A2 by adopting a histogram equalization method, wherein the preprocessing is used for enhancing the local contrast of the image without affecting the overall contrast; image histogram definition: a digital image having a gray level in the range 0, l-1, the histogram of which is a discrete function:
p(r k )=n k /n (1)
where p is a probability function, r k Is the kth gray level, k=0, 1,2, …, L-1, L is the total number of gray levels, n k Is the total number of pixels in the image for the kth gray level, n is the total number of pixels in the image, and then a preprocessed image is obtained.
5. The image motion blur removal method based on the multi-scale improved residual block CNN according to claim 1, wherein the step S2 is specifically:
b1: will source data set A 1 Cutting the image in the image, wherein the resolution of the original image is 2688 x 1520, a fixed-size photo frame moving method is adopted to cut the area containing the power inspection target, the resolution of the obtained image is greater than 256 pixels in length and greater than 256 pixels in width, and the resolution of the clear and blurred image pairs is completely equal;
b2: b1, performing image compression on the image obtained in the step B1, ensuring that the length of the image is greater than 256 pixels, the width of the image is greater than 256 pixels, and the resolution of the clear and fuzzy image pairs are completely equal, so that the clear and fuzzy image pairs are obtained and are recorded as a power inspection deblurring data set A 2
6. The image motion blur removal method based on the multi-scale improved residual block CNN according to claim 5, wherein the step S3 is specifically:
c1: constructing a multi-scale improved residual block convolutional neural network, and determining a network structure;
c2: selecting a network optimizer and defuzzifying data set A using public data set GoPro and power inspection 2 Training the multi-scale improved residual block convolutional neural network to obtain a trained multi-scale improved residual block convolutional neural network model.
7. The image deblurring method based on the multi-scale improved residual block CNN according to claim 6, wherein the structure of the multi-scale improved residual block convolutional neural network in step C1 is composed of 3 different-sized-scale neural network structures, a clear image is generated at each scale, and the clear image is taken as a sub-problem of the deblurring task, the sub-problem takes as input a blurred image and an initial deblurring result, and it is presumed that at this scale, the clear image is:
I m ,h m =Net(B m ,I m+1 ,h m+1 ;θ) (2)
wherein I is m For the deblurred image output by the layer of network under the scale of the present level, h m For the hidden layer characteristics under the present scale, net is the multi-scale improved residual block convolution neural network provided by the invention, B m For the fuzzy image input by the layer of network under the scale of the present level, I m+1 A deblurred image output for the previous scale, h m+1 And (3) as the hidden layer characteristics of the upper-level scale, θ is a parameter to be trained, m is the current-level scale, and m+1 is the upper-level scale.
8. The image motion blur removing method based on the multi-scale improved residual block CNN according to claim 7, wherein the neural network structure at any scale in the step C1 is the same, and the specific construction method is as follows:
the first part is an input module comprising a convolutional layer and three modified residual blocks;
the second part is a coding structure formed by two coding blocks, wherein any one coding block comprises a convolution layer, a linear rectification function and three improved residual blocks;
the third part consists of two decoding blocks, wherein any one decoding block consists of three improved residual blocks, a deconvolution layer and a linear rectification function;
the fourth part is the output module, comprising three modified residual blocks and one convolutional layer.
9. The image deblurring method based on multi-scale improved residual block CNN according to claim 8, wherein the first portion is the inputIn the module, the invention replaces the element-by-element addition operation of the original improved residual block by using the splicing operation, and comprises two convolution layers, a linear rectification function and a layer jump splicing; the first part of input comes from the public data set GoPro and the power inspection deblurring data set A 2 Each image in the data set is processed by an OpenCV program to obtain a local image with 256 x 256 sizes, the local image is sent to an input module, a first convolution layer is input, the convolution kernels of 64 3*3 are selected, the step size is 1, the filling quantity is 1, and the calculation formula of the output size of the convolution layer is shown as the following formula:
wherein Z is the length of the convolution output data, W is the length of the convolution input data, P is the filling number, F is the length of the convolution kernel, and S represents the step size; for a first code block in the first part, calculating the output size of the first convolution layer by a calculation formula (3) to obtain that the output size of the first convolution layer is 256×256×64; since the convolution kernel adopted by the convolution layer of the first part has the same size and format, the parameters are the same, and so on, the final output sizes of the convolution layer of the first part and the three improved residual blocks are 256×256×64;
in the coding structure of the second part, for the first coding block in the second part, 128 convolution kernels of 3*3 are selected for the layer, the step size is 2, the number of fills is 0.5, the output size of the first convolution layer calculates equation (3), calculating to obtain a first convolution layer the output size of (2) is 128 x 128; using a linear rectification function as an activation function after the first convolution layer of the coding block, feeding the data subjected to the activation function into a second convolution layer of the coding block, the second convolution layer employing 64 3*3 convolution kernels, step length is 2, the filling quantity is 0.5, and according to the output size calculation formula (3) of the convolution layer, second part of the second convolution layer the output size is also 128×128×128; since the convolution kernel employed by the second portion of the convolution layer is the same size and format, the parameters are the same, and so on, the final output size of the first encoded block of the second portion is 128 x 128; similarly, the second coding block of the second part selects 256 convolution kernels of 3*3, the step length is 2, the filling quantity is 0.5, and the output size of the first convolution layer is calculated by a calculation formula (3) to obtain that the output size of the first convolution layer is 64×64×256; because the convolution kernel adopted by the convolution layer of the second part has the same size and format, the parameters are the same, and so on, the final output size of the second coding block of the second part is 64 x 256;
in the two decoding blocks of the third portion, the calculation formula of the output size of the deconvolution layer is shown as the following formula:
Z 1 =(W 1 -1)×S 1 -2×P 1 +F 1 (4)
wherein Z is 1 Is the length of the convolution output data, W 1 Is the length of the convolved input data, P 1 Is the filling quantity, F 1 Is the length of the convolution kernel, S 1 Representing a step size; according to an output size calculation formula (4) of the deconvolution layer, 256 convolution kernels of 3*3 are selected in the deconvolution layer, the step length is 2, the filling quantity is 0.5, and the output size of the first deconvolution layer of the third part is 128×128×256; because the convolution kernel adopted by the deconvolution layer of the third part has the same size and format, the parameters are the same, and so on, the final output size of the first decoding block of the third part is 128 by 256; similarly, the second decoding block of the third part selects 128 convolution kernels of 3*3, the step length is 2, the filling quantity is 0.5, the output size of the first deconvolution layer is calculated by a calculation formula (4), and the output size of the first deconvolution layer is 256×256×128; because the convolution kernel adopted by the deconvolution layer of the second decoding block has the same size and format, the parameters are the same, and so on, the final output size of the second decoding block of the third part is 256×256×128;
the output module of the fourth part selects 64 3*3 convolution kernels, the step length is 1, the filling quantity is 1, and the output size of the first convolution layer of the fourth part is 256×256×64; since the convolution kernel size and format adopted by the convolution layer of the fourth portion are the same, the parameters are the same, and so on, the final output size of the decoding block of the fourth portion is 256×256×64.
10. The image motion blur removal method based on the multi-scale improved residual block CNN according to claim 7, wherein the step C2 is specifically:
deblurring the public dataset GoPro and the Power inspection data set A 2 Dividing the training set and the testing set according to the ratio of 4:1, wherein in the training process, the loss function uses a four-loss weight balance function, and the four-loss weight balance function is defined as follows:
L=L p +k 1 L 1 +k 2 L 2 +k 3 L 3 (5)
wherein L is a four-loss weight balance function, k 1 ,k 2 And k is equal to 3 L is the fusion coefficient p To perceive the loss function, L 1 As L1 norm loss function, L 2 As L2 norm loss function, L 3 Reconstructing a loss function, phi, for multi-scale frequencies i,j Is the feature map, W, taken by the jth convolution after activation in VGG19 networks before the ith layer maximum pooling i,j Is the width of the feature map, H i,j Is the height of the feature map, S representsA clear image, I represents the generated deblurred image, x is the coordinate value of the width direction of the feature map, y is the coordinate value of the height direction of the feature map, n is the scale batch, I n S for deblurring the image output by the nth scale in the network n T is the clear image output by the n-th scale in the network k Representing the total number of elements at the kth scale, I.I 1 Is an L1 distance calculator that is used to calculate the distance between the two devices, I.I 2 Is an L2 distance calculator, F represents a fast fourier transform that transmits an image signal to a frequency domain;
the network optimizer adopts Adam optimizer, training batch size is 24, and learning rate is 5×10 -5 And fully training the network to obtain a trained multi-scale improved residual block convolutional neural network model.
CN202311444732.5A 2023-11-01 2023-11-01 Image de-motion blur method based on multi-scale improved residual block CNN Pending CN117391973A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311444732.5A CN117391973A (en) 2023-11-01 2023-11-01 Image de-motion blur method based on multi-scale improved residual block CNN

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311444732.5A CN117391973A (en) 2023-11-01 2023-11-01 Image de-motion blur method based on multi-scale improved residual block CNN

Publications (1)

Publication Number Publication Date
CN117391973A true CN117391973A (en) 2024-01-12

Family

ID=89464610

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311444732.5A Pending CN117391973A (en) 2023-11-01 2023-11-01 Image de-motion blur method based on multi-scale improved residual block CNN

Country Status (1)

Country Link
CN (1) CN117391973A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN118918364A (en) * 2024-07-15 2024-11-08 哈尔滨工业大学 Visual identification method for pipe gallery bolt missing under robot moving shooting condition

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN118918364A (en) * 2024-07-15 2024-11-08 哈尔滨工业大学 Visual identification method for pipe gallery bolt missing under robot moving shooting condition

Similar Documents

Publication Publication Date Title
CN113362223B (en) Image Super-Resolution Reconstruction Method Based on Attention Mechanism and Two-Channel Network
CN110084165B (en) Intelligent identification and early warning method for abnormal events in open scene of power field based on edge calculation
CN110751630B (en) Power transmission line foreign matter detection method and device based on deep learning and medium
CN114494050B (en) A self-supervised video deblurring and image interpolation method based on event camera
CN111339902B (en) A method and device for identifying digital display numbers on an LCD screen of a digital display instrument
CN112396635B (en) A multi-target detection method in complex environment based on multi-device
CN112070728A (en) Steel rope surface defect identification method based on target detection algorithm
CN116129291A (en) Unmanned aerial vehicle animal husbandry-oriented image target recognition method and device
CN117727046A (en) Novel mountain torrent front-end instrument and meter reading automatic identification method and system
CN117854036A (en) Water surface obstacle detection method based on improved YOLOv3
CN117391973A (en) Image de-motion blur method based on multi-scale improved residual block CNN
CN112990215A (en) Image denoising method, device, equipment and storage medium
CN112308087A (en) Integrated imaging recognition system and method based on dynamic vision sensor
CN118865351B (en) A road parking license plate recognition method and system based on 5G neural network
CN115424017A (en) Building internal and external contour segmentation method, device and storage medium
CN118410724B (en) Transmission line foreign matter identification method, system, computer equipment and medium
CN117593702B (en) Remote monitoring method, device, equipment and storage medium
CN112418229A (en) A real-time segmentation method of unmanned ship marine scene images based on deep learning
CN117745709A (en) Railway foreign matter intrusion detection method, system, equipment and medium
CN114494931B (en) A method and system for intelligent classification and processing of video image faults
CN117893795A (en) GIS equipment inner surface metal particle detection method based on improved YOLOv5
CN117523311A (en) Three-way pipe detection method and system based on improved YOLOv7
CN116597503A (en) Classroom behavior detection method based on space-time characteristics
CN111160255B (en) Fishing behavior identification method and system based on three-dimensional convolution network
CN111127587B (en) Reference-free image quality map generation method based on countermeasure generation network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination