CN112085024A - A method for character recognition on the surface of a tank - Google Patents
A method for character recognition on the surface of a tank Download PDFInfo
- Publication number
- CN112085024A CN112085024A CN202010998223.7A CN202010998223A CN112085024A CN 112085024 A CN112085024 A CN 112085024A CN 202010998223 A CN202010998223 A CN 202010998223A CN 112085024 A CN112085024 A CN 112085024A
- Authority
- CN
- China
- Prior art keywords
- image
- character
- text
- training
- model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 46
- 238000013528 artificial neural network Methods 0.000 claims abstract description 36
- 239000008267 milk Substances 0.000 claims abstract description 35
- 210000004080 milk Anatomy 0.000 claims abstract description 35
- 235000013336 milk Nutrition 0.000 claims abstract description 35
- 239000000843 powder Substances 0.000 claims abstract description 35
- 238000001514 detection method Methods 0.000 claims abstract description 32
- 102100032202 Cornulin Human genes 0.000 claims abstract description 17
- 101000920981 Homo sapiens Cornulin Proteins 0.000 claims abstract description 17
- 238000012937 correction Methods 0.000 claims abstract description 11
- 238000012549 training Methods 0.000 claims description 48
- 238000012360 testing method Methods 0.000 claims description 23
- 230000009466 transformation Effects 0.000 claims description 14
- 230000008569 process Effects 0.000 claims description 12
- 238000012545 processing Methods 0.000 claims description 11
- 238000013527 convolutional neural network Methods 0.000 claims description 10
- 238000007781 pre-processing Methods 0.000 claims description 8
- 239000011159 matrix material Substances 0.000 claims description 7
- 238000010606 normalization Methods 0.000 claims description 5
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 claims description 3
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 3
- 239000006185 dispersion Substances 0.000 claims description 3
- 238000003064 k means clustering Methods 0.000 claims description 3
- 238000013507 mapping Methods 0.000 claims description 3
- 238000012805 post-processing Methods 0.000 claims description 3
- 238000004088 simulation Methods 0.000 claims description 3
- 238000013518 transcription Methods 0.000 claims description 3
- 230000035897 transcription Effects 0.000 claims description 3
- 238000012216 screening Methods 0.000 claims 1
- 230000011218 segmentation Effects 0.000 claims 1
- 238000000844 transformation Methods 0.000 claims 1
- 238000004519 manufacturing process Methods 0.000 abstract description 3
- 230000006870 function Effects 0.000 description 5
- 230000008859 change Effects 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 4
- 230000007547 defect Effects 0.000 description 3
- 238000013461 design Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 238000009776 industrial production Methods 0.000 description 3
- 238000011176 pooling Methods 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 2
- 229930091051 Arenine Natural products 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 238000011897 real-time detection Methods 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 239000013598 vector Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/14—Image acquisition
- G06V30/148—Segmentation of character regions
- G06V30/153—Segmentation of character regions using recognition of characters or words
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/62—Text, e.g. of license plates, overlay texts or captions on TV images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
- Character Discrimination (AREA)
Abstract
Description
技术领域technical field
本发明涉及一种自动化生产线上字符识别领域,特别是一种基于YOLO-V3与改进CRNN的字符识别方法。The invention relates to the field of character recognition on automatic production lines, in particular to a character recognition method based on YOLO-V3 and improved CRNN.
背景技术Background technique
在实际工业生产的过程中,字符识别技术是非常重要的一个方面,字符识别技术已经被成功的应用于工业生产与组装,如许多小型电子元器件、电路板和一些大型零件表面的字符喷码和标签,厂商可通过这些喷码和标签来辨认追踪产品信息。而由于人眼难以做到长时间的高效率检测,某些高温高压等恶劣的作业环境更给人工检测识别带来了困难,因此,采用机器视觉技术进行自动准确的字符定位与识别成为工业生产流程中的一个重要环节。In the process of actual industrial production, character recognition technology is a very important aspect. Character recognition technology has been successfully applied to industrial production and assembly, such as character printing on the surface of many small electronic components, circuit boards and some large parts. And labels, manufacturers can identify and track product information through these codes and labels. Since it is difficult for the human eye to achieve high-efficiency detection for a long time, some harsh operating environments such as high temperature and high pressure have brought difficulties to manual detection and recognition. Therefore, the use of machine vision technology for automatic and accurate character positioning and recognition has become an industrial production process. an important part of the process.
基于轮廓检测和边缘处理的图像处理方法也可以将字符区域提取,但是对每一种类型的图像都需要进行针对性的版面分析,轮廓的选取也受图像背景的影响很大,因而其泛化能力不强。使用深度卷积神经网络对版面中的文本信息进行定位提取,可以对各类证件的文本区域进行定位及提取,使字符识别定位的适用范围更广泛,对复杂背景的图像识别更强。Image processing methods based on contour detection and edge processing can also extract character regions, but each type of image requires targeted layout analysis, and the selection of contours is also greatly affected by the image background, so its generalization The ability is not strong. Using the deep convolutional neural network to locate and extract the text information in the layout can locate and extract the text area of various documents, so that the applicable scope of character recognition and localization is wider, and the image recognition of complex backgrounds is stronger.
当前应用于工业界的OCR识别方法主要存在以下问题:The current OCR identification methods used in the industry mainly have the following problems:
1)利用传统算法提取字符区域、分割字符文本时的适应性较差,对于亮度不均匀的成像、字符粘连、字符模糊等情况难以准确提取;1) Using traditional algorithms to extract character regions and segment character texts has poor adaptability, and it is difficult to accurately extract images with uneven brightness, character sticking, and character blurring;
2)需要针对提取到的字符人工设计提取特征的方式,并且检测新字体的通用性较差。所以提供一种基于YOLO-V3与CRNN的字符定位与识别的方法。2) It is necessary to manually design a feature extraction method for the extracted characters, and the generality of detecting new fonts is poor. Therefore, a method for character location and recognition based on YOLO-V3 and CRNN is provided.
发明内容SUMMARY OF THE INVENTION
为达到上述目的,本发明提供一种罐表面字符识别方法,包括:实时采集罐表面图像;利用预先建立的YOLO-V3神经网络检测模型对实时采集的罐表面图像进行字符定位,得到包括是否含有字符和字符位置信息;输出字符预测框位置,并裁剪出字符区域;裁剪出的字符区域存在与实际文本区域贴合不够紧密的问题,使用文本矫正矫正算法进行矫正;通过直方图投影,分割出每一行的文本字符;对裁剪出来的文本区域使用改进的端到端不定长文本CRNN模型进行识别;根据当时系统情况,如果是在线模式,识别结果上传到数据库;如果是离线模式,识别结果将保存在本地。In order to achieve the above purpose, the present invention provides a method for recognizing characters on the surface of a tank, which includes: collecting an image of the tank surface in real time; using a pre-established YOLO-V3 neural network detection model to perform character positioning on the image of the tank surface collected in real time, and obtaining information including whether it contains Character and character position information; output the position of the character prediction frame, and crop out the character area; the cropped character area is not close enough to the actual text area, use the text correction correction algorithm to correct it; through histogram projection, segment out Text characters of each line; use the improved end-to-end variable-length text CRNN model to identify the cropped text area; according to the system conditions at the time, if it is in online mode, the recognition result will be uploaded to the database; if it is in offline mode, the recognition result will be Save locally.
上述方案中,罐表面字符识别方法还包括:执行机构切换下一个待识别工件。In the above solution, the method for recognizing characters on the surface of the tank further includes: the actuator switches to the next workpiece to be recognized.
上述方案中,罐表面包括:奶粉罐表面,罐头罐表面。In the above solution, the surface of the can includes: the surface of the milk powder can and the surface of the can.
上述方案中,YOLO-V3神经网络检测模型的建立方法包括:In the above scheme, the establishment method of the YOLO-V3 neural network detection model includes:
S1:采集大量奶粉罐表面字符图像;S1: Collect a large number of character images on the surface of milk powder cans;
S2:对奶粉罐表面图像进行预处理;预处理包括对奶粉罐表面图像进行归一化处理:将等比例缩放后的奶粉罐表面图像填入预设像素大小的空白图像中,然后将空白图像上除奶粉罐图像外的像素全部填充为设定颜色;空白图像的像素大小为416*416,空白图像能够覆盖缩放后的奶粉罐图像;所述设定颜色包括灰色;S2: Preprocessing the surface image of the milk powder can; the preprocessing includes normalizing the surface image of the milk powder can: filling the proportionally scaled surface image of the milk powder can into a blank image with a preset pixel size, and then adding the blank image All pixels except the image of the milk powder can are filled with the set color; the pixel size of the blank image is 416*416, and the blank image can cover the zoomed image of the milk powder can; the set color includes gray;
S3:对预处理后的图像进行字符区域标定,得到各图像的字符位置标签;通过人工框选方式对预处理后的奶粉罐图像进行字符区域标定,将各标定区域的左上角点和右下角点坐标和相应区域的字符存储为字符标签;S3: Perform character area calibration on the preprocessed image to obtain character position labels of each image; perform character area calibration on the preprocessed milk powder can image by manual frame selection, and set the upper left corner and lower right corner of each calibration area Point coordinates and characters in the corresponding area are stored as character labels;
S4:将图像及其位置信息标签划分为训练集样本和测试集样本,划分比例为9:1;利用训练集样本和测试机样本,采用计算机图形处理器GPU对预先搭建的YOLO-V3深度神经网络进行训练,得到YOLO-V3神经网络检测模型。S4: Divide the image and its position information labels into training set samples and test set samples, with a division ratio of 9:1; use the training set samples and test machine samples, and use the computer graphics processor GPU to analyze the pre-built YOLO-V3 deep neural network. The network is trained to obtain the YOLO-V3 neural network detection model.
上述方案中,预处理还包括对归一化处理后得到的罐表面字符图像进行数据增强;罐表面字符图像进行数据增强包括:In the above solution, the preprocessing also includes data enhancement on the character images on the tank surface obtained after normalization; the data enhancement on the character images on the tank surface includes:
几何变换类:包括翻转,旋转,裁剪,变形,缩放;Geometric transformation class: including flipping, rotating, cropping, deforming, scaling;
颜色变换类:包括噪声、模糊、颜色变换、擦除、填充;Color transformation class: including noise, blur, color transformation, erase, fill;
根据图片样式模拟生成样本图片。Generate a sample image based on the image style simulation.
上述方案中,对预先搭建的YOLO-V3神经网络进行训练包括:In the above scheme, training the pre-built YOLO-V3 neural network includes:
S1:设置训练参数:将迭代步数epochs设置为10000,将学习率优化器optimizer设置为'adam',将批训练样本数batch_size设置为64,以及设置9个先验框anchor box的大小;S1: Set the training parameters: set the number of iteration steps epochs to 10000, set the learning rate optimizer optimizer to 'adam', set the number of batch training samples batch_size to 64, and set the size of the 9 a priori box anchor boxes;
S2:训练YOLO-V3神经网络:把训练集样本数据和各样本对应的奶粉罐表面字符数据输入到YOLO-V3神经网络中进行模型训练;S2: Train the YOLO-V3 neural network: input the training set sample data and the corresponding character data on the surface of the milk powder can into the YOLO-V3 neural network for model training;
S3:模型测试:每次训练结束后,利用测试集样本对训练得到的模型进行测试,若模型对测试集样本中的瑕疵检出率超过95%且检测准确率不低于95%,则将最后一次训练得到的模型作为最终YOLO-V3神经网络检测模型;否则将最后一次训练得到的模型作为当前待训练的YOLO-V3神经网络,重复步骤S2-S3,直至得到最终YOLO-V3神经网络检测模型。S3: Model test: After each training, use the test set samples to test the trained model. If the model has a defect detection rate of more than 95% in the test set samples and the detection accuracy is not less than 95%, the model will be tested. The model obtained from the last training is used as the final YOLO-V3 neural network detection model; otherwise, the model obtained from the last training is used as the current YOLO-V3 neural network to be trained, and steps S2-S3 are repeated until the final YOLO-V3 neural network detection model is obtained. Model.
上述方案中,先验框的大小通过k-means聚类方法设置,包括:随机选取9个字符区域矩形框对象作为训练集的聚类中心,根据各对象与各个聚类中心之间的距离,把各对象分配给距离最近的聚类中心,每分配一个样本,聚类中心重新计算一次,重复分配过程,直到没有聚类中心再发生变化,当前9个聚类中心的值即9个先验框的大小。In the above scheme, the size of the prior frame is set by the k-means clustering method, including: randomly selecting 9 character area rectangular frame objects as the clustering center of the training set, and according to the distance between each object and each clustering center, Allocate each object to the nearest cluster center, each time a sample is assigned, the cluster center is recalculated once, and the assignment process is repeated until no cluster center changes. The current values of the nine cluster centers are nine priors. the size of the box.
上述方案中,文本矫正算法包括:In the above scheme, the text correction algorithm includes:
S1:裁剪后的图像进行灰度化处理,字符为白色,背景为色;S1: The cropped image is grayscaled, the characters are white, and the background is colored;
S2:进行高斯模糊,使文本部分能连成一块,再对图像进行阈值处理得到二值图像;S2: Perform Gaussian blurring so that the text parts can be connected together, and then perform threshold processing on the image to obtain a binary image;
S3:使用最小矩形拟合,然后通过矩形面积大小,筛选得到文本区域部分的面积,此部分就是文本最小外接矩形;S3: Use the smallest rectangle to fit, and then filter the area of the text area by the size of the rectangle area, which is the smallest bounding rectangle of the text;
S4:得到最小外接矩形的中心点,对文本区域进行仿射变换来实现校正倾斜文本。S4: Obtain the center point of the minimum circumscribed rectangle, and perform affine transformation on the text area to correct the oblique text.
上述方案中,直方图水平投影算法包括:In the above scheme, the histogram horizontal projection algorithm includes:
S1:对步骤四得到矫正后的图像进行二值化;S1: Binarize the corrected image obtained in step 4;
S2:对二值化后的图像计算水平方向上像素和,绘制出横轴为图像纵轴坐标范围为0~图像宽度,纵轴为像素和的直方图;S2: Calculate the pixel sum in the horizontal direction for the binarized image, and draw a histogram with the horizontal axis as the image and the vertical axis as the coordinate range from 0 to the image width, and the vertical axis as the pixel sum;
S3:找出直方图谷底部分,此部分为分割线处;S3: Find the bottom part of the histogram, which is the dividing line;
S4:根据分割线,分割出每一行文本图像。S4: According to the dividing line, each line of text image is divided.
上述方案中,改进的端到端不定长文本CRNN模型进行识别包括:In the above scheme, the improved end-to-end variable-length text CRNN model for recognition includes:
S1:每一行文本图像灰度化处理,将三通道的图像转为单通道的灰度图;S1: Grayscale processing of each line of text image, converting the three-channel image into a single-channel grayscale image;
S2:长度比例缩放和填充,将输入图像的尺寸固定为32*width,按比例进行缩放或扩充,这样能够与训练的样本相类似;S2: Length scaling and padding, fixing the size of the input image to 32*width, scaling or expanding proportionally, so that it can be similar to the training samples;
S3:数据标签稀疏矩阵转化,对标签矩阵进行处理,将其转化为需要的数据格式;S3: Data label sparse matrix conversion, process the label matrix and convert it into the required data format;
S4:搭建模型,更改结构,将原来的CNN+RNN+CTC转录层的思想,更改为CNN+CTC,主干网络在VGG上做修改设计7层卷积层,5层池化层的小型卷积神经网络,并在中间层的加入两次批量正则化,避免模型梯度弥散,加速收敛;S4: Build the model, change the structure, change the original idea of CNN+RNN+CTC transcription layer to CNN+CTC, and modify the backbone network on VGG to design a 7-layer convolutional layer and a 5-layer pooling layer. Small convolution Neural network, and adding batch regularization twice in the middle layer to avoid model gradient dispersion and accelerate convergence;
S5:数据后处理,将对应真实值的字典数组的下标,做一个映射变换变成实际值。S5: Data post-processing, transform the subscript of the dictionary array corresponding to the real value into the actual value by doing a mapping transformation.
本发明所提供的一种罐表面字符识别方法,利用YOLO-V3与CRNN深度学习方法实现了文本字符实时识别,能快速框选出字符所在的区域,并能够识别。因为运用了深度学习,所以该方法具有鲁棒性高,准确率高,速度快的优点,为奶粉罐生产企业提供智能字符定位的检测方案,解决人工识别效率低、速度慢、人员成本高等问题。The method for recognizing characters on the surface of a tank provided by the present invention realizes real-time recognition of text characters by using the YOLO-V3 and CRNN deep learning method, and can quickly frame and select the area where the characters are located, and can recognize them. Because of the use of deep learning, this method has the advantages of high robustness, high accuracy and fast speed. It provides a detection solution for intelligent character positioning for milk powder can manufacturers, and solves the problems of low manual recognition efficiency, slow speed and high personnel cost. .
附图说明Description of drawings
图1为本发明实施例提供的奶粉罐表面字符识别流程示意图;Fig. 1 is the schematic flow chart of the character recognition process on the surface of the milk powder can provided by the embodiment of the present invention;
图2为本发明实施例提供的YOLO-V3训练过程图;Fig. 2 is a YOLO-V3 training process diagram provided by an embodiment of the present invention;
图3为本发明实施例提供的改进CRNN部分网络结构图。FIG. 3 is a partial network structure diagram of the improved CRNN provided by an embodiment of the present invention.
具体实施方式Detailed ways
为了能够更加详尽地了解本发明的特点与技术内容,下面结合附图对本发明的实现进行详细阐述,所附附图仅供参考说明之用,并非用来限定本发明。In order to be able to understand the features and technical content of the present invention in more detail, the implementation of the present invention is described in detail below with reference to the accompanying drawings, which are for reference only and are not intended to limit the present invention.
本发明提供一种罐表面字符识别方法,如图1所示,图1为本发明实施例提供的奶粉罐表面字符识别流程示意图;罐表面字符识别包括:实时采集罐表面图像;数据采集阶段应当利用摄像机等尽可能多角度的拍摄奶粉罐表面图像,以涵盖各种角度字符以及不同字符使得训练得到的模型检出率和准确率都比较高。The present invention provides a method for recognizing characters on the surface of a can, as shown in FIG. 1 , which is a schematic diagram of a process for recognizing characters on the surface of a milk powder can provided by an embodiment of the present invention; the character recognition on the surface of the can includes: collecting images of the surface of the can in real time; The surface images of the milk powder cans are taken from as many angles as possible, such as cameras, to cover characters from various angles and different characters, so that the detection rate and accuracy of the trained model are relatively high.
利用预先建立的YOLO-V3神经网络检测模型对实时采集的罐表面图像进行字符定位,得到包括是否含有字符和字符位置信息;输出字符预测框位置,并裁剪出字符区域;裁剪出的字符区域存在与实际文本区域贴合不够紧密的问题,使用文本矫正矫正算法进行矫正;通过直方图投影,分割出每一行的文本字符;对裁剪出来的文本区域使用改进的端到端不定长文本CRNN模型进行识别;根据当时系统情况,如果是在线模式,识别结果上传到数据库;如果是离线模式,识别结果将保存在本地。Use the pre-established YOLO-V3 neural network detection model to locate the characters on the surface image of the tank collected in real time, and obtain the information including whether it contains characters and the position of the characters; output the position of the character prediction frame, and crop out the character area; the cropped character area exists The problem of not fitting closely with the actual text area is corrected by using a text correction correction algorithm; the text characters of each line are segmented through histogram projection; the cropped text area is processed using an improved end-to-end variable-length text CRNN model. Recognition; according to the system conditions at the time, if it is in online mode, the recognition result will be uploaded to the database; if it is in offline mode, the recognition result will be saved locally.
其中,罐表面字符识别方法还包括:执行机构切换下一个待识别工件。Wherein, the method for recognizing characters on the surface of the tank further includes: the actuator switches over the next workpiece to be recognized.
在本发明提供的实施例中,罐表面可以包括:奶粉罐表面,罐头罐表面,一起其他罐装包装的表面。In the embodiment provided by the present invention, the surface of the can includes: the surface of the milk powder can, the surface of the can, and the surface of other cans.
在本发明提供的实施例中,YOLO-V3神经网络检测模型的建立方法包括:In the embodiment provided by the present invention, the establishment method of the YOLO-V3 neural network detection model includes:
S1:采集大量奶粉罐表面字符图像;S1: Collect a large number of character images on the surface of milk powder cans;
S2:对奶粉罐表面图像进行预处理;预处理包括对奶粉罐表面图像进行归一化处理:将等比例缩放后的奶粉罐表面图像填入预设像素大小的空白图像中,然后将空白图像上除奶粉罐图像外的像素全部填充为设定颜色;空白图像的像素大小为416*416,空白图像能够覆盖缩放后的奶粉罐图像;归一化过程按图像原有长宽比缩放到相应大小,填入到416*416空白图像中间,其他像素全部填为白色,以尽可能保持图像原有纹理特征,并通过翻转、平移策略对归一化后的瑕疵图像进行数据增强,设定颜色包括灰色;S2: Preprocessing the surface image of the milk powder can; the preprocessing includes normalizing the surface image of the milk powder can: filling the proportionally scaled surface image of the milk powder can into a blank image with a preset pixel size, and then adding the blank image All pixels except the image of the milk powder can are filled with the set color; the pixel size of the blank image is 416*416, and the blank image can cover the zoomed image of the milk powder can; the normalization process is scaled to the corresponding image according to the original aspect ratio of the image. Size, fill in the middle of the 416*416 blank image, and fill all other pixels with white, so as to keep the original texture characteristics of the image as much as possible, and use the flipping and translation strategies to enhance the normalized defect image and set the color. including grey;
S3:对预处理后的图像进行字符区域标定,得到各图像的字符位置标签;通过人工框选方式对预处理后的奶粉罐图像进行字符区域标定,将各标定区域的左上角点和右下角点坐标和相应区域的字符存储为字符标签;在图像标定阶段,字符图像标签的制作包括:可使用标记软件,用矩形框人工标定字符区域;将矩形框的左上角点和右下角点的横纵坐标存于txt文件中,所述标记软件可采用现有图形绘制修改软件,labelImg等软件,只为方便在人工框选后,能够自动或人为获取相应的坐标数据。S3: Perform character area calibration on the preprocessed image to obtain character position labels of each image; perform character area calibration on the preprocessed milk powder can image by manual frame selection, and set the upper left corner and lower right corner of each calibration area The point coordinates and the characters in the corresponding area are stored as character labels; in the image calibration stage, the production of character image labels includes: labeling software can be used to manually demarcate the character area with a rectangular frame; The ordinate is stored in the txt file, and the labeling software can adopt existing graphics drawing modification software, labelImg and other software, only for the convenience of automatically or artificially obtaining the corresponding coordinate data after manual frame selection.
S4:将图像及其位置信息标签划分为训练集样本和测试集样本,训练集样本和测试机样本随机划分,划分比例为9:1,可上下浮动;利用训练集样本和测试机样本,采用计算机图形处理器GPU对预先搭建的YOLO-V3深度神经网络进行训练,得到YOLO-V3神经网络检测模型。S4: Divide the image and its position information labels into training set samples and test set samples, and randomly divide the training set samples and test machine samples, and the division ratio is 9:1, which can float up and down; using training set samples and test machine samples, adopt The computer graphics processor GPU trains the pre-built YOLO-V3 deep neural network to obtain the YOLO-V3 neural network detection model.
在本发明提供的实施例中,预处理还包括对归一化处理后得到的罐表面字符图像进行数据增强;罐表面字符图像进行数据增强包括:In the embodiment provided by the present invention, the preprocessing further includes performing data enhancement on the character image on the tank surface obtained after the normalization process; performing data enhancement on the character image on the tank surface includes:
几何变换类:包括翻转,旋转,裁剪,变形,缩放;Geometric transformation class: including flipping, rotating, cropping, deforming, scaling;
颜色变换类:包括噪声、模糊、颜色变换、擦除、填充;Color transformation class: including noise, blur, color transformation, erase, fill;
根据图片样式模拟生成样本图片。Generate a sample image based on the image style simulation.
在本发明提供的实施例中,对预先搭建的YOLO-V3神经网络进行训练包括:In the embodiment provided by the present invention, training the pre-built YOLO-V3 neural network includes:
S1:设置训练参数:将迭代步数epochs设置为10000,将学习率优化器optimizer设置为'adam',将批训练样本数batch_size设置为64,以及设置9个先验框anchor box的大小;S1: Set the training parameters: set the number of iteration steps epochs to 10000, set the learning rate optimizer optimizer to 'adam', set the number of batch training samples batch_size to 64, and set the size of the 9 a priori box anchor boxes;
S2:训练YOLO-V3神经网络:把训练集样本数据和各样本对应的奶粉罐表面字符数据输入到YOLO-V3神经网络中进行模型训练;S2: Train the YOLO-V3 neural network: input the training set sample data and the corresponding character data on the surface of the milk powder can into the YOLO-V3 neural network for model training;
S3:模型测试:每次训练结束后,利用测试集样本对训练得到的模型进行测试,若模型对测试集样本中的瑕疵检出率超过95%且检测准确率不低于95%,则将最后一次训练得到的模型作为最终YOLO-V3神经网络检测模型;否则将最后一次训练得到的模型作为当前待训练的YOLO-V3神经网络,重复步骤S2-S3,直至得到最终YOLO-V3神经网络检测模型。S3: Model test: After each training, use the test set samples to test the trained model. If the model has a defect detection rate of more than 95% in the test set samples and the detection accuracy is not less than 95%, the model will be tested. The model obtained from the last training is used as the final YOLO-V3 neural network detection model; otherwise, the model obtained from the last training is used as the current YOLO-V3 neural network to be trained, and steps S2-S3 are repeated until the final YOLO-V3 neural network detection model is obtained. Model.
后续即可利用最终YOLO-V3神经网络检测模型进行字符的实时检测:利用摄像头实时采集待检测的奶粉罐表面图像,根据训练好的YOLO-V3网络,实时检测奶粉罐表面字符位置。Afterwards, the final YOLO-V3 neural network detection model can be used to perform real-time detection of characters: the camera is used to collect the surface image of the milk powder can to be detected in real time, and according to the trained YOLO-V3 network, the position of the characters on the surface of the milk powder can is detected in real time.
在本发明提供的实施例中,先验框的大小通过k-means聚类方法设置,包括:随机选取9个字符区域矩形框对象作为训练集的聚类中心,根据各对象与各个聚类中心之间的距离,把各对象分配给距离最近的聚类中心,每分配一个样本,聚类中心重新计算一次,重复分配过程,直到没有聚类中心再发生变化,当前9个聚类中心的值即9个先验框的大小。In the embodiment provided by the present invention, the size of the prior frame is set by the k-means clustering method, including: randomly selecting 9 character area rectangular frame objects as the clustering centers of the training set, and according to each object and each clustering center The distance between each object is assigned to the nearest cluster center, each time a sample is assigned, the cluster center is recalculated, and the assignment process is repeated until no cluster center changes, and the current value of the 9 cluster centers That is, the size of the 9 a priori boxes.
为了更好地理解基于YOLO-V3网络的字符位置检测方法,这里对YOLO-V3网络的工作原理进行简要说明:In order to better understand the character position detection method based on the YOLO-V3 network, here is a brief description of the working principle of the YOLO-V3 network:
a、YOLO-V3网络将输入的图像平均划分成S×S个单元格;a. The YOLO-V3 network evenly divides the input image into S×S cells;
b、每个单元格会预测B个边界框(Bounding Box),以向量的形式给出这些边界框的信息。边界框的信息包括了位置信息(矩形框中心点坐标,宽和高),置信度(Confidence)以及预测物体的类别信息。b. Each cell will predict B bounding boxes, and give the information of these bounding boxes in the form of vectors. The information of the bounding box includes position information (coordinates of the center point of the rectangular box, width and height), confidence and category information of the predicted object.
c、对于训练数据,图像和标签输入后,将单元格输出的五个参数(矩形框中心点坐标,宽和高,置信度)代入损失函数(损失函数用来计算YOLO-V3网络计算出的五个参数与标注的五个参数之间的差距)计算,通过反向传播来调整权重,使正确字符区域的置信度升高,非正确字符区域的置信度降低;对于实时采集的数据在输入YOLO-V3网络后,也会算出五个参数(矩形框中心点坐标,宽和高,置信度),将这五个数据通过损失函数计算后,会得到使损失函数值最低的边界框,也就是最后需要的分类框,即检测结果。c. For the training data, after the image and label are input, the five parameters output by the cell (coordinates of the center point of the rectangular box, width and height, confidence) are substituted into the loss function (the loss function is used to calculate the value calculated by the YOLO-V3 network. The difference between the five parameters and the five marked parameters) is calculated, and the weight is adjusted by backpropagation, so that the confidence of the correct character area is increased, and the confidence of the incorrect character area is decreased; for the real-time collected data in the input After the YOLO-V3 network, five parameters (coordinates of the center point of the rectangular box, width and height, confidence) will also be calculated. After calculating these five data through the loss function, the bounding box with the lowest loss function value will be obtained. It is the final classification frame required, that is, the detection result.
在本发明提供的实施例中,文本矫正算法包括:In the embodiment provided by the present invention, the text correction algorithm includes:
S1:裁剪后的图像进行灰度化处理,字符为白色,背景为色;S1: The cropped image is grayscaled, the characters are white, and the background is colored;
S2:进行高斯模糊,使文本部分能连成一块,再对图像进行阈值处理得到二值图像;S2: Perform Gaussian blurring so that the text parts can be connected together, and then perform threshold processing on the image to obtain a binary image;
S3:使用最小矩形拟合,然后通过矩形面积大小,筛选得到文本区域部分的面积,此部分就是文本最小外接矩形;S3: Use the smallest rectangle to fit, and then filter the area of the text area by the size of the rectangle area, which is the smallest bounding rectangle of the text;
S4:得到最小外接矩形的中心点,对文本区域进行仿射变换来实现校正倾斜文本。S4: Obtain the center point of the minimum circumscribed rectangle, and perform affine transformation on the text area to correct the oblique text.
在本发明提供的实施例中,直方图水平投影算法包括:In the embodiment provided by the present invention, the histogram horizontal projection algorithm includes:
S1:对步骤四得到矫正后的图像进行二值化;S1: Binarize the corrected image obtained in step 4;
S2:对二值化后的图像计算水平方向上像素和,绘制出横轴为图像纵轴坐标范围为0~图像宽度,纵轴为像素和的直方图;S2: Calculate the pixel sum in the horizontal direction for the binarized image, and draw a histogram with the horizontal axis as the image and the vertical axis as the coordinate range from 0 to the image width, and the vertical axis as the pixel sum;
S3:找出直方图谷底部分,此部分为分割线处;S3: Find the bottom part of the histogram, which is the dividing line;
S4:根据分割线,分割出每一行文本图像。S4: According to the dividing line, each line of text image is divided.
在本发明提供的实施例中,改进的端到端不定长文本CRNN模型进行识别包括:In the embodiment provided by the present invention, the identification of the improved end-to-end variable-length text CRNN model includes:
S1:每一行文本图像灰度化处理,将三通道的图像转为单通道的灰度图;S1: Grayscale processing of each line of text image, converting the three-channel image into a single-channel grayscale image;
S2:长度比例缩放和填充,将输入图像的尺寸固定为32*width,按比例进行缩放或扩充,这样能够与训练的样本相类似;S2: Length scaling and padding, fixing the size of the input image to 32*width, scaling or expanding proportionally, so that it can be similar to the training samples;
S3:数据标签稀疏矩阵转化,对标签矩阵进行处理,将其转化为需要的数据格式;S3: Data label sparse matrix conversion, process the label matrix and convert it into the required data format;
S4:搭建模型,更改结构,将原来的CNN+RNN+CTC转录层的思想,更改为CNN+CTC,主干网络在VGG上做修改设计7层卷积层,5层池化层的小型卷积神经网络,并在中间层的加入两次批量正则化,避免模型梯度弥散,加速收敛;S4: Build the model, change the structure, change the original idea of CNN+RNN+CTC transcription layer to CNN+CTC, and modify the backbone network on VGG to design a 7-layer convolutional layer and a 5-layer pooling layer. Small convolution Neural network, and adding batch regularization twice in the middle layer to avoid model gradient dispersion and accelerate convergence;
S5:数据后处理,将对应真实值的字典数组的下标,做一个映射变换变成实际值。S5: Data post-processing, transform the subscript of the dictionary array corresponding to the real value into the actual value by doing a mapping transformation.
本发明提供的实施例中,预先搭建的改进CRNN神经网络的网络结构如下:In the embodiment provided by the present invention, the network structure of the pre-built improved CRNN neural network is as follows:
S1、其结构层次包括卷积神经网络CNN,删除原来的RNN结构,CNN包括7层卷积层,除第三层和第七层卷积层以外,每层卷积层之后都会进行最大池化处理;S1. Its structure level includes convolutional neural network CNN, delete the original RNN structure, CNN includes 7 convolutional layers, except for the third and seventh convolutional layers, each convolutional layer will be followed by maximum pooling deal with;
S2、调整CRNN网络中参数,将batch大小设置为16或32,learning rate设置为0.00001,epoch设置为100;S2. Adjust the parameters in the CRNN network, set the batch size to 16 or 32, the learning rate to 0.00001, and the epoch to 100;
S3、将上一步得到的文本图像进行训练集和测试集的分类,根据图片里面文本特征,编写代码生成相识的图片与各种情况的文本图片,这样有助于增加模型的泛化能力,将文本数据集集进行打乱,随机分成两部分:其中测试集10000张,训练集100000张;S3. Classify the text images obtained in the previous step into the training set and the test set. According to the text features in the pictures, write codes to generate familiar pictures and text pictures of various situations, which will help to increase the generalization ability of the model. The text dataset is scrambled and randomly divided into two parts: 10,000 for the test set and 100,000 for the training set;
S4、将训练集放入CRNN网络中进行训练,并在训练的同时对当前模型进行测试;S4. Put the training set into the CRNN network for training, and test the current model while training;
S5、当训练以及测试的损失函数收敛时,停止训练,获取CRNN模型S5. When the training and testing loss functions converge, stop training and obtain the CRNN model
S6、使用CRNN识别模型对裁剪后的文本进行识别;’S6. Use the CRNN recognition model to recognize the cropped text;'
S7、根据当时系统情况,如果是在线模式,识别结果上传到数据库;如果是离线模式,识别结果将保存在本地;S7. According to the system situation at the time, if it is in online mode, the recognition result will be uploaded to the database; if it is in offline mode, the recognition result will be saved locally;
S8、执行机构将切换下一个待识别工件;S8. The actuator will switch to the next workpiece to be identified;
本发明所提供的一种罐表面字符识别方法,利用YOLO-V3与CRNN深度学习方法实现了文本字符实时识别,能快速框选出字符所在的区域,并能够识别。因为运用了深度学习,所以该方法具有鲁棒性高,准确率高,速度快的优点,为奶粉罐生产企业提供智能字符定位的检测方案,解决人工识别效率低、速度慢、人员成本高等问题。The method for recognizing characters on the surface of a tank provided by the present invention realizes real-time recognition of text characters by using the YOLO-V3 and CRNN deep learning method, and can quickly select the area where the characters are located, and can recognize them. Because of the use of deep learning, this method has the advantages of high robustness, high accuracy and fast speed. It provides a detection solution for intelligent character positioning for milk powder can manufacturers, and solves the problems of low manual recognition efficiency, slow speed and high personnel cost. .
如上所述,仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本发明的保护范围之内。因此,本发明的保护范围应以所述权利要求的保护范围为准。As mentioned above, it is only a specific embodiment of the present invention, but the protection scope of the present invention is not limited to this. should be included within the protection scope of the present invention. Therefore, the protection scope of the present invention should be based on the protection scope of the claims.
Claims (10)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010998223.7A CN112085024A (en) | 2020-09-21 | 2020-09-21 | A method for character recognition on the surface of a tank |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010998223.7A CN112085024A (en) | 2020-09-21 | 2020-09-21 | A method for character recognition on the surface of a tank |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112085024A true CN112085024A (en) | 2020-12-15 |
Family
ID=73738292
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010998223.7A Pending CN112085024A (en) | 2020-09-21 | 2020-09-21 | A method for character recognition on the surface of a tank |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112085024A (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112613527A (en) * | 2020-12-17 | 2021-04-06 | 西北大学 | Minimum quantization feature detection method based on unsupervised learning |
CN112818823A (en) * | 2021-01-28 | 2021-05-18 | 建信览智科技(北京)有限公司 | Text extraction method based on bill content and position information |
CN112883965A (en) * | 2021-02-08 | 2021-06-01 | 中兴盛达电气技术(郑州)有限公司 | Date detection method on packaging vessel, electronic device and computer-readable storage medium |
CN113436180A (en) * | 2021-07-07 | 2021-09-24 | 京东科技控股股份有限公司 | Method, device, system, equipment and medium for detecting spray codes on production line |
CN113627190A (en) * | 2021-08-27 | 2021-11-09 | 上海复深蓝软件股份有限公司 | Visualized data conversion method and device, computer equipment and storage medium |
CN113971779A (en) * | 2021-10-29 | 2022-01-25 | 中国水利水电科学研究院 | An automatic reading method of water gauge based on deep learning |
CN114005127A (en) * | 2021-11-15 | 2022-02-01 | 中再云图技术有限公司 | Image optical character recognition method based on deep learning, storage device and server |
CN114299502A (en) * | 2022-03-07 | 2022-04-08 | 科大智能物联技术股份有限公司 | Method for correcting and identifying inclination of code-spraying characters on end face of round casting blank and storage medium |
CN114330247A (en) * | 2021-11-09 | 2022-04-12 | 世纪保众(北京)网络科技有限公司 | Automatic insurance clause analysis method based on image recognition |
CN115578695A (en) * | 2022-11-21 | 2023-01-06 | 昆明理工大学 | A water gauge water level machine vision detection method and device with free shooting angle |
KR20230128813A (en) * | 2022-02-28 | 2023-09-05 | 에스케이텔레콤 주식회사 | Method And Apparatus for Augmentation of Learning Data for Text Recognition Model |
CN116912845A (en) * | 2023-06-16 | 2023-10-20 | 广东电网有限责任公司佛山供电局 | Intelligent content identification and analysis method and device based on NLP and AI |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110097048A (en) * | 2019-04-02 | 2019-08-06 | 江苏理工学院 | A kind of SOT chip image quickly corrects and character identifying method |
CN110490874A (en) * | 2019-09-04 | 2019-11-22 | 河海大学常州校区 | Weaving cloth surface flaw detecting method based on YOLO neural network |
CN110956171A (en) * | 2019-11-06 | 2020-04-03 | 广州供电局有限公司 | Automatic nameplate identification method and device, computer equipment and storage medium |
CN111444908A (en) * | 2020-03-25 | 2020-07-24 | 腾讯科技(深圳)有限公司 | Image recognition method, device, terminal and storage medium |
-
2020
- 2020-09-21 CN CN202010998223.7A patent/CN112085024A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110097048A (en) * | 2019-04-02 | 2019-08-06 | 江苏理工学院 | A kind of SOT chip image quickly corrects and character identifying method |
CN110490874A (en) * | 2019-09-04 | 2019-11-22 | 河海大学常州校区 | Weaving cloth surface flaw detecting method based on YOLO neural network |
CN110956171A (en) * | 2019-11-06 | 2020-04-03 | 广州供电局有限公司 | Automatic nameplate identification method and device, computer equipment and storage medium |
CN111444908A (en) * | 2020-03-25 | 2020-07-24 | 腾讯科技(深圳)有限公司 | Image recognition method, device, terminal and storage medium |
Non-Patent Citations (4)
Title |
---|
CV.EXP: "主流ocr算法:CNN+BLSTM+CTC架构", pages 1 - 3, Retrieved from the Internet <URL:https://blog.csdn.net/forest_world/article/details/78566737> * |
刘晓杰 等: "图像处理技术在船舶水尺刻度线检测中的应用", 《舰船科学技术》, vol. 39, no. 16, pages 10 - 12 * |
钟巧: "基于图论的扫描图像文本行分割与矫正", 《中国优秀硕士学位论文全文数据库 信息科技辑》, no. 7, pages 138 - 1469 * |
陈玉新: "基于深度学习的场景文字识别研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》, no. 5, pages 138 - 166 * |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112613527B (en) * | 2020-12-17 | 2023-07-28 | 西北大学 | Minimum quantitative feature detection method based on unsupervised learning |
CN112613527A (en) * | 2020-12-17 | 2021-04-06 | 西北大学 | Minimum quantization feature detection method based on unsupervised learning |
CN112818823A (en) * | 2021-01-28 | 2021-05-18 | 建信览智科技(北京)有限公司 | Text extraction method based on bill content and position information |
CN112818823B (en) * | 2021-01-28 | 2024-04-12 | 金科览智科技(北京)有限公司 | Text extraction method based on bill content and position information |
CN112883965A (en) * | 2021-02-08 | 2021-06-01 | 中兴盛达电气技术(郑州)有限公司 | Date detection method on packaging vessel, electronic device and computer-readable storage medium |
CN113436180A (en) * | 2021-07-07 | 2021-09-24 | 京东科技控股股份有限公司 | Method, device, system, equipment and medium for detecting spray codes on production line |
CN113627190A (en) * | 2021-08-27 | 2021-11-09 | 上海复深蓝软件股份有限公司 | Visualized data conversion method and device, computer equipment and storage medium |
CN113971779B (en) * | 2021-10-29 | 2022-07-01 | 中国水利水电科学研究院 | Water gauge automatic reading method based on deep learning |
CN113971779A (en) * | 2021-10-29 | 2022-01-25 | 中国水利水电科学研究院 | An automatic reading method of water gauge based on deep learning |
CN114330247A (en) * | 2021-11-09 | 2022-04-12 | 世纪保众(北京)网络科技有限公司 | Automatic insurance clause analysis method based on image recognition |
CN114005127A (en) * | 2021-11-15 | 2022-02-01 | 中再云图技术有限公司 | Image optical character recognition method based on deep learning, storage device and server |
KR20230128813A (en) * | 2022-02-28 | 2023-09-05 | 에스케이텔레콤 주식회사 | Method And Apparatus for Augmentation of Learning Data for Text Recognition Model |
KR102740679B1 (en) * | 2022-02-28 | 2024-12-10 | 에스케이텔레콤 주식회사 | Method And Apparatus for Augmentation of Learning Data for Text Recognition Model |
CN114299502A (en) * | 2022-03-07 | 2022-04-08 | 科大智能物联技术股份有限公司 | Method for correcting and identifying inclination of code-spraying characters on end face of round casting blank and storage medium |
CN114299502B (en) * | 2022-03-07 | 2022-06-17 | 科大智能物联技术股份有限公司 | Method for correcting and identifying inclination of code-spraying characters on end face of round casting blank and storage medium |
CN115578695A (en) * | 2022-11-21 | 2023-01-06 | 昆明理工大学 | A water gauge water level machine vision detection method and device with free shooting angle |
CN116912845A (en) * | 2023-06-16 | 2023-10-20 | 广东电网有限责任公司佛山供电局 | Intelligent content identification and analysis method and device based on NLP and AI |
CN116912845B (en) * | 2023-06-16 | 2024-03-19 | 广东电网有限责任公司佛山供电局 | Intelligent content identification and analysis method and device based on NLP and AI |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112085024A (en) | A method for character recognition on the surface of a tank | |
CN111027547B (en) | Automatic detection method for multi-scale polymorphic target in two-dimensional image | |
CN109961049B (en) | Cigarette brand identification method under complex scene | |
CN108918536B (en) | Tire mold surface character defect detection method, device, equipment and storage medium | |
CN112036231B (en) | A detection and recognition method of lane lines and road signs based on vehicle video | |
CN110287960A (en) | Detection and Recognition Method of Curved Characters in Natural Scene Images | |
CN100565559C (en) | Image text location method and device based on connected component and support vector machine | |
CN113160192A (en) | Visual sense-based snow pressing vehicle appearance defect detection method and device under complex background | |
CN111310760B (en) | Oracle Bone Inscription Text Detection Method Combining Local Prior Features and Deep Convolution Features | |
CN117854072B (en) | Automatic labeling method for industrial visual defects | |
CN111539330B (en) | Transformer substation digital display instrument identification method based on double-SVM multi-classifier | |
CN110807775A (en) | Traditional Chinese medicine tongue image segmentation device and method based on artificial intelligence and storage medium | |
CN110363196B (en) | Method for accurately recognizing characters of inclined text | |
CN115331245B (en) | Table structure identification method based on image instance segmentation | |
CN113221956B (en) | Target identification method and device based on improved multi-scale depth model | |
CN112926694A (en) | Method for automatically identifying pigs in image based on improved neural network | |
CN111027538A (en) | Container detection method based on instance segmentation model | |
CN110929795A (en) | Method for quickly identifying and positioning welding spot of high-speed wire welding machine | |
CN113065396A (en) | Automated filing processing system and method for scanned archive images based on deep learning | |
CN113158977A (en) | Image character editing method for improving FANnet generation network | |
CN116342525A (en) | SOP chip pin defect detection method and system based on Lenet-5 model | |
CN116993985A (en) | Method for realizing Zero-Shot automatic cutting of safety belt based on CLIP | |
CN117745708A (en) | Deep learning algorithm-based wood floor surface flaw detection method | |
CN117576699A (en) | Locomotive work order information intelligent recognition method and system based on deep learning | |
CN111754525A (en) | Industrial character detection process based on non-precise segmentation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20201215 |
|
RJ01 | Rejection of invention patent application after publication |