Nothing Special   »   [go: up one dir, main page]

CN112329771B - Deep learning-based building material sample identification method - Google Patents

Deep learning-based building material sample identification method Download PDF

Info

Publication number
CN112329771B
CN112329771B CN202011201983.7A CN202011201983A CN112329771B CN 112329771 B CN112329771 B CN 112329771B CN 202011201983 A CN202011201983 A CN 202011201983A CN 112329771 B CN112329771 B CN 112329771B
Authority
CN
China
Prior art keywords
building material
sample
roi
image
feature map
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011201983.7A
Other languages
Chinese (zh)
Other versions
CN112329771A (en
Inventor
赵力
程荣
张亦明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Suzhou Institute Of Building Science Group Co ltd
Yuanzhun Intelligent Technology Suzhou Co ltd
Original Assignee
Suzhou Institute Of Building Science Group Co ltd
Yuanzhun Intelligent Technology Suzhou Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Suzhou Institute Of Building Science Group Co ltd, Yuanzhun Intelligent Technology Suzhou Co ltd filed Critical Suzhou Institute Of Building Science Group Co ltd
Priority to CN202011201983.7A priority Critical patent/CN112329771B/en
Publication of CN112329771A publication Critical patent/CN112329771A/en
Application granted granted Critical
Publication of CN112329771B publication Critical patent/CN112329771B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a building material sample recognition method based on deep learning, which comprises a model training stage and a sample recognition stage, wherein the model training stage comprises the steps of manufacturing a building material sample data set and constructing a multi-scale information fusion convolutional neural network for sample recognition; enhancing the data of the sample data set to obtain the best model performance; the sample identification stage comprises the steps of inputting processed building material sample images into a model, carrying out feature extraction to generate an optimal size feature map, correcting the generated feature map to be used as an ROI, transmitting the ROI into ROI pooling layers according to different scales, mapping the ROI into the same size proposal, projecting the image to generate proposal feature map on an original building material sample image, carrying out BBOX and CLS branch processing and other steps to generate a building material detection frame with accurate positions, and identifying the material performance state of the sample. The method carries out information extraction through the multi-scale feature map, well learns target feature information with different scales, has good identification performance and universality, and has wide application prospect in the field of construction engineering.

Description

Deep learning-based building material sample identification method
Technical Field
The invention belongs to the field of construction engineering, and particularly relates to a building material sample identification method based on deep learning.
Background
With the increase of data information processing demands and the rapid development of artificial intelligence technology, people begin to try to identify building material samples by using a method based on machine learning or deep learning, such as a clustering neural network, a support vector machine, a wavelet transformation neural network and other shallow network algorithms. However, the shallow network algorithm needs various complex algorithms to extract and determine sample identification characteristic information from echo information; the computational complexity and the consumption of computational resources are high and therefore the versatility is low. A convolutional neural network is one of important models in the field of deep learning, and its network structure has high invariance to image data having characteristics of translation, inversion, affine transformation, and the like, and therefore, in recent years, the convolutional neural network has been widely used in various fields in computer vision and has achieved excellent results. However, in the identification process of the traditional single linear convolutional neural network, only the last layer of output is aimed at, namely: the single-scale feature map is used for information extraction, so that target feature information with different scales cannot be well learned obviously; it is therefore difficult to achieve good recognition performance in complex building material scenarios.
Disclosure of Invention
In order to solve the problems in the prior art, the invention provides a building material sample identification method based on deep learning.
The aim of the invention is achieved by the following technical scheme:
a building material sample recognition method based on deep learning, which comprises a model training stage and a sample recognition stage,
The model training phase comprises the following steps:
S1, collecting and marking building material samples, and manufacturing a building material sample data set; the data set comprises building material samples of each category and is divided into a training set, a testing set and an evaluation set;
s2, constructing a multi-scale information fusion convolutional neural network for building material sample identification;
S3, carrying out data enhancement on the sample data set in the S1: to obtain optimal model performance;
The sample identification phase comprises the following steps:
s1, inputting the processed building material sample image into a model, pre-training through imagenet, removing Resnet on the top layer, extracting features, and generating a feature map with the optimal size;
S2, the feature images generated in the S1 are respectively processed by RPN2,3,4 and 5to generate candidate anchors with different sizes; the anchor area is set, and all the anchors generated by the RPN uniformly adopt 1: 1. 1: 2. 2:1, generating a plurality of candidate anchors, and then screening out the most complete anchor containing the target by utilizing a two-class function and bounding box regerssion function according to the RPN and real labeling, and correcting the most complete anchor to be used as the ROI;
S3, using the feature image layers output by different residual convolution modules as the input of ROI pooling layers, outputting the feature image by a deep convolution module by using the large-scale ROI, and using the ROI leave as a criterion output by a corresponding layer convolution module:
Wherein w and h are the length and width of the ROI area, and K 0 is the reference leave;
S4, transmitting the ROI generated in the S3 into ROI pooling layers, uniformly mapping the multi-scale ROI into proposal with the same size by ROI pooling, and projecting the multi-scale ROI on an original building material sample image to generate proposal feature map, so that subsequent BBOX and CLS branches are convenient to process;
S5, calculating the category to which each sample belongs through the full connection layer and softmax, and outputting the highest category probability as a confidence coefficient by the CLS branch pair proposal feature map;
And S6, BBOX, correcting the proposal area by utilizing the bounding box regression function, generating a building material detection frame with more accurate positions, and identifying the material performance state of the building material sample.
Preferably, the model training phase S3 comprises the steps of:
s31, building material sample images under different scales and scenes are constructed by utilizing a combination of a plurality of data enhancement methods, existing data are expanded to simulate a complex identification scene, the learning of the model on detail characteristic information is improved, and the universality of the model is enhanced;
s32, setting initial weight as pre-training weight on imagenet, setting initial learning rate as 0.001, learning rate attenuation index as 0.1, batch_size as 16, and inputting image size;
s33, on a loss function, the RPN series module adopts two kinds of loss and regression loss; the CLS branch adopts multi-classification loss, and the BBOX branch adopts regression loss;
s34, training on a training set and a testing set by adopting an SGD random gradient descent optimizer until the model performance reaches the best.
Preferably, the method for enhancing data in S31 includes:
S311, random Erasing algorithm:
(3) IRE (Image-aware Random Erasing), randomly selecting an occlusion position on the whole target Image;
(4) ORE (Object-aware Random Erasing), randomly selecting an occlusion position within the Object's bounding-box region;
(3) Combining both IRE and ORE;
s312, HIDE AND SEEK algorithm:
Dividing the picture into S-S grids, hiding each grid according to probability, hiding different grid groups for each batch of the same picture in training;
S313, grid Mask algorithm:
in order to avoid the problem that the complete target is deleted or the context information is deleted because the over-deleted area exists in the step S311 and the step S312; setting four parameters of x, y, r and d through a Grid Mask:
Wherein r is the size of mask, M is the reserved pixel number, H, W is the image size; x and y are the area coordinates randomly generated on the image; the value of the non-shielding area is 1, the value of the shielding area is 0, a mask with the same resolution as the original image is generated, and then the mask is multiplied with the original image to obtain an image;
s314, mixup algorithm
The method comprises the steps of carrying out mixed enhancement on images, and mixing the images among different classes; the algorithm can be summarized as follows:
Where x 1、x2 is the pixel of the different image and λ is the weight; Is the output pixel after the mixed class;
s315, cutmix algorithm
A portion of the region is randomly cropped and region pixel values of other data in the training set are randomly populated.
Preferably, the step of generating the best feature map after the sample is extracted in the sample identifying stage S1 includes:
S11, marking the last plurality of residual convolution modules in Resnet as { C 1,C2,C3,…Cn }, and respectively extracting output characteristic graphs of the residual modules as { P 1,P2,P3,…,Pn };
S12, performing 2 times nearest neighbor up-sampling on the deepest feature map P 5;
S13, extracting an output characteristic diagram P n-1 of a residual convolution module C n-1 adjacent to the C n, and performing 1*1 convolution dimension reduction processing;
S14, fusing the feature map P n by a pixel value adding method of the corresponding part;
S15, reducing an aliasing effect caused by up-sampling through 3*3 convolution of the fusion feature map;
S16, iterating the processes of S11 to S15 until the optimal size characteristic diagram is generated.
The beneficial effects of the invention are as follows: the method of the invention extracts information through the multi-scale feature map, well learns the target feature information with different scales, has good identification performance and universality, and has wide application prospect in the field of construction engineering.
Detailed Description
The technical proposal of the invention is specifically described below by combining the embodiment, the invention discloses a building material sample recognition method based on deep learning, which comprises a model training stage and a sample recognition stage,
The model training phase comprises the following steps:
S1, collecting and marking building material samples, and manufacturing a building material sample data set; the data set comprises building material samples of each category and is divided into a training set, a testing set and an evaluation set;
s2, constructing a multi-scale information fusion convolutional neural network for building material sample identification;
S3, carrying out data enhancement on the sample data set in the S1 to obtain the optimal model performance;
in particular, the method comprises the steps of,
S31, building material sample images under different scales and scenes are constructed by utilizing a combination of a plurality of data enhancement methods, existing data are expanded to simulate a complex identification scene, the learning of the model on detail characteristic information is improved, and the universality of the model is enhanced;
S32, setting initial weight as pre-training weight on imagenet, setting initial learning rate to be 0.001, learning rate attenuation index to be 0.1, batch_size to be 16, and input image size to be 224 x 224.
S33, on a loss function, the RPN series module adopts two kinds of loss and regression loss; the CLS branch uses multiple classification losses and the BBOX branch uses regression losses.
And S34, training 20 epochs on the training set and the testing set by adopting an SGD random gradient descent optimizer in the training until the performance of the model is optimal.
Wherein, the enhancing method in S31 includes the following steps:
S311, random Erasing algorithm:
(5) IRE, randomly selecting a shielding position on the whole target image;
(6) ORE, randomly selecting a shielding position in a binding-box area of the target;
(3) Combining both IRE and ORE;
s312, HIDE AND SEEK algorithm:
the picture is segmented into S-S grids, each grid is hidden by adopting a certain probability, the same picture is hidden in training, and different grid groups are hidden in each batch;
S313, grid Mask algorithm:
in order to avoid the problem that the complete target is deleted or the context information is deleted because the excessive deleting area exists in the step S31 and the step S32; setting four parameters of x, y, r and d through a Grid Mask:
Wherein r is the size of mask, M is the reserved pixel number, H, W is the image size; x and y are the area coordinates randomly generated on the image; the value of the non-shielding area is 1, the value of the shielding area is 0, a mask with the same resolution as the original image is generated, and then the mask is multiplied with the original image to obtain an image;
s314, mixup algorithm
The method comprises the steps of carrying out mixed enhancement on images, and mixing the images among different classes; the algorithm can be summarized as follows:
Where x 1、x2 is the pixel of the different image, lambda is the weight, Is the output pixel after the mixed class;
s315, cutmix algorithm
A portion of the region is randomly cropped and region pixel values of other data in the training set are randomly populated.
The sample identification phase comprises the following steps:
s1, inputting the processed building material sample image into a model, and extracting features through Resnet of pre-training and top layer removal on imagenet.
S2, marking the last 5 residual convolution modules in Resnet as { C 1,C2,C3,C4,C5 }, and respectively extracting output characteristic graphs of the 5 residual modules as { P 1,P2,P3,P4,P5 }; generating a feature map of an optimal size;
the step of generating the feature map comprises the following steps:
S21, performing 2 times nearest neighbor up-sampling on the deepest feature map P 5;
S22, extracting an output characteristic diagram P 4 of a residual convolution module C4 adjacent to the C5, and performing convolution dimension reduction processing of 1*1;
S23, fusing the pixel values of the corresponding parts with the feature map P 5 by a pixel value adding method of the corresponding parts;
s24, reducing an aliasing effect brought by up-sampling through 3*3 convolution of the fusion feature map;
s25, iterating the processes of S11 to S15 until a feature map with the optimal size is generated;
S3, generating candidate anchors with different sizes through RPN2,3,4 and 5 respectively; anchor areas are respectively set to 32 x 32, 64 x 64, 128 x 128, 256 x 256, and all the RPNs are generated by uniformly adopting 1: 1. 1: 2. 2:1, generating a plurality of candidate anchors, and then screening out the most complete anchor containing the target by utilizing a two-class function and bounding box regerssion function according to the RPN and real labeling, and correcting the most complete anchor to be used as the ROI;
S4, using the feature image layers output by different residual convolution modules as the input of ROI pooling layers, outputting the feature image by a deep convolution module by using the large-scale ROI, and using the ROI leave as a criterion output by a certain layer convolution module: Wherein w and h are the length and width of the ROI region, K 0 is the reference leave, and the small-scale ROI is set to be 5 by using the output characteristic diagram of the depth shallow convolution module, and the size of the representative characteristic diagram P 5 is set.
S5, transmitting the ROI generated in the S4 into ROI pooling layers, uniformly mapping the multi-scale ROI into proposal with the size of 7*7 by ROI pooling, and projecting the multi-scale ROI on an original building material sample image to generate proposal feature map, so that subsequent BBOX and CLS branches are convenient to process;
S6, calculating the category to which each sample belongs through the full connection layer and softmax, and outputting the highest category probability as a confidence coefficient by the CLS branch pair proposal feature map;
And S7, BBOX, correcting the proposal area by utilizing the bounding box regression function, generating a building material detection frame with more accurate positions, and identifying the material performance state of the building material sample.
There are, of course, many specific embodiments of the invention, not set forth herein. All technical solutions formed by equivalent substitution or equivalent transformation fall within the scope of the invention claimed.

Claims (2)

1. A building material sample identification method based on deep learning is characterized in that: comprises a model training stage and a sample recognition stage,
The model training phase comprises the following steps:
S1, collecting and marking building material samples, and manufacturing a building material sample data set; the data set comprises building material samples of each category and is divided into a training set, a testing set and an evaluation set;
s2, constructing a multi-scale information fusion convolutional neural network for building material sample identification;
S3, carrying out data enhancement on the sample data set in the S1: to obtain optimal model performance;
The sample identification phase comprises the following steps:
s1, inputting the processed building material sample image into a model, pre-training through imagenet, removing Resnet on the top layer, extracting features, and generating a feature map with the optimal size;
S2, the feature images generated in the S1 are respectively processed by RPN2,3,4 and 5to generate candidate anchors with different sizes; the anchor area is set, and all the anchors generated by the RPN uniformly adopt 1: 1. 1: 2. 2:1, generating a plurality of candidate anchors, and then screening out the most complete anchor containing the target by utilizing a two-class function and bounding box regerssion function according to the RPN and real labeling, and correcting the most complete anchor to be used as the ROI;
S3, using the feature image layers output by different residual convolution modules as the input of ROI pooling layers, outputting the feature image by a deep convolution module by using the large-scale ROI, and using the ROI leave as a criterion output by a corresponding layer convolution module:
Wherein w and h are the length and width of the ROI area, and K 0 is the reference leave;
S4, transmitting the ROI generated in the S3 into ROI pooling layers, uniformly mapping the multi-scale ROI into proposal with the same size by ROI pooling, and projecting the multi-scale ROI on an original building material sample image to generate proposal feature map, so that subsequent BBOX and CLS branches are convenient to process;
S5, calculating the category to which each sample belongs through the full connection layer and softmax, and outputting the highest category probability as a confidence coefficient by the CLS branch pair proposal feature map;
s6, BBOX, correcting a proposal area by utilizing a bounding box regression function, generating a building material detection frame with more accurate positions, and identifying the material performance state of a building material sample;
The model training stage S3 comprises the steps of:
s31, building material sample images under different scales and scenes are constructed by utilizing a combination of a plurality of data enhancement methods, existing data are expanded to simulate a complex identification scene, the learning of the model on detail characteristic information is improved, and the universality of the model is enhanced;
s32, setting initial weight as pre-training weight on imagenet, setting initial learning rate as 0.001, learning rate attenuation index as 0.1, batch_size as 16, and inputting image size;
s33, on a loss function, the RPN series module adopts two kinds of loss and regression loss; the CLS branch adopts multi-classification loss, and the BBOX branch adopts regression loss;
s34, training on a training set and a testing set by adopting an SGD random gradient descent optimizer in training until the performance of the model is optimal;
The step of generating the optimal feature map after the sample is extracted in the sample recognition stage S1 comprises the following steps:
S11, marking the last plurality of residual convolution modules in Resnet as { C 1,C2,C3,…Cn }, and respectively extracting output characteristic graphs of the residual modules as { P 1,P2,P3,…,Pn };
S12, performing 2 times nearest neighbor up-sampling on the deepest feature map P 5;
S13, extracting an output characteristic diagram P n-1 of a residual convolution module C n-1 adjacent to the C n, and performing 1*1 convolution dimension reduction processing;
S14, fusing the feature map P n by a pixel value adding method of the corresponding part;
S15, reducing an aliasing effect caused by up-sampling through 3*3 convolution of the fusion feature map;
S16, iterating the processes of S11 to S15 until the optimal size characteristic diagram is generated.
2. A method for identifying a sample of building material based on deep learning as claimed in claim 1, wherein: the method for enhancing the data in the S31 comprises the following steps:
S311, random Erasing algorithm:
(1) IRE, randomly selecting a shielding position on the whole target image;
(2) ORE, randomly selecting a shielding position in a binding-box area of the target;
(3) Combining both IRE and ORE;
s312, HIDE AND SEEK algorithm:
Dividing the picture into S-S grids, hiding each grid according to probability, hiding different grid groups for each batch of the same picture in training;
S313, grid Mask algorithm:
To avoid the problem that the complete target is deleted or the context information is lost due to the existence of the over-deleted region in S311 and S312; setting four parameters of x, y, r and d through a Grid Mask:
Wherein r is the size of mask, M is the reserved pixel number, H, W is the image size; x and y are the area coordinates randomly generated on the image; the value of the non-shielding area is 1, the value of the shielding area is 0, a mask with the same resolution as the original image is generated, and then the mask is multiplied with the original image to obtain an image;
s314, mixup algorithm
The method comprises the steps of carrying out mixed enhancement on images, and mixing the images among different classes; the algorithm is as follows:
Where x 1、x2 is the pixel of the different image and λ is the weight; Is the output pixel after the mixed class;
s315, cutmix algorithm
A portion of the region is randomly cropped and region pixel values of other data in the training set are randomly populated.
CN202011201983.7A 2020-11-02 2020-11-02 Deep learning-based building material sample identification method Active CN112329771B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011201983.7A CN112329771B (en) 2020-11-02 2020-11-02 Deep learning-based building material sample identification method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011201983.7A CN112329771B (en) 2020-11-02 2020-11-02 Deep learning-based building material sample identification method

Publications (2)

Publication Number Publication Date
CN112329771A CN112329771A (en) 2021-02-05
CN112329771B true CN112329771B (en) 2024-05-14

Family

ID=74323985

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011201983.7A Active CN112329771B (en) 2020-11-02 2020-11-02 Deep learning-based building material sample identification method

Country Status (1)

Country Link
CN (1) CN112329771B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112967296B (en) * 2021-03-10 2022-11-15 重庆理工大学 Point cloud dynamic region graph convolution method, classification method and segmentation method
CN113657202B (en) * 2021-07-28 2022-10-11 万翼科技有限公司 Component identification method, training set construction method, device, equipment and storage medium
CN113762229B (en) * 2021-11-10 2022-02-08 山东天亚达新材料科技有限公司 Intelligent identification method and system for building equipment in building site

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10303981B1 (en) * 2018-10-04 2019-05-28 StradVision, Inc. Learning method and testing method for R-CNN based object detector, and learning device and testing device using the same
CN110533024A (en) * 2019-07-10 2019-12-03 杭州电子科技大学 Biquadratic pond fine granularity image classification method based on multiple dimensioned ROI feature
CN111160249A (en) * 2019-12-30 2020-05-15 西北工业大学深圳研究院 Multi-class target detection method of optical remote sensing image based on cross-scale feature fusion

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10303981B1 (en) * 2018-10-04 2019-05-28 StradVision, Inc. Learning method and testing method for R-CNN based object detector, and learning device and testing device using the same
CN110533024A (en) * 2019-07-10 2019-12-03 杭州电子科技大学 Biquadratic pond fine granularity image classification method based on multiple dimensioned ROI feature
CN111160249A (en) * 2019-12-30 2020-05-15 西北工业大学深圳研究院 Multi-class target detection method of optical remote sensing image based on cross-scale feature fusion

Also Published As

Publication number Publication date
CN112329771A (en) 2021-02-05

Similar Documents

Publication Publication Date Title
CN109977918B (en) Target detection positioning optimization method based on unsupervised domain adaptation
CN108416266B (en) Method for rapidly identifying video behaviors by extracting moving object through optical flow
CN111753828B (en) Natural scene horizontal character detection method based on deep convolutional neural network
CN111027493B (en) Pedestrian detection method based on deep learning multi-network soft fusion
CN112329771B (en) Deep learning-based building material sample identification method
CN111179217A (en) Attention mechanism-based remote sensing image multi-scale target detection method
CN110738207A (en) character detection method for fusing character area edge information in character image
Kadam et al. Detection and localization of multiple image splicing using MobileNet V1
CN114758288B (en) Power distribution network engineering safety control detection method and device
CN108345850A (en) The scene text detection method of the territorial classification of stroke feature transformation and deep learning based on super-pixel
CN112926652B (en) Fish fine granularity image recognition method based on deep learning
CN107944459A (en) A kind of RGB D object identification methods
CN110390308B (en) Video behavior identification method based on space-time confrontation generation network
CN107784288A (en) A kind of iteration positioning formula method for detecting human face based on deep neural network
CN112052772A (en) Face shielding detection algorithm
CN105574545B (en) The semantic cutting method of street environment image various visual angles and device
CN109635726A (en) A kind of landslide identification method based on the symmetrical multiple dimensioned pond of depth network integration
CN118230175B (en) Real estate mapping data processing method and system based on artificial intelligence
CN115410081A (en) Multi-scale aggregated cloud and cloud shadow identification method, system, equipment and storage medium
CN114782417A (en) Real-time detection method for digital twin characteristics of fan based on edge enhanced image segmentation
CN115661754A (en) Pedestrian re-identification method based on dimension fusion attention
CN110135435B (en) Saliency detection method and device based on breadth learning system
Yang et al. An improved algorithm for the detection of fastening targets based on machine vision
CN113011506B (en) Texture image classification method based on deep fractal spectrum network
CN114445691A (en) Model training method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant