CN110046599A - Intelligent control method based on depth integration neural network pedestrian weight identification technology - Google Patents
Intelligent control method based on depth integration neural network pedestrian weight identification technology Download PDFInfo
- Publication number
- CN110046599A CN110046599A CN201910330924.0A CN201910330924A CN110046599A CN 110046599 A CN110046599 A CN 110046599A CN 201910330924 A CN201910330924 A CN 201910330924A CN 110046599 A CN110046599 A CN 110046599A
- Authority
- CN
- China
- Prior art keywords
- neural network
- pedestrian
- depth
- training
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 60
- 238000000034 method Methods 0.000 title claims abstract description 46
- 238000005516 engineering process Methods 0.000 title claims abstract description 27
- 230000010354 integration Effects 0.000 title claims abstract description 26
- 238000012549 training Methods 0.000 claims abstract description 42
- 230000009467 reduction Effects 0.000 claims abstract description 22
- 230000008569 process Effects 0.000 claims abstract description 9
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 5
- 239000013598 vector Substances 0.000 claims description 26
- 238000004422 calculation algorithm Methods 0.000 claims description 16
- 238000000605 extraction Methods 0.000 claims description 15
- 230000004913 activation Effects 0.000 claims description 13
- 238000010606 normalization Methods 0.000 claims description 9
- 230000006870 function Effects 0.000 claims description 8
- 230000004927 fusion Effects 0.000 claims description 8
- 238000012545 processing Methods 0.000 claims description 4
- 230000002708 enhancing effect Effects 0.000 claims description 3
- 238000013508 migration Methods 0.000 claims description 3
- 230000005012 migration Effects 0.000 claims description 3
- 238000012544 monitoring process Methods 0.000 description 5
- 238000012360 testing method Methods 0.000 description 5
- 238000010586 diagram Methods 0.000 description 3
- 235000013399 edible fruits Nutrition 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 238000000354 decomposition reaction Methods 0.000 description 1
- 238000009795 derivation Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000005538 encapsulation Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 238000011478 gradient descent method Methods 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000016507 interphase Effects 0.000 description 1
- 238000011835 investigation Methods 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 210000005036 nerve Anatomy 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000004801 process automation Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000012163 sequencing technique Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/56—Extraction of image or video features relating to colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Multimedia (AREA)
- General Health & Medical Sciences (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The present invention provides a kind of intelligent control method based on depth integration neural network pedestrian weight identification technology.After the method for the present invention carries out the pretreatment of color enhancement to the image of acquisition, traditional-handwork feature is extracted to it and depth residual error convolutional neural networks extract image, the traditional-handwork feature after dimensionality reduction and the depth characteristic for training intact neural network to extract are merged later, complete the identification to target pedestrian.Compared with prior art, the present invention accuracy of identification greatly improves, the pedestrian's weight recognizer for including in the present invention promotes recognition success rate to 81.74%, so that the technology completely can be practical.The process that pedestrian identifies again is to be automatically performed, and will not lose details because of personnel's fatigue.
Description
Technical field
The present invention relates to intelligent monitoring technology fields, more particularly to one kind to be identified again based on depth integration neural network pedestrian
The intelligent control method of technology.
Background technique
Pedestrian identify again also referred to as pedestrian identify again, be judged using computer vision technique be in image or video sequence
It is no that there are the technologies of specific pedestrian.Due to the difference between different picture pick-up devices, pedestrian has both characteristic rigid and flexible, appearance
Vulnerable to dress, scale, block, posture and visual angle etc. influence, therefore, pedestrian identify again as one in computer vision field both
With researching value extremely challenging heat subject again.
Currently, there are mainly two types of the research methods that pedestrian identifies again: method based on manual feature and being based on depth characteristic
Method.Method based on manual feature will usually find a kind of good feature of robustness to solve illumination and visual angle change band
The influence come, this method are mostly combined with metric learning;And the method based on depth characteristic is then obtained from training data emphatically
To the adaptive structure for being matched with label, to realize the identification end-to-end to target pedestrian.In recent years, convolutional neural networks mention
The depth characteristic taken has been proved to have good robustness.From pedestrian is introduced within 2014 again behind identification field, based on volume
The depth characteristic study of product neural network and depth distance measurement become the mainstream that pedestrian identifies again.Hereafter pedestrian identifies calculation again
Fado is to have carried out the improvement in structure to it, including Weihua Chen in proposition in 2017 based on triplet loss's
Intersect neural network, the singular value neural network (SVDNet) based on singular value decomposition that Yifan Sun was proposed in 2017 and
SpindleNet based on human synovial structure feature.These algorithms have original performance, but still have in robustness
It is to be hoisted.2016, Zheng Weishi proposed to be identified again in such a way that manual feature and depth characteristic merge.The algorithmic derivation
Loss function also demonstrates constraint of the traditional-handwork feature to neural network parameter to the back-propagation process of each parameter,
But the neural network is difficult to train in practical applications, convergence rate it is difficult to ensure that, and by hand feature and depth characteristic
It couples very faint.
Summary of the invention
According to technical problem set forth above, and provide a kind of based on depth integration neural network pedestrian's again identification technology
Intelligent monitor system.The present invention sets about in terms of manual feature and depth characteristic two, to effectively improve matching rate.The present invention adopts
Technological means is as follows:
As shown in Figure 1, a kind of intelligent control method based on depth integration neural network pedestrian weight identification technology, including such as
Lower step:
S1, image is obtained frame by frame from monitor video;
S2, the pretreatment that color enhancement is carried out to image send it to the local maximum for extracting manual feature later
In the neural network of structure and extraction depth characteristic;
S3A, traditional-handwork feature is extracted, specifically: feature vector is extracted using local maxima pond algorithm, using default
The grid of scale traverses whole picture with a fixed step size, and to the image zooming-out color characteristic and textural characteristics in each grid,
Then dimension-reduction treatment is carried out to entire feature vector;
S3B, image is extracted using depth residual error convolutional neural networks and Gauss pond;
S4, the traditional-handwork feature after dimensionality reduction and the depth characteristic for training intact neural network to extract are melted
It closes, completes the identification to target pedestrian;
S5, the pedestrian recognized is tracked, and repeats above-mentioned stream when tracking creditability is reduced to some threshold value
Cheng Jinhang identifies again, meanwhile, the mark for carrying out the preset style to the result of identification is shown.
Further, the pretreatment of the color enhancement specifically: use three respectively using multi-scale image enhancing algorithm
The Gaussian parameter of kind different scale carries out color enhancement to obtained image.
Further, in the extraction of the traditional-handwork feature, the extraction of color characteristic include: will carry out it is pretreated
Image is converted into hsv color space by RGB, and obtains the color histogram of image on this basis;The extraction of textural characteristics has
Body is to be extracted using three value pattern-coding technologies of Scale invariant, the dimension-reduction treatment specifically: using Gauss pond, by data
Stripping and slicing is carried out with default size, the moment of the orign and central moment for obtaining stripping and slicing data respectively later are to indicate data in block, thus real
Now to the dimensionality reduction of entire feature vector.
Further, the S3B specifically: dynamic fuzzy neural network cuts off pretreated image at random, later
It readjusts to pre-set dimension, and is passed in deep neural network ResNet50, input picture passes sequentially through convolutional layer, part
The processing of layer, the linear activation primitive of amendment, maximum pond is normalized, enters convolution module later and carries out dimension-reduction treatment, every process
One dimensionality reduction module, the size of image can all reduce 2*2.
Further, the S4 specifically: be separately connected the defeated of two vectors using the full connection fused layer of 4096 dimensions
Out, the operations such as regularization, batch normalization, nonlinear activation are carried out later, classifier are finally coupled to, using softmax letter
Number receives result and provides predicted value, to complete the identification to target pedestrian.
Further, the step S4 specifically comprises the following steps:
S41, the traditional-handwork feature after dimensionality reduction and the depth characteristic for training intact neural network to extract are carried out
Fusion, total character representation after fusion are as follows:
Fz=[Fh, Fd]
Wherein, FdFor the depth characteristic vector obtained in the ResNet50 model, FhFor by after the dimensionality reduction of Gauss pond
Traditional-handwork feature;
S42, traditional-handwork feature is connected in splicing layer, after being sequentially connected batch normalization, nonlinear activation function
The attended operation for carrying out classifier, is embodied as:
H () represents activation primitive in formula,And bfIndicate the weight coefficient and migration vector of articulamentum,
μzAnd σzThe characteristic for representing mean value and variance, is connected to classifier later, is used herein as the conduct point of softmax structure
One layer of class, specifically:
Wherein, x indicates that the vector that the output of one layer of neural network is constituted, θ are the parameter vector obtained by training, point
Female presence is in order to which prediction output is normalized.
Using cross entropy loss function, it is J that note, which intersects entropy loss, then has
Wherein, pkIndicate that the corresponding softmax output valve of k-th of output (also is understood as neural network prediction as k knot
The probability of fruit).
Further, the training of model is based on gradient and freezes coaching method and be trained, and specifically comprises the following steps:
Training ResNet50 depth residual error neural network and the full articulamentum of 2048 dimensions being correspondingly connected with point on data set
Class device;
After training to model convergence, the parameter of trained ResNet50 is imported in depth integration neural network, weight
New connection new full link sort and softmax network, then be trained on identical data set;It, will in training
Lower learning rate, emphasis training fusion ring is arranged in the network parameter for the extraction depth characteristic crossed using pre-training model initialization
The weight parameter of section and classifier, is finally completed the training to entire converged network.
Further, the data set is specially that Market1501 pedestrian identifies data set again.
The invention has the following advantages that
1, compared with prior art, the present invention accuracy of identification greatly improves, by depth characteristic ResNet50 and manual feature
Its accuracy of depth integration neural network after LOMO is combined improves 30% using ResNet50 than simple, the present invention
In include pedestrian's weight recognizer recognition success rate is promoted to 81.74% so that the technology completely can be practical.
2, the present invention using gradient freeze the training method time reduce 10 interphases so that fused neural network in training
Temporal disadvantage is eased, and improves convergent speed on the basis of guaranteeing accuracy rate.
3, the present invention only need on the basis of artificial monitoring installation one can mouse-based operation program, user is without being entangled with
The algorithm details being related to inside technology is identified and is tracked automatically to pedestrian, thus is used relatively simple.
4, the degree of automation of security field is improved.Conventional security monitoring needs a large amount of manpower to carry out naked eyes investigation,
Certain personnel that solve a case even will intently, and this technology can be by this process automation, and all processes are automatic complete
At, will not because personnel fatigue and lose details.
The present invention can be widely popularized in intelligent monitoring technology field based on the above reasons.
Detailed description of the invention
In order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, to embodiment or will show below
There is attached drawing needed in technical description to do simply to introduce, it should be apparent that, the accompanying drawings in the following description is this hair
Bright some embodiments for those of ordinary skill in the art without any creative labor, can be with
It obtains other drawings based on these drawings.
Fig. 1 is a kind of intelligent control method process based on depth integration neural network pedestrian weight identification technology of the present invention
Figure.
Fig. 2 is LOMO of the embodiment of the present invention and the specific design flow diagram of ResNet50.
Fig. 3 is depth integration of embodiment of the present invention detail figure.
Fig. 4 is that gradient of the embodiment of the present invention freezes coaching method superiority schematic diagram.
Fig. 5 is the comparison diagram of the embodiment of the present invention and other methods detection.
Specific embodiment
In order to make the object, technical scheme and advantages of the embodiment of the invention clearer, below in conjunction with the embodiment of the present invention
In attached drawing, technical scheme in the embodiment of the invention is clearly and completely described, it is clear that described embodiment is
A part of the embodiment of the present invention, instead of all the embodiments.Based on the embodiments of the present invention, those of ordinary skill in the art
Every other embodiment obtained without making creative work, shall fall within the protection scope of the present invention.
As shown in Figure 1, present embodiments providing a kind of intelligence based on depth integration neural network pedestrian weight identification technology
Monitoring method includes the following steps:
S1, it obtains image frame by frame from monitor video, specifically, reads the video flowing in camera and obtain letter frame by frame
Breath;Realize that obtain the image tensor of each frame, data are stayed for the reading to video flowing using the Python encapsulation of opencv
It is standby to use.
S2, the pretreatment that color enhancement is carried out to image send it to the local maximum for extracting manual feature later
In the neural network of structure and extraction depth characteristic;The pretreatment of the color enhancement specifically: enhanced using multi-scale image
Algorithm carries out color enhancement to obtained image using the Gaussian parameter of three kinds of different scales respectively.
S3A, traditional-handwork feature is extracted, specifically: feature vector is extracted using local maxima pond algorithm, using default
The grid of scale traverses whole picture with a fixed step size, and to the image zooming-out color characteristic and textural characteristics in each grid,
Then dimension-reduction treatment is carried out to entire feature vector;In the extraction of the traditional-handwork feature, the extraction of color characteristic include: by
It carried out pretreated image and hsv color space is converted by RGB, and obtain the color histogram of image on this basis;Line
The extraction for managing feature is specially to be extracted using three value pattern-coding technologies of Scale invariant, the dimension-reduction treatment specifically: use
Data are carried out stripping and slicing with default size by Gauss pond, and the moment of the orign and central moment for obtaining stripping and slicing data respectively later are to indicate
Data in block, to realize the dimensionality reduction to entire feature vector.
S3B, image is extracted using depth residual error convolutional neural networks and Gauss pond;Specifically, dynamic fuzzy
Neural network cuts off pretreated image at random, readjusts later to pre-set dimension, and is passed to deep neural network
In ResNet50, input picture passes sequentially through convolutional layer, part normalization layer, the place for correcting linear activation primitive, maximum pond
Reason enters convolution module later and carries out dimension-reduction treatment, and every to pass through a dimensionality reduction module, the size of image can all reduce 2*2.
As shown in Figure 2 and Figure 3, S4, by after dimensionality reduction traditional-handwork feature and training intact neural network extract
Depth characteristic is merged, and the identification to target pedestrian is completed;Specifically, connected respectively using the full connection fused layer of 4096 dimensions
The output of two vectors is connect, the operations such as regularization, batch normalization, nonlinear activation is carried out later, is finally coupled to classifier,
Result is received using softmax function and provides predicted value, to complete the identification to target pedestrian.
S5, the pedestrian recognized is tracked, and repeats above-mentioned stream when tracking creditability is reduced to some threshold value
Cheng Jinhang identifies again, meanwhile, the mark for carrying out the preset style to the result of identification is shown, specifically, will be identified using ssd algorithm
Obtained pedestrian tracks, and is just known again using above-mentioned process again when tracking creditability is reduced to some threshold value
Not, the result identified during this will be outlined by rectangle frame to be come.User interface is write using pyQt5, and is translated into operation
System executable file (such as exe file), it is user-friendly.
The step S4 specifically comprises the following steps:
S41, the traditional-handwork feature after dimensionality reduction and the depth characteristic for training intact neural network to extract are carried out
Fusion, total character representation after fusion are as follows:
Fz=[Fh, Fd]
Wherein, FdFor the depth characteristic vector obtained in the ResNet50 model, FhFor by after the dimensionality reduction of Gauss pond
Traditional-handwork feature;
S42, traditional-handwork feature is connected in splicing layer, after being sequentially connected batch normalization, nonlinear activation function
The attended operation for carrying out classifier, is embodied as:
H () represents activation primitive in formula,And bfIndicate the weight coefficient and migration vector of articulamentum,
μzAnd σzThe characteristic for representing mean value and variance, is connected to classifier later, is used herein as the conduct point of softmax structure
One layer of class, specifically:
Wherein, x indicates that the vector that the output of one layer of neural network is constituted, θ are the parameter vector obtained by training, point
Female presence is in order to which prediction output is normalized.
Using cross entropy loss function, it is J that note, which intersects entropy loss, then has
Wherein, pkIndicate that the corresponding softmax output valve of k-th of output (also is understood as neural network prediction as k knot
The probability of fruit).
As preferred embodiment, the training of model is freezed coaching method based on gradient and is trained, and specifically includes as follows
Step:
Training ResNet50 depth residual error neural network and the full articulamentum of 2048 dimensions being correspondingly connected with point on data set
Class device;
After training to model convergence, the parameter of trained ResNet50 is imported in depth integration neural network, weight
New connection new full link sort and softmax network, then be trained on identical data set;It, will in training
Lower learning rate, emphasis training fusion ring is arranged in the network parameter for the extraction depth characteristic crossed using pre-training model initialization
The weight parameter of section and classifier, is finally completed the training to entire converged network.
Wherein, the data set is specially that Market1501 pedestrian identifies data set again.The model of this paper is use
Market1501 data set is trained (train), test (test) and verifying (validation).Market1501 data set
It is data set to be identified by the pedestrian that Liang Zheng is issued again in 2015, at present and largest data set.The data set
It finds a view in supermarket, Tsinghua University, totally 38195 pedestrian's pictures that 6 different cameras take are related to including 1501 rows altogether
People, wherein training image has 12936, and from 751 pedestrians, test image shares 19732, from other 750
Pedestrian.
Embodiment 1:
S3B, for an original image having a size of 128*64, DFNN first by its it is random cut off, readjust later
Size is to 256*128 and is passed in deep neural network ResNet50.Input picture passes sequentially through convolutional layer, part normalization
The processing such as layer, the linear activation primitive of amendment, maximum pond, enter 16 convolution modules later.Although these convolution modules are extracted
Feature become increasingly complex, but DFFN simultaneously can the size to feature carry out keep and dimensionality reduction, convolution module (conv
Block) it is used to carry out dimensionality reduction to data.It is also 3 below identical with this dimensionality reduction module, every to pass through a dimensionality reduction module,
The size of image can all reduce 2*2.Therefore, it needs to input before input and output are connected by adder and utilizes convolutional layer
Reduce the processing of size, to guarantee normally to be added.
After the pretreatment by multi-scale image enhancing algorithm, a part can be represented as local scale for S3A, image
Constant background differential encoding, another part can be converted into the color characteristic of HSV space, later using both LOMO algorithm abstractions
Feature Descriptor, obtain the manual feature vector of 70770 dimensions.
Manual characteristic dimension is 70770, if being directly accessed full articulamentum, parameter amount can be very huge.Even if adopting
Take traditional dimensionality reduction means, it is also difficult to guarantee accurately to characterize entire traditional characteristic.Therefore one rationally may be used data
The characterization leaned on is very necessary.Maximum point is utilized in view of using in the LOMO algorithm that uses when extracting traditional characteristic
It measures to characterize the means of the distribution of all components, thus herein no longer using maximum pond, but to the mould of certain length
Block uses average pond.In order to which further more reliable guarantee main information is not lost, the classifier of characteristic is assumed herein
One-dimensional gaussian profile is closed, thus borrows expectation and two data of variance to characterize local distribution, 70 data are divided into 1 group, are asked
Its Gaussian parameter is taken, is linked together later, the feature vector of 2022 dimensions has just been obtained, while guaranteeing precision as far as possible
Complete dimensionality reduction.
S4, the traditional-handwork feature after dimensionality reduction and the depth characteristic for training intact neural network to extract are melted
It closes, is connected in the splicing layer of whole 2022+2048=4070 dimension, realize the strong confrontation and close coupling of two features.
Freeze to obtain final model after coaching method is trained based on gradient, it is tested, so that it is excellent to assess its
More property, specifically, pedestrian identify that common appraisal procedure is CMC (Accumulative Match Characteristic) again.
The appraisal procedure thinks that pedestrian's weight identification mission is target image and the sequencing of similarity task to selected digital image.CMC (k) is
It is the picture that similarity degree is k in weight identification mission, computer is just correctly obtained the frequency quilt of conclusion in first time identification
Referred to as Rank1, Rank1 can be used to measure the accuracy rate that pedestrian identifies again.
In multi-tag image classification, it can not be used simply using Average Accuracy used in single labeling
MAP (mean Average Precision) appraisal procedure.The appraisal procedure is to seek putting down by calculating accuracy rate and recall rate
Equal accuracy rate (AP, Average Precision), acquires mAP value again later.General mAP numerical value outline is lower than Rank1 numerical value.
As shown in figure 5, in each row, the image on the left side is sample, all recognition results arrange in descending order, red block
Indicate erroneous matching, blue box indicates correct matching.That is in the DFNN of the application, correct matching is indicated in addition to the 4th.And
SOMAnet only has the 2nd, 7 expressions correctly, other equal mistakes.For second group of example, in the DFNN of the application, the 2nd, 4,7,10
Indicate correct, other indicate mistake, and SOMAnet only has the 6th identification correctly, other identify mistake.Utilize what is introduced
Totally 12936 pictures are trained the 751 pedestrian ID given in Market1501 data set, utilize additional 750 later
Test by totally 19732 pictures by pedestrian ID.
In training process, detailed process is as follows for the training method that gradient is freezed:
1) training ResNet50 depth residual error neural network and 2048 dimensions being correspondingly connected on Market1501 data set
Spend full articulamentum classifier;
2) parameter that trained ResNet50 is imported in depth integration neural network reconnects new full connection point
Class and softmax network, then be trained on identical data set.
As shown in figure 4, when only using ResNet50 network training, ResNet50 e-learning rate, depth integration nerve
The learning rate of the full link sort device of network is disposed as 0.01.When to deep neural network training, the learning rate of ResNet50
It is arranged to lower 0.001.Optimization method in this model is based on the modified small lot gradient descent method of momentum, momentum
(momentum) numerical value is 0.9, and weight delay is 0.0005.After the completion of ResNet training, it is tested, Zhi Hou
After being trained to deep neural network, and second is carried out and has tested.Experiments have shown that obtained using ResNet50 merely
Rank1 is 51%, and the use of the Rank1 that depth integration neural network is tested is 82%, and 31% is improved from precision.
It is compared by the application depth integration neural network algorithm (DFNN) with pedestrian's weight recognizer in recent years,
The performance of this algorithm is verified, test result is as shown in table 1.
Table 1
As shown in table 1, the Rank1 of DFNN is 81.74%, is higher than any other algorithm in table.DFNN model is better than LOMO+
The classic algorithms such as XQDA, because being extracted depth characteristic using ResNet50.In addition, DFNN model ratio SpindleNet or
The models such as Triplet CNN will be got well, because having imported manual feature in a model to constrain deep neural network.As can be seen that
There is recognition accuracy after depth integration is obviously improved.
Finally, it should be noted that the above embodiments are only used to illustrate the technical solution of the present invention., rather than its limitations.To the greatest extent
Present invention has been described in detail with reference to the aforementioned embodiments for pipe, those skilled in the art should understand that it is still
It is possible to modify the technical solutions described in the foregoing embodiments, or some or all of the technical features is carried out
Equivalent replacement, and these are modified or replaceed, it does not separate the essence of the corresponding technical solution various embodiments of the present invention technical side
The range of case.
Claims (8)
1. a kind of intelligent control method based on depth integration neural network pedestrian weight identification technology, which is characterized in that including such as
Lower step:
S1, image is obtained frame by frame from monitor video;
S2, the pretreatment that color enhancement is carried out to image send it to the local maximum structure for extracting manual feature later
In the neural network of extraction depth characteristic;
S3A, traditional-handwork feature is extracted, specifically: feature vector is extracted using local maxima pond algorithm, uses default scale
Grid whole picture is traversed with a fixed step size, and to the image zooming-out color characteristic and textural characteristics in each grid, then
Dimension-reduction treatment is carried out to entire feature vector;
S3B, image is extracted using depth residual error convolutional neural networks and Gauss pond;
S4, the traditional-handwork feature after dimensionality reduction and the depth characteristic for training intact neural network to extract are merged,
Complete the identification to target pedestrian;
S5, the pedestrian recognized is tracked, and repeat when tracking creditability is reduced to some threshold value above-mentioned process into
It is capable to identify again, meanwhile, the mark for carrying out the preset style to the result of identification is shown.
2. the intelligent control method according to claim 1 based on depth integration neural network pedestrian weight identification technology,
It is characterized in that, the pretreatment of the color enhancement specifically: use three kinds of different rulers respectively using multi-scale image enhancing algorithm
The Gaussian parameter of degree carries out color enhancement to obtained image.
3. the intelligent control method according to claim 2 based on depth integration neural network pedestrian weight identification technology,
Be characterized in that, in the extraction of the traditional-handwork feature, the extraction of color characteristic include: will carry out pretreated image by
RGB is converted into hsv color space, and obtains the color histogram of image on this basis;The extraction of textural characteristics is specially to make
It is extracted with three value pattern-coding technologies of Scale invariant, the dimension-reduction treatment specifically: using Gauss pond, by data with default
Size carries out stripping and slicing, and the moment of the orign and central moment for obtaining stripping and slicing data respectively later are to indicate data in block, to realize to whole
The dimensionality reduction of a feature vector.
4. the intelligent control method according to claim 2 based on depth integration neural network pedestrian weight identification technology,
It is characterized in that, the S3B specifically: dynamic fuzzy neural network cuts off pretreated image at random, readjusts later
It to pre-set dimension, and is passed in deep neural network ResNet50, input picture passes sequentially through convolutional layer, part normalization
The processing of layer, the linear activation primitive of amendment, maximum pond enters convolution module later and carries out dimension-reduction treatment, every by a drop
Module is tieed up, the size of image can all reduce 2*2.
5. the intelligence prison according to any one of claims 1 to 4 based on depth integration neural network pedestrian weight identification technology
Prosecutor method, which is characterized in that the S4 specifically: be separately connected the defeated of two vectors using the full connection fused layer of 4096 dimensions
Out, the operations such as regularization, batch normalization, nonlinear activation are carried out later, classifier are finally coupled to, using softmax letter
Number receives result and provides predicted value, to complete the identification to target pedestrian.
6. the intelligent control method according to claim 5 based on depth integration neural network pedestrian weight identification technology,
It is characterized in that, the step S4 specifically comprises the following steps:
S41, the traditional-handwork feature after dimensionality reduction and the depth characteristic for training intact neural network to extract are merged,
Total character representation after fusion are as follows:
Fz=[Fh, Fd]
Wherein, FdFor the depth characteristic vector obtained in the ResNet50 model, FhFor by the tradition after the dimensionality reduction of Gauss pond
Manual feature;
S42, traditional-handwork feature is connected in splicing layer, is carried out after being sequentially connected batch normalization, nonlinear activation function
The attended operation of classifier, is embodied as:
H () represents activation primitive in formula,And bfIndicate the weight coefficient and migration vector of articulamentum,
μzAnd σzThe characteristic for representing mean value and variance, is connected to classifier later, is used herein as softmax structure as classification
One layer, specifically:
Wherein, x indicates that the vector that the output of one layer of neural network is constituted, θ are the parameter vector obtained by training, denominator
In the presence of be in order to prediction output be normalized.
Using cross entropy loss function, it is J that note, which intersects entropy loss, then has
Wherein, pkIndicate that (also be understood as neural network prediction is k result to the corresponding softmax output valve of k-th of output
Probability).
7. the intelligent control method according to claim 1 based on depth integration neural network pedestrian weight identification technology,
It is characterized in that, the training of model is freezed coaching method based on gradient and is trained, and specifically comprises the following steps:
Training ResNet50 depth residual error neural network and the full articulamentum classifier of 2048 dimensions being correspondingly connected on data set;
After training to model convergence, the parameter of trained ResNet50 is imported in depth integration neural network, is connected again
New full link sort and softmax network are connect, then is trained on identical data set;In training, will utilize
The network parameter of the extraction depth characteristic that pre-training model initialization is crossed is arranged lower learning rate, emphasis training fusion link and
The weight parameter of classifier is finally completed the training to entire converged network.
8. the intelligent control method according to claim 7 based on depth integration neural network pedestrian weight identification technology,
It is characterized in that, the data set is specially that Market1501 pedestrian identifies data set again.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910330924.0A CN110046599A (en) | 2019-04-23 | 2019-04-23 | Intelligent control method based on depth integration neural network pedestrian weight identification technology |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910330924.0A CN110046599A (en) | 2019-04-23 | 2019-04-23 | Intelligent control method based on depth integration neural network pedestrian weight identification technology |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110046599A true CN110046599A (en) | 2019-07-23 |
Family
ID=67278848
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910330924.0A Pending CN110046599A (en) | 2019-04-23 | 2019-04-23 | Intelligent control method based on depth integration neural network pedestrian weight identification technology |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110046599A (en) |
Cited By (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110414430A (en) * | 2019-07-29 | 2019-11-05 | 郑州信大先进技术研究院 | A kind of pedestrian recognition methods and device again based on the fusion of more ratios |
CN110473185A (en) * | 2019-08-07 | 2019-11-19 | Oppo广东移动通信有限公司 | Image processing method and device, electronic equipment, computer readable storage medium |
CN110688966A (en) * | 2019-09-30 | 2020-01-14 | 华东师范大学 | Semantic-guided pedestrian re-identification method |
CN110728238A (en) * | 2019-10-12 | 2020-01-24 | 安徽工程大学 | Personnel re-detection method of fusion type neural network |
CN111079666A (en) * | 2019-12-20 | 2020-04-28 | 广州市鑫广飞信息科技有限公司 | Ground object identification method, device, equipment and storage medium |
CN111274873A (en) * | 2020-01-09 | 2020-06-12 | 济南浪潮高新科技投资发展有限公司 | Pedestrian re-identification method based on artificial feature and depth feature fusion |
CN111325709A (en) * | 2019-12-26 | 2020-06-23 | 联博智能科技有限公司 | Wireless capsule endoscope image detection system and detection method |
CN111460891A (en) * | 2020-03-01 | 2020-07-28 | 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) | Automatic driving-oriented vehicle-road cooperative pedestrian re-identification method and system |
CN111488973A (en) * | 2020-04-09 | 2020-08-04 | 陕西师范大学 | Preprocessing method and device for neural network data |
CN111610518A (en) * | 2020-06-09 | 2020-09-01 | 电子科技大学 | Secondary radar signal denoising method based on depth residual separation convolutional network |
CN111681676A (en) * | 2020-06-09 | 2020-09-18 | 杭州星合尚世影视传媒有限公司 | Method, system and device for identifying and constructing audio frequency by video object and readable storage medium |
CN111797732A (en) * | 2020-06-22 | 2020-10-20 | 电子科技大学 | Video motion identification anti-attack method insensitive to sampling |
CN111985504A (en) * | 2020-08-17 | 2020-11-24 | 中国平安人寿保险股份有限公司 | Copying detection method, device, equipment and medium based on artificial intelligence |
CN112287965A (en) * | 2020-09-21 | 2021-01-29 | 卓尔智联(武汉)研究院有限公司 | Image quality detection model training method and device and computer equipment |
CN112749747A (en) * | 2021-01-13 | 2021-05-04 | 上海交通大学 | Garbage classification quality evaluation method and system |
CN112766180A (en) * | 2021-01-22 | 2021-05-07 | 重庆邮电大学 | Pedestrian re-identification method based on feature fusion and multi-core learning |
CN113112403A (en) * | 2021-03-31 | 2021-07-13 | 国网山东省电力公司枣庄供电公司 | Infrared image splicing method, system, medium and electronic equipment |
CN113420697A (en) * | 2021-07-01 | 2021-09-21 | 中科人工智能创新技术研究院(青岛)有限公司 | Reloading video pedestrian re-identification method and system based on appearance and shape characteristics |
CN113536995A (en) * | 2021-06-30 | 2021-10-22 | 河南大学 | Pedestrian re-identification method based on feature mapping space and sample judgment |
WO2024160215A1 (en) * | 2023-01-31 | 2024-08-08 | 华为技术有限公司 | Data processing method and apparatus |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9401020B1 (en) * | 2015-05-01 | 2016-07-26 | London Health Science Centre Research Inc | Multi-modality vertebra recognition |
CN107301380A (en) * | 2017-06-01 | 2017-10-27 | 华南理工大学 | One kind is used for pedestrian in video monitoring scene and knows method for distinguishing again |
US20180181842A1 (en) * | 2016-12-22 | 2018-06-28 | TCL Research America Inc. | Method and device for quasi-gibbs structure sampling by deep permutation for person identity inference |
CN108596010A (en) * | 2017-12-31 | 2018-09-28 | 厦门大学 | The implementation method of pedestrian's weight identifying system |
CN108960141A (en) * | 2018-07-04 | 2018-12-07 | 国家新闻出版广电总局广播科学研究院 | Pedestrian's recognition methods again based on enhanced depth convolutional neural networks |
CN109271888A (en) * | 2018-08-29 | 2019-01-25 | 汉王科技股份有限公司 | Personal identification method, device, electronic equipment based on gait |
CN109508731A (en) * | 2018-10-09 | 2019-03-22 | 中山大学 | A kind of vehicle based on fusion feature recognition methods, system and device again |
-
2019
- 2019-04-23 CN CN201910330924.0A patent/CN110046599A/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9401020B1 (en) * | 2015-05-01 | 2016-07-26 | London Health Science Centre Research Inc | Multi-modality vertebra recognition |
US20180181842A1 (en) * | 2016-12-22 | 2018-06-28 | TCL Research America Inc. | Method and device for quasi-gibbs structure sampling by deep permutation for person identity inference |
CN107301380A (en) * | 2017-06-01 | 2017-10-27 | 华南理工大学 | One kind is used for pedestrian in video monitoring scene and knows method for distinguishing again |
CN108596010A (en) * | 2017-12-31 | 2018-09-28 | 厦门大学 | The implementation method of pedestrian's weight identifying system |
CN108960141A (en) * | 2018-07-04 | 2018-12-07 | 国家新闻出版广电总局广播科学研究院 | Pedestrian's recognition methods again based on enhanced depth convolutional neural networks |
CN109271888A (en) * | 2018-08-29 | 2019-01-25 | 汉王科技股份有限公司 | Personal identification method, device, electronic equipment based on gait |
CN109508731A (en) * | 2018-10-09 | 2019-03-22 | 中山大学 | A kind of vehicle based on fusion feature recognition methods, system and device again |
Non-Patent Citations (3)
Title |
---|
SHANGXUAN WU.ET AL: ""An Enhanced Deep Feature Representation for Person Re-identification"", 《IEEE》 * |
SHENGCAI LIAO.ET AL: ""Modeling Pixel Process with Scale Invariant Local Patterns for Background Subtraction in Complex Scenes"", 《IEEE》 * |
雷峰网LEIPHONE: ""如何用PyTorch训练图像分类器"", 《HTTPS://BAIJIAHAO.BAIDU.COM/S?ID=1618188613267031927&WFR=SPIDER&FOR=PC》 * |
Cited By (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110414430A (en) * | 2019-07-29 | 2019-11-05 | 郑州信大先进技术研究院 | A kind of pedestrian recognition methods and device again based on the fusion of more ratios |
CN110414430B (en) * | 2019-07-29 | 2022-10-04 | 郑州信大先进技术研究院 | Pedestrian re-identification method and device based on multi-proportion fusion |
CN110473185B (en) * | 2019-08-07 | 2022-03-15 | Oppo广东移动通信有限公司 | Image processing method and device, electronic equipment and computer readable storage medium |
CN110473185A (en) * | 2019-08-07 | 2019-11-19 | Oppo广东移动通信有限公司 | Image processing method and device, electronic equipment, computer readable storage medium |
CN110688966A (en) * | 2019-09-30 | 2020-01-14 | 华东师范大学 | Semantic-guided pedestrian re-identification method |
CN110688966B (en) * | 2019-09-30 | 2024-01-09 | 华东师范大学 | Semantic guidance pedestrian re-recognition method |
CN110728238A (en) * | 2019-10-12 | 2020-01-24 | 安徽工程大学 | Personnel re-detection method of fusion type neural network |
CN111079666A (en) * | 2019-12-20 | 2020-04-28 | 广州市鑫广飞信息科技有限公司 | Ground object identification method, device, equipment and storage medium |
CN111079666B (en) * | 2019-12-20 | 2024-03-19 | 广州市鑫广飞信息科技有限公司 | Ground object identification method, device, equipment and storage medium |
CN111325709A (en) * | 2019-12-26 | 2020-06-23 | 联博智能科技有限公司 | Wireless capsule endoscope image detection system and detection method |
CN111274873A (en) * | 2020-01-09 | 2020-06-12 | 济南浪潮高新科技投资发展有限公司 | Pedestrian re-identification method based on artificial feature and depth feature fusion |
CN111460891A (en) * | 2020-03-01 | 2020-07-28 | 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) | Automatic driving-oriented vehicle-road cooperative pedestrian re-identification method and system |
CN111488973A (en) * | 2020-04-09 | 2020-08-04 | 陕西师范大学 | Preprocessing method and device for neural network data |
CN111488973B (en) * | 2020-04-09 | 2023-08-18 | 陕西师范大学 | Preprocessing method and device for neural network data |
CN111610518B (en) * | 2020-06-09 | 2022-08-05 | 电子科技大学 | Secondary radar signal denoising method based on depth residual separation convolutional network |
CN111610518A (en) * | 2020-06-09 | 2020-09-01 | 电子科技大学 | Secondary radar signal denoising method based on depth residual separation convolutional network |
CN111681676A (en) * | 2020-06-09 | 2020-09-18 | 杭州星合尚世影视传媒有限公司 | Method, system and device for identifying and constructing audio frequency by video object and readable storage medium |
CN111681676B (en) * | 2020-06-09 | 2023-08-08 | 杭州星合尚世影视传媒有限公司 | Method, system, device and readable storage medium for constructing audio frequency by video object identification |
CN111797732B (en) * | 2020-06-22 | 2022-03-25 | 电子科技大学 | Video motion identification anti-attack method insensitive to sampling |
CN111797732A (en) * | 2020-06-22 | 2020-10-20 | 电子科技大学 | Video motion identification anti-attack method insensitive to sampling |
CN111985504B (en) * | 2020-08-17 | 2021-05-11 | 中国平安人寿保险股份有限公司 | Copying detection method, device, equipment and medium based on artificial intelligence |
CN111985504A (en) * | 2020-08-17 | 2020-11-24 | 中国平安人寿保险股份有限公司 | Copying detection method, device, equipment and medium based on artificial intelligence |
CN112287965A (en) * | 2020-09-21 | 2021-01-29 | 卓尔智联(武汉)研究院有限公司 | Image quality detection model training method and device and computer equipment |
CN112749747B (en) * | 2021-01-13 | 2022-11-11 | 上海交通大学 | Garbage classification quality evaluation method and system |
CN112749747A (en) * | 2021-01-13 | 2021-05-04 | 上海交通大学 | Garbage classification quality evaluation method and system |
CN112766180B (en) * | 2021-01-22 | 2022-07-12 | 重庆邮电大学 | Pedestrian re-identification method based on feature fusion and multi-core learning |
CN112766180A (en) * | 2021-01-22 | 2021-05-07 | 重庆邮电大学 | Pedestrian re-identification method based on feature fusion and multi-core learning |
CN113112403A (en) * | 2021-03-31 | 2021-07-13 | 国网山东省电力公司枣庄供电公司 | Infrared image splicing method, system, medium and electronic equipment |
CN113536995A (en) * | 2021-06-30 | 2021-10-22 | 河南大学 | Pedestrian re-identification method based on feature mapping space and sample judgment |
CN113536995B (en) * | 2021-06-30 | 2022-11-18 | 河南大学 | Pedestrian re-identification method based on feature mapping space and sample judgment |
CN113420697A (en) * | 2021-07-01 | 2021-09-21 | 中科人工智能创新技术研究院(青岛)有限公司 | Reloading video pedestrian re-identification method and system based on appearance and shape characteristics |
WO2024160215A1 (en) * | 2023-01-31 | 2024-08-08 | 华为技术有限公司 | Data processing method and apparatus |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110046599A (en) | Intelligent control method based on depth integration neural network pedestrian weight identification technology | |
CN108520226B (en) | Pedestrian re-identification method based on body decomposition and significance detection | |
CN108921107B (en) | Pedestrian re-identification method based on sequencing loss and Simese network | |
CN110532970B (en) | Age and gender attribute analysis method, system, equipment and medium for 2D images of human faces | |
CN102324025B (en) | Human face detection and tracking method based on Gaussian skin color model and feature analysis | |
CN101339607B (en) | Human face recognition method and system, human face recognition model training method and system | |
CN109146831A (en) | Remote sensing image fusion method and system based on double branch deep learning networks | |
CN111368690B (en) | Deep learning-based video image ship detection method and system under influence of sea waves | |
CN109740413A (en) | Pedestrian recognition methods, device, computer equipment and computer storage medium again | |
CN110674785A (en) | Multi-person posture analysis method based on human body key point tracking | |
CN110059694A (en) | The intelligent identification Method of lteral data under power industry complex scene | |
CN109101865A (en) | A kind of recognition methods again of the pedestrian based on deep learning | |
CN109190475B (en) | Face recognition network and pedestrian re-recognition network collaborative training method | |
CN113673510B (en) | Target detection method combining feature point and anchor frame joint prediction and regression | |
CN109635727A (en) | A kind of facial expression recognizing method and device | |
CN107145845A (en) | The pedestrian detection method merged based on deep learning and multi-characteristic points | |
CN111563452A (en) | Multi-human body posture detection and state discrimination method based on example segmentation | |
CN102262729B (en) | Fused face recognition method based on integrated learning | |
CN110263768A (en) | A kind of face identification method based on depth residual error network | |
CN111275058B (en) | Safety helmet wearing and color identification method and device based on pedestrian re-identification | |
CN110046544A (en) | Digital gesture identification method based on convolutional neural networks | |
CN109325408A (en) | A kind of gesture judging method and storage medium | |
CN111368768A (en) | Human body key point-based employee gesture guidance detection method | |
CN113780145A (en) | Sperm morphology detection method, sperm morphology detection device, computer equipment and storage medium | |
CN113221770A (en) | Cross-domain pedestrian re-identification method and system based on multi-feature hybrid learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190723 |