Nothing Special   »   [go: up one dir, main page]

CN111738164A - Pedestrian detection method based on deep learning - Google Patents

Pedestrian detection method based on deep learning Download PDF

Info

Publication number
CN111738164A
CN111738164A CN202010586392.XA CN202010586392A CN111738164A CN 111738164 A CN111738164 A CN 111738164A CN 202010586392 A CN202010586392 A CN 202010586392A CN 111738164 A CN111738164 A CN 111738164A
Authority
CN
China
Prior art keywords
pedestrian
matrix
pixel
height
deep learning
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010586392.XA
Other languages
Chinese (zh)
Other versions
CN111738164B (en
Inventor
陈凌霄
廖宏
肖杨
杜奕霖
杨程
彭一峰
黄铭斌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangxi Computing Center Co ltd
Original Assignee
Guangxi Computing Center Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangxi Computing Center Co ltd filed Critical Guangxi Computing Center Co ltd
Priority to CN202010586392.XA priority Critical patent/CN111738164B/en
Publication of CN111738164A publication Critical patent/CN111738164A/en
Application granted granted Critical
Publication of CN111738164B publication Critical patent/CN111738164B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to the technical field of pedestrian detection systems, in particular to a pedestrian detection method based on deep learning. The method comprises the following steps: s1, arranging a camera, acquiring a video of a current analysis scene, and marking a surrounding frame of a pedestrian in the video as a training data set for deep learning; s2, calculating a corresponding pedestrian pixel width matrix W and a pedestrian pixel height matrix H according to the pixel width and height of the pedestrian in each position of the image in the training data set; s3, calculating the Scale Scale and the aspect Ratio of the anchor frame of each region position in the image by using the matrix W and the matrix H; s4 deep learning Faster Rcnn model training is carried out; s5: and acquiring the position coordinates and width and height of each pedestrian, and counting the total number of people or the local density of the current scene by adopting a trained Faster Rcnn model. The invention improves the detection accuracy and reduces the false detection.

Description

Pedestrian detection method based on deep learning
Technical Field
The invention relates to the field of computer vision and deep learning, in particular to the technical field of pedestrian detection systems for security monitoring and intelligent video analysis, and particularly relates to a pedestrian detection method based on deep learning.
Background
With the rapid development of computer science technology, the application of computer vision technology to automatically and intelligently analyze pedestrian targets in a monitoring scene gradually becomes a research hotspot. Through pedestrian's detection, can carry out high density state warning, count pedestrian's flow etc. and replace traditional manpower control on duty. In recent years, the deep learning neural network has been developed in a striding way, and gradually replaces the traditional target detection method. Traditional algorithms often require manual design of features according to a specific domain and generally can only detect a certain class. The deep learning enables the deep neural network to automatically learn the target characteristics through supervised learning of a large-scale sample set, and not only has higher precision than that of the traditional algorithm, but also can simultaneously detect various targets. At present, target detection methods based on deep learning are divided into two types, one is a target detection method based on a one-step method, and position coordinates of a target, such as SSD, YOLO and the like, are directly regressed from an image, but the algorithm is often low in detection precision of small targets, and a large amount of missing detection occurs; in another method, which is a target detection method based on a two-step method, such as fast-RCNN, the algorithm first extracts a feature map from an image through a base network, then generates a candidate region through an RPN network, and then performs target classification through a classification algorithm. The multi-target deep learning detection algorithm is called a general target detection algorithm, and has poor detection effect on pedestrians in a monitoring scene. Disclosure of Invention
In order to solve the problems, the invention provides a pedestrian detection method based on deep learning, which comprises the following specific technical scheme:
a pedestrian detection method based on deep learning comprises the following steps:
s1: arranging a camera at a scene to be analyzed, acquiring a video of the current analysis scene, marking a surrounding frame of a pedestrian in the video, and storing a pedestrian marking data file as a training data set for deep learning;
s2: calculating a corresponding pedestrian pixel width matrix W and a pedestrian pixel height matrix H according to the pixel width and height of the pedestrian at each position in the image in the training data set;
s3: calculating the Scale and the aspect Ratio of an anchor frame of each region position in the image by using the pedestrian pixel width matrix W and the pedestrian pixel height matrix H;
s4: performing deep learning FasterRcnn model training on a current scene image data set to obtain weight parameters of a trained FasterRcnn model;
s5: deploying a pedestrian detection system for the current scene, acquiring the position coordinate and width and height of each pedestrian, and counting the total number of people or the local density of the current scene by adopting a trained Faster Rcnn model.
Preferably, the step S2 includes the steps of:
s21: reading a locally stored pedestrian marking data file;
s22: creating a pedestrian pixel width matrix W and a pedestrian pixel height matrix H, wherein the dimensions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H respectively correspond to the pixel width and height of the camera image;
s23: if a certain pixel point in the training data set image is the central point of a pedestrian bounding box, writing the pixel width value and the pixel height value of the bounding box into corresponding positions of a pedestrian pixel width matrix W and a pedestrian pixel height matrix H respectively; if the corresponding positions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H have numerical values, calculating the addition and average value of the written value and the stored value, and then writing the sum and average value into the corresponding positions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H;
s24: and checking whether 0 value points exist in the pedestrian pixel width matrix W and the pedestrian pixel height matrix H obtained in the steps, if so, performing adjacent pixel distance interpolation to calculate the pixel value of the current position, and writing the pixel value into the corresponding positions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H.
Preferably, the step S3 includes the steps of:
s31: performing n-n meshing on the pedestrian pixel width matrix W and the pedestrian pixel height matrix H, and traversing the training data set to obtain the pedestrian maximum pixel height H1 and the pedestrian minimum pixel height H2 in the current analysis scene, wherein n = H2/H1, and n is an integer obtained by downward rounding;
s32: calculating the average pixel value in each grid of the pedestrian pixel width matrix W as the Scale of the corresponding training RPN network;
s33: calculating the average value of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H in each grid respectively as
Figure 100002_DEST_PATH_IMAGE001
And
Figure 18813DEST_PATH_IMAGE002
then the aspect Ratio within each grid is {1:
Figure 100002_DEST_PATH_IMAGE003
/
Figure 322755DEST_PATH_IMAGE002
}。
preferably, in the step S4, in the RPN network training process, the Scale and the aspect Ratio of the anchor frame are obtained through coordinates of a center of a candidate frame generated by the RPN network.
Preferably, the step S5 includes the steps of:
s51: automatically acquiring a real-time video stream of the analysis scene by using the camera sdk, and decoding to obtain an image of the analysis scene;
s52: loading the weight parameters of the trained FasterRcnn model, and inputting the decoded image into the trained FasterRcnn model to obtain a convolution characteristic diagram;
s53: and performing RPN network reasoning on the convolution characteristic diagram to obtain a target candidate region, and performing Fast RCNN classification and region frame regression to obtain a target classification score and pedestrian width and height.
The invention has the beneficial effects that: the target pedestrian detection effect under the monitoring scene is improved, the detection accuracy is improved, and the false detection is reduced.
Drawings
FIG. 1 is a schematic flow diagram of the present invention;
FIG. 2 is a schematic diagram of selecting an anchor point box according to the present invention.
Detailed Description
For a better understanding of the present invention, reference is made to the following detailed description taken in conjunction with the accompanying drawings in which:
as shown in fig. 1, a pedestrian detection method based on deep learning includes the steps of:
s1: arranging a camera at a scene needing analysis, acquiring a video record of the current analysis scene, performing frame skipping processing on the video every 50 frames because pedestrians walk slowly in the scene, and extracting images containing the pedestrians from the video, wherein the images form an original data set. And (3) carrying out surrounding frame labeling on the pedestrians in the video by using open source labeling software Labelimg, and storing a pedestrian labeling data file as a training data set for deep learning.
S2: calculating a corresponding pedestrian pixel width matrix W and a pedestrian pixel height matrix H according to the pixel width and height of the pedestrian at each position in the image in the training data set; the method comprises the following steps:
s21: reading a locally stored pedestrian marking data file;
s22: creating a pedestrian pixel width matrix W and a pedestrian pixel height matrix H, wherein the dimensions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H respectively correspond to the pixel width and height of the camera image;
s23: if a certain pixel point in the training data set image is the central point of a pedestrian bounding box, writing the pixel width value and the pixel height value of the bounding box into corresponding positions of a pedestrian pixel width matrix W and a pedestrian pixel height matrix H respectively; if the corresponding positions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H have numerical values, calculating the addition and average value of the written value and the stored value, and then writing the sum and average value into the corresponding positions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H;
s24: and checking whether 0 value points exist in the pedestrian pixel width matrix W and the pedestrian pixel height matrix H obtained in the steps, if so, performing adjacent pixel distance interpolation to calculate the pixel value of the current position, and writing the pixel value into the corresponding positions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H.
S3: calculating the Scale and the aspect Ratio of an anchor frame of each region position in the image by using the pedestrian pixel width matrix W and the pedestrian pixel height matrix H; the method comprises the following steps:
s31: performing n-n gridding on the pedestrian pixel width matrix W and the pedestrian pixel height matrix H, and traversing a training data set to obtain the maximum pixel height H1 (near target) and the minimum pixel height H2 (far target) of a pedestrian in the current analysis scene, wherein n = H2/H1, and n is an integer obtained by downward integration;
s32: calculating the average pixel value in each grid of the pedestrian pixel width matrix W as the Scale of the corresponding training RPN network;
s33: calculating the average value of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H in each grid respectively as
Figure 660195DEST_PATH_IMAGE001
And
Figure 365983DEST_PATH_IMAGE002
then the aspect Ratio within each grid is {1:
Figure 701150DEST_PATH_IMAGE003
/
Figure 492388DEST_PATH_IMAGE002
}。
s4: performing deep learning fast Rcnn model training on a current scene image data set to obtain weight parameters of a trained fast Rcnn model; the method specifically comprises the following steps: the disclosed VGG-16 model was used as a feature extraction infrastructure network structure, the base of which was used for 13 convolutional layers. In the RPN network training process, the Scale and the aspect Ratio of the anchor frame are obtained through the center coordinates of the candidate frame generated by the RPN network. As shown in fig. 2, where a rectangular frame 1 is generated by the method, pedestrians are compactly wrapped. The rectangular frames 2-4 are rectangular frames generated by a conventional method, and have a large difference from the actual pedestrian size. Because the pixel width and height of the candidate frame generated in the RPN network training process are similar to the pedestrian pixel height in the analysis scene, the accuracy of the RPN network generating the candidate frame is improved, and the final FasterRCNN pedestrian detection accuracy is further improved. In the step, a final model is trained by adopting a fast RCNN end-to-end back propagation algorithm. After 10 ten thousand iterations, the weight parameters of the trained FasterRcnn model are saved to the local hard disk.
S5: deploying a pedestrian detection system for the current scene, acquiring the position coordinate and width and height of each pedestrian, and counting the total number of people or local density of the current scene by adopting a trained Faster Rcnn model; the method comprises the following steps:
s51: automatically acquiring a real-time video stream of the analysis scene by using the camera sdk, and decoding to obtain an image of the analysis scene;
s52: loading the weight parameters of the trained FasterRcnn model, and inputting the decoded image into the trained FasterRcnn model to obtain a convolution characteristic diagram;
s53: and performing RPN network reasoning on the convolution characteristic diagram to obtain a target candidate region, and performing Fast RCNN classification and region frame regression to obtain a target classification score and pedestrian width and height.
In this embodiment, intel i7-8700k CPU and intevada 1080Ti GPU are used for training, and a traditional fast-rcnn neural network algorithm and the method of the present invention are used to detect and compare the citrypersons data set, and the results are shown in table 1:
TABLE 1 comparative results
CityPersons dataset
Figure DEST_PATH_IMAGE005
Faster-rcnn 76.3
Method for producing a composite material 82.6
Figure 633519DEST_PATH_IMAGE006
The calculation method comprises the following steps:
1. setting an IOU threshold for a certain category, where the threshold is set to 0.5 in this embodiment, (meaning that the intersection ratio of a prediction boundary frame detected as a human target and a real boundary frame is greater than 0.5, the boundary frame greater than the threshold is regarded as TP, and the other boundary frames are regarded as FP, TP represents a detected positive sample, FP represents a detected positive sample, FN is obtained by subtracting TP from the number of real positive samples in a test sample, FN is a missed positive sample), and counting the number of group truth frames (denominator for calculating a recall ratio) M and the number of detection frames N; the IOU represents the intersection-to-union ratio, i.e., the ratio of the intersection and union of the target prediction box and the real box.
2. Slave meterComputer memory initialization of a two-dimensional array
Figure DEST_PATH_IMAGE007
(i =1,2, 3...., N.j =1, 2), the first column stores the prediction score of the target classification, and the second column is used to mark whether the detection box is a TP.
3. Initializing precision, recall result matrices from computer memory
Figure 256131DEST_PATH_IMAGE008
(i =1,2, 3.... N.j =1, 2), a first column stores recall values and a second column stores corresponding precision values. precision = TP/(TP + FP); recall = TP/(TP + FN).
4. Calculating IOU for each detection frame and ground Truth of each image, wherein TP is greater than threshold value, FP is greater than threshold value, and the value is assigned to the array
Figure 27778DEST_PATH_IMAGE007
5. For two-dimensional array
Figure 40733DEST_PATH_IMAGE007
The first column of prediction scores is ordered from large to small (the second column also follows the first column ordering).
6. For two-dimensional array
Figure 923238DEST_PATH_IMAGE007
Calculating precision and recall ratio of current time line by line to obtain a group of
Figure DEST_PATH_IMAGE009
) Is assigned a value to
Figure 406215DEST_PATH_IMAGE008
Row i of (2).
7. Calculating the AP (Average Precision) under the current category, specifically by
Figure 348764DEST_PATH_IMAGE008
DrawingPR curve, then using interpolation method (0, 0.1, 0.2.. 1) to make 11 interpolation points or all points on the curve undergo the process of interpolation to obtain the area under the curve, namely
Figure 849015DEST_PATH_IMAGE010
The area under the P-R curve can be used to evaluate the capability of the model to which the curve corresponds, i.e., the larger the area the better the model performance.
The present invention is not limited to the above-described embodiments, which are merely preferred embodiments of the present invention, and the present invention is not limited thereto, and any modification, equivalent replacement, and improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (5)

1. A pedestrian detection method based on deep learning is characterized in that: the method comprises the following steps:
s1: arranging a camera at a scene to be analyzed, acquiring a video of the current analysis scene, marking a surrounding frame of a pedestrian in the video, and storing a pedestrian marking data file as a training data set for deep learning;
s2: calculating a corresponding pedestrian pixel width matrix W and a pedestrian pixel height matrix H according to the pixel width and height of the pedestrian at each position in the image in the training data set;
s3: calculating the Scale and the aspect Ratio of an anchor frame of each region position in the image by using the pedestrian pixel width matrix W and the pedestrian pixel height matrix H;
s4: performing deep learning FasterRcnn model training on a current scene image data set to obtain weight parameters of a trained FasterRcnn model;
s5: deploying a pedestrian detection system for the current scene, acquiring the position coordinate and width and height of each pedestrian, and counting the total number of people or the local density of the current scene by adopting a trained Faster Rcnn model.
2. The pedestrian detection method based on deep learning of claim 1, wherein: the step S2 includes the steps of:
s21: reading a locally stored pedestrian marking data file;
s22: creating a pedestrian pixel width matrix W and a pedestrian pixel height matrix H, wherein the dimensions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H respectively correspond to the pixel width and height of the camera image;
s23: if a certain pixel point in the training data set image is the central point of a pedestrian bounding box, writing the pixel width value and the pixel height value of the bounding box into corresponding positions of a pedestrian pixel width matrix W and a pedestrian pixel height matrix H respectively; if the corresponding positions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H have numerical values, calculating the addition and average value of the written value and the stored value, and then writing the sum and average value into the corresponding positions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H;
s24: and checking whether 0 value points exist in the pedestrian pixel width matrix W and the pedestrian pixel height matrix H obtained in the steps, if so, performing adjacent pixel distance interpolation to calculate the pixel value of the current position, and writing the pixel value into the corresponding positions of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H.
3. The pedestrian detection method based on deep learning of claim 1, wherein: the step S3 includes the steps of:
s31: performing n-n meshing on the pedestrian pixel width matrix W and the pedestrian pixel height matrix H, and traversing the training data set to obtain the pedestrian maximum pixel height H1 and the pedestrian minimum pixel height H2 in the current analysis scene, wherein n = H2/H1, and n is an integer obtained by downward rounding;
s32: calculating the average pixel value in each grid of the pedestrian pixel width matrix W as the Scale of the corresponding training RPN network;
s33: calculating the average value of the pedestrian pixel width matrix W and the pedestrian pixel height matrix H in each grid respectively as
Figure DEST_PATH_IMAGE001
And
Figure 1243DEST_PATH_IMAGE002
then the aspect Ratio within each grid is {1:
Figure DEST_PATH_IMAGE003
/
Figure 843297DEST_PATH_IMAGE002
}。
4. the pedestrian detection method based on deep learning of claim 1, wherein: in the step S4, in the RPN network training process, the Scale and the aspect Ratio of the anchor frame are obtained through the coordinates of the center of the candidate frame generated by the RPN network.
5. The pedestrian detection method based on deep learning of claim 1, wherein: the step S5 includes the steps of:
s51: automatically acquiring a real-time video stream of the analysis scene by using the camera sdk, and decoding to obtain an image of the analysis scene;
s52: loading the weight parameters of the trained FasterRcnn model, and inputting the decoded image into the trained FasterRcnn model to obtain a convolution characteristic diagram;
s53: and performing RPN network reasoning on the convolution characteristic diagram to obtain a target candidate region, and performing Fast RCNN classification and region frame regression to obtain a target classification score and pedestrian width and height.
CN202010586392.XA 2020-06-24 2020-06-24 Pedestrian detection method based on deep learning Active CN111738164B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010586392.XA CN111738164B (en) 2020-06-24 2020-06-24 Pedestrian detection method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010586392.XA CN111738164B (en) 2020-06-24 2020-06-24 Pedestrian detection method based on deep learning

Publications (2)

Publication Number Publication Date
CN111738164A true CN111738164A (en) 2020-10-02
CN111738164B CN111738164B (en) 2021-02-26

Family

ID=72650864

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010586392.XA Active CN111738164B (en) 2020-06-24 2020-06-24 Pedestrian detection method based on deep learning

Country Status (1)

Country Link
CN (1) CN111738164B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112598738A (en) * 2020-12-25 2021-04-02 南京大学 Figure positioning method based on deep learning
CN112633168A (en) * 2020-12-23 2021-04-09 长沙中联重科环境产业有限公司 Garbage truck and method and device for identifying barrel turning action of garbage truck
CN113361370A (en) * 2021-06-02 2021-09-07 南京工业大学 Abnormal behavior detection method based on deep learning

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106874894A (en) * 2017-03-28 2017-06-20 电子科技大学 A kind of human body target detection method based on the full convolutional neural networks in region
CN107437099A (en) * 2017-08-03 2017-12-05 哈尔滨工业大学 A kind of specific dress ornament image recognition and detection method based on machine learning
US20180096595A1 (en) * 2016-10-04 2018-04-05 Street Simplified, LLC Traffic Control Systems and Methods
CN108021848A (en) * 2016-11-03 2018-05-11 浙江宇视科技有限公司 Passenger flow volume statistical method and device
CN108830152A (en) * 2018-05-07 2018-11-16 北京红云智胜科技有限公司 The pedestrian detection method and system that deep learning network and manual features are combined
CN109117806A (en) * 2018-08-22 2019-01-01 歌尔科技有限公司 A kind of gesture identification method and device
US20190043178A1 (en) * 2018-07-10 2019-02-07 Intel Corporation Low-light imaging using trained convolutional neural networks
CN109902806A (en) * 2019-02-26 2019-06-18 清华大学 Method is determined based on the noise image object boundary frame of convolutional neural networks
CN109977812A (en) * 2019-03-12 2019-07-05 南京邮电大学 A kind of Vehicular video object detection method based on deep learning
CN110136098A (en) * 2019-04-15 2019-08-16 江苏大学 A kind of order of cables detection method based on deep learning
CN110263712A (en) * 2019-06-20 2019-09-20 江南大学 A kind of coarse-fine pedestrian detection method based on region candidate

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180096595A1 (en) * 2016-10-04 2018-04-05 Street Simplified, LLC Traffic Control Systems and Methods
CN108021848A (en) * 2016-11-03 2018-05-11 浙江宇视科技有限公司 Passenger flow volume statistical method and device
CN106874894A (en) * 2017-03-28 2017-06-20 电子科技大学 A kind of human body target detection method based on the full convolutional neural networks in region
CN107437099A (en) * 2017-08-03 2017-12-05 哈尔滨工业大学 A kind of specific dress ornament image recognition and detection method based on machine learning
CN108830152A (en) * 2018-05-07 2018-11-16 北京红云智胜科技有限公司 The pedestrian detection method and system that deep learning network and manual features are combined
US20190043178A1 (en) * 2018-07-10 2019-02-07 Intel Corporation Low-light imaging using trained convolutional neural networks
CN109117806A (en) * 2018-08-22 2019-01-01 歌尔科技有限公司 A kind of gesture identification method and device
CN109902806A (en) * 2019-02-26 2019-06-18 清华大学 Method is determined based on the noise image object boundary frame of convolutional neural networks
CN109977812A (en) * 2019-03-12 2019-07-05 南京邮电大学 A kind of Vehicular video object detection method based on deep learning
CN110136098A (en) * 2019-04-15 2019-08-16 江苏大学 A kind of order of cables detection method based on deep learning
CN110263712A (en) * 2019-06-20 2019-09-20 江南大学 A kind of coarse-fine pedestrian detection method based on region candidate

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112633168A (en) * 2020-12-23 2021-04-09 长沙中联重科环境产业有限公司 Garbage truck and method and device for identifying barrel turning action of garbage truck
CN112633168B (en) * 2020-12-23 2023-10-31 长沙中联重科环境产业有限公司 Garbage truck and method and device for identifying garbage can overturning action of garbage truck
CN112598738A (en) * 2020-12-25 2021-04-02 南京大学 Figure positioning method based on deep learning
CN112598738B (en) * 2020-12-25 2024-03-19 南京大学 Character positioning method based on deep learning
CN113361370A (en) * 2021-06-02 2021-09-07 南京工业大学 Abnormal behavior detection method based on deep learning
CN113361370B (en) * 2021-06-02 2023-06-23 南京工业大学 Abnormal behavior detection method based on deep learning

Also Published As

Publication number Publication date
CN111738164B (en) 2021-02-26

Similar Documents

Publication Publication Date Title
CN113963445B (en) Pedestrian falling action recognition method and equipment based on gesture estimation
CN111738164B (en) Pedestrian detection method based on deep learning
Li et al. Adaptive deep convolutional neural networks for scene-specific object detection
CN107633226B (en) Human body motion tracking feature processing method
CN110135354B (en) Change detection method based on live-action three-dimensional model
CN107784288B (en) Iterative positioning type face detection method based on deep neural network
CN101470809B (en) Moving object detection method based on expansion mixed gauss model
CN114648665B (en) Weak supervision target detection method and system
CN111814597A (en) Urban function partitioning method coupling multi-label classification network and YOLO
CN113657414B (en) Object identification method
CN114758288A (en) Power distribution network engineering safety control detection method and device
CN111275010A (en) Pedestrian re-identification method based on computer vision
CN113033516A (en) Object identification statistical method and device, electronic equipment and storage medium
CN112419202A (en) Wild animal image automatic identification system based on big data and deep learning
CN117437382B (en) Updating method and system for data center component
CN111091101A (en) High-precision pedestrian detection method, system and device based on one-step method
CN114332473A (en) Object detection method, object detection device, computer equipment, storage medium and program product
CN112149665A (en) High-performance multi-scale target detection method based on deep learning
CN112861970A (en) Fine-grained image classification method based on feature fusion
CN114495266A (en) Non-standing posture detection method and device, computer equipment and storage medium
CN116091946A (en) Yolov 5-based unmanned aerial vehicle aerial image target detection method
CN113139540B (en) Backboard detection method and equipment
CN116805415A (en) Cage broiler health status identification method based on lightweight improved YOLOv5
Li A crowd density detection algorithm for tourist attractions based on monitoring video dynamic information analysis
CN117115824A (en) Visual text detection method based on stroke region segmentation strategy

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant