CN111967433A - Action identification method based on self-supervision learning network - Google Patents
Action identification method based on self-supervision learning network Download PDFInfo
- Publication number
- CN111967433A CN111967433A CN202010894661.9A CN202010894661A CN111967433A CN 111967433 A CN111967433 A CN 111967433A CN 202010894661 A CN202010894661 A CN 202010894661A CN 111967433 A CN111967433 A CN 111967433A
- Authority
- CN
- China
- Prior art keywords
- action
- model
- information
- training
- positive
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000009471 action Effects 0.000 title claims abstract description 51
- 238000000034 method Methods 0.000 title claims abstract description 13
- 238000001514 detection method Methods 0.000 claims abstract description 13
- 238000013145 classification model Methods 0.000 claims abstract description 7
- 230000006870 function Effects 0.000 claims description 12
- 238000013528 artificial neural network Methods 0.000 claims description 5
- 230000004913 activation Effects 0.000 claims description 3
- 238000007477 logistic regression Methods 0.000 claims description 3
- 238000000605 extraction Methods 0.000 claims description 2
- 238000003062 neural network model Methods 0.000 claims 5
- 238000002372 labelling Methods 0.000 claims 2
- 210000000988 bone and bone Anatomy 0.000 claims 1
- 238000013136 deep learning model Methods 0.000 abstract 1
- 230000009194 climbing Effects 0.000 description 12
- 239000000523 sample Substances 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 4
- 230000006399 behavior Effects 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000020595 eating behavior Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000003238 somatosensory effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Probability & Statistics with Applications (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Image Analysis (AREA)
Abstract
本发明提供了一种基于自监督学习网络的动作识别办法,第一步使用OpenPose对视频流里的人体骨架信息进行提取,并将人体骨架信息制作成正负样本数据集,使用数据集训练动作分类模型ResNet‑56,初步训练好的动作分类模型用于初步判断输入动作和进一步的自监督训练;在训练动作分类器的同时利用深度学习模型YOLOv4检测视频流中的物体信息,将检测到的物体信息打上正负标签传入动作分类模型ResNet‑56中进行自监督学习训练,经过自监督后的动作分类模型有更高的检测精度和可靠性。
The invention provides an action recognition method based on a self-supervised learning network. The first step is to use OpenPose to extract human skeleton information in a video stream, make the human skeleton information into a positive and negative sample data set, and use the data set to train actions The classification model ResNet‑56, the preliminarily trained action classification model is used to preliminarily judge the input action and further self-supervised training; while training the action classifier, the deep learning model YOLOv4 is used to detect the object information in the video stream, and the detected The object information is labeled with positive and negative labels and sent to the action classification model ResNet‑56 for self-supervised learning training. The action classification model after self-supervision has higher detection accuracy and reliability.
Description
技术领域technical field
本发明属于计算机智能学习领域,更具体地说是一种结合OpenPose和YOLOv4的自监督人体动作识别办法。The invention belongs to the field of computer intelligent learning, and more specifically relates to a self-supervised human action recognition method combining OpenPose and YOLOv4.
背景技术Background technique
计算机视觉技术是人工智能的分支方向,主要通过视觉图像信息的处理,达到识别和理解图像内容的目的。智能视频监控借助于计算机强大的计算能力。依靠计算机视觉技术,从监控图像或视频中提取关键信息并快速分析,并将分析结果反馈给监控系统进行处理,从而达到智能识别、理解和处理监控视频的目的。Computer vision technology is a branch of artificial intelligence, mainly through the processing of visual image information, to achieve the purpose of identifying and understanding image content. Intelligent video surveillance relies on the powerful computing power of computers. Relying on computer vision technology, key information is extracted from surveillance images or videos and quickly analyzed, and the analysis results are fed back to the surveillance system for processing, so as to achieve the purpose of intelligently identifying, understanding and processing surveillance videos.
人体动作识别里人体关键点的分析是最为重要的之一,也被统称为人体姿态估计。目前,有很多种技术可以实现对人体关键点的信息进行读取,例如深度体感相机,如Kinect等,再人们穿上设备之后可以对人体d额关键点进行实时的信息采集。不足之处是一次可采集的人数太少,并且价格较为昂贵。相比之下采用OpenPose来对人体骨骼关键点进行信息采集则方便的多,一次性也可采集多人,并且不需要太贵的价格。The analysis of human key points in human action recognition is one of the most important, and is also collectively referred to as human pose estimation. At present, there are many technologies that can read the information of the key points of the human body, such as depth somatosensory cameras, such as Kinect, etc. After people put on the device, they can collect real-time information on the key points of the human body. The disadvantage is that the number of people that can be collected at one time is too small, and the price is relatively expensive. In contrast, it is much more convenient to use OpenPose to collect information on key points of human skeleton, and it can collect multiple people at one time, and it does not require too expensive price.
OpenPose是目前主流的人体姿态估计采用的模型算法,大多数的动作都可以用OpenPose采集人体骨骼关键点信息并加以训练后可以识别出视频流中的动作,但是,相对于Kinect的高精准性,OpenPose在检测精度上略有不足;因为大多数的模型都是基于单调的OpenPose来进行动作识别,例如,一种基于OpenPose的吃东西行为识别方法CN201911150648.6,基于OpenPose的击剑动作获取方法及计算机存储介质CN201810338998.4等,上述两篇专利都是基于只基于OpenPose来提取信息就完成了动作识别的模型训练OpenPose is the current mainstream model algorithm for human pose estimation. Most actions can use OpenPose to collect key point information of human skeletons and train them to recognize actions in video streams. However, compared to Kinect's high accuracy, OpenPose is slightly insufficient in detection accuracy; because most models are based on monotonic OpenPose for action recognition, for example, an OpenPose-based eating behavior recognition method CN201911150648.6, OpenPose-based fencing action acquisition method and computer The storage medium CN201810338998.4, etc., the above two patents are based on the model training of action recognition only based on OpenPose to extract information
为了提高OpenPose在实际运用中的检测准确性,本专利提供了一种基于自监督学习网络的动作识别办法,在传统的单调OpenPose动作识别的基础上,加入了YOLOv4目标检测技术,将检测到的物品信息分类为阳性、阴性样本后再自监督模型训练,以此提高动作识别的准确度。In order to improve the detection accuracy of OpenPose in practical applications, this patent provides an action recognition method based on a self-supervised learning network. On the basis of the traditional monotonous OpenPose action recognition, YOLOv4 target detection technology is added to detect the detected objects. Item information is classified into positive and negative samples and then self-supervised model training is performed to improve the accuracy of action recognition.
发明内容SUMMARY OF THE INVENTION
本发明基于目前单调的OpenPose提取人体骨架信息再用于识别的精确度不高,提出了一种基于自监督学习网络的动作识别办法。The present invention proposes an action recognition method based on a self-supervised learning network based on the low accuracy of the current monotonous OpenPose extracting human skeleton information and then using it for recognition.
本发明具体内容如下:The specific content of the present invention is as follows:
步骤S1:收集所人体骨骼关节点特征信息:OpenPose通过前馈网络预测出图片中人体部位置信度S,同时预测出部位的亲和力向量场L(人体骨骼各个关节的关系),集合S=(S1,S2,SJ)J表示每个骨骼关节点有J个身体部位置信图。集合L(L1,L2,LC)每个肢体有C个部位亲和力向量场。得到集合J和L后使用Greedy algorithm算法将人体骨骼关节点信息找出。将提取到的骨架特征信息制作成所需识别动作的正负样本数据集。如图1所示。Step S1: Collect the feature information of all human skeleton joint points: OpenPose predicts the position reliability S of the human body in the picture through the feedforward network, and at the same time predicts the affinity vector field L of the part (the relationship between each joint of the human skeleton), the set S = ( S1, S2, SJ) J means that each skeleton joint point has J body part position information maps. The set L(L1, L2, LC) has C site affinity vector fields for each limb. After the sets J and L are obtained, the Greedy algorithm is used to find out the joint point information of the human skeleton. The extracted skeleton feature information is made into a positive and negative sample data set of the desired action to be recognized. As shown in Figure 1.
步骤S2:将步骤S1.1中制作的正负样本数据集作为输入,训练基于ResNet-56动作分类器模型。在ResNet-56模型中,使用Softmax损失进行分类,目标函数如下: Step S2: Use the positive and negative sample data set made in Step S1.1 as input, and train an action classifier model based on ResNet-56. In the ResNet-56 model, the Softmax loss is used for classification, and the objective function is as follows:
N为样本个数,λ是回归损失权重,i是批次中提议窗口的索引,P是预测概率。N is the number of samples, λ is the regression loss weight, i is the index of the proposed window in the batch, and P is the predicted probability.
神经络的输入是x,期望输出是H(x),残差网络F(x)=H(x)-x,训练ResNet-56模型直至F(x)接近0则模型训练结束。将训练数据集传入步骤S2.1所训练好的模型中,得到视频流训练集的动作识别结果。The input of the neural network is x, the expected output is H(x), the residual network F(x)=H(x)-x, and the ResNet-56 model is trained until F(x) is close to 0. The model training ends. The training data set is passed into the model trained in step S2.1, and the action recognition result of the video stream training set is obtained.
步骤S3.1:物体数据集的建立:对需检测动作中所包含除人体以外的物体数据集进行收集分类。如:将需识别的动作设为A,则和动作A相关的物品设为B1、B2……依次排列,可以从多种不同视频或图片中收集物品B1、B2的信息,并将其制作成标准VOC数据集。Step S3.1: Establishment of Object Data Set: Collect and classify the object data set other than human body included in the action to be detected. For example: set the action to be recognized as A, then the items related to action A are set as B1, B2... Arranged in order, the information of items B1 and B2 can be collected from a variety of different videos or pictures, and made into Standard VOC dataset.
步骤S3.2:将上述步骤S3.1中制作好的物品VOC数据集输入到标准的YOLOv4模型中进行训练。YOLOv4模型中运用卷积的级联结构,网络的输入层设计为448*448。Step S3.2: Input the item VOC data set prepared in the above step S3.1 into the standard YOLOv4 model for training. The convolution cascade structure is used in the YOLOv4 model, and the input layer of the network is designed to be 448*448.
逻辑回归的代价函数为: The cost function of logistic regression is:
其中hθ是sigmoid函数,作为激活函数在网络中。当物品检测的准确率达到95%以上时,模型训练完毕,随即可以用于检测视频流中的物品和行人信息并记录。where h θ is the sigmoid function as the activation function in the network. When the accuracy rate of item detection reaches more than 95%, the model is trained and can be used to detect and record the information of items and pedestrians in the video stream.
步骤S4.1:将步骤S3中识别出的物品信息标签划分为正负标签,若物品信息和动作信息有相似则视为阳性样本,若不同则视为阴性样本,对于这些物品信息,对比方法可以根据公式score(f(x),f(x+))>>score(f(x),f(x-))来衡量两个功能相似之间的相似指标。Step S4.1: Divide the item information labels identified in step S3 into positive and negative labels. If the item information and action information are similar, they are regarded as positive samples, and if they are different, they are regarded as negative samples. The similarity index between two functional similarities can be measured according to the formula score(f(x), f(x + )) >> score(f(x), f(x - )).
构造softmax分类器来正确的对正负样本进行分类:Construct a softmax classifier to correctly classify positive and negative samples:
步骤S4.2:将物品信息分为正负样本后,正负样本将会被传入步骤S3中的ResNet-56中进行自监督训练:当阳性样本信息较多的时候则最终识别的动作的置信度会增加,反之,若阴性样本数量较多的时候则预测的改动作置信度会降低。经过不断的输入由物品信息制作成的阳性阴性标签训练模型,动作分类器模型的识别率会逐步上升,直至达到理想水平。Step S4.2: After dividing the item information into positive and negative samples, the positive and negative samples will be passed to ResNet-56 in step S3 for self-supervised training: when there is a lot of positive sample information, the final identified action The confidence will increase, on the contrary, if the number of negative samples is large, the confidence of the predicted change will decrease. After continuous input of positive and negative labels made from item information to train the model, the recognition rate of the action classifier model will gradually increase until it reaches an ideal level.
附图说明Description of drawings
图1为OpenPose提取人体骨架信息流程;Figure 1 shows the process of extracting human skeleton information from OpenPose;
图2为人体周围物品检测;Figure 2 shows the detection of objects around the human body;
图3为自监督训练动作分类模型图Figure 3 is a diagram of the self-supervised training action classification model
具体实施方式Detailed ways
为了更清楚地说明本发明实施例的技术方案,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,应当理解,所描述的实施例仅仅是本发明的一部分实施例,而不是全部的实施例,因此不应被看作是对保护范围的限定。基于本发明中的实施例,本领域普通技术工作人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the following will clearly and completely describe the technical solutions in the embodiments of the present invention with reference to the accompanying drawings in the embodiments of the present invention. It should be understood that the described embodiments are only Some, but not all, embodiments of the present invention should therefore not be construed as limiting the scope of protection. Based on the embodiments of the present invention, all other embodiments obtained by those of ordinary skill in the art without creative work shall fall within the protection scope of the present invention.
实施例1:(以人在攀爬为例)Example 1: (take a person climbing as an example)
步骤S1:收集人在攀爬围墙、攀爬梯子等和攀爬相关的人体骨骼关节点特征信息:OpenPose通过前馈网络预测出攀爬人体的关键节点部位置信度S,同时预测出部位的亲和力向量场L(人体骨骼各个关节的关系),集合S=(S1,S2,SJ)J表示每个骨骼关节点有J个身体部位置信图。集合L(L1,L2,LC)每个肢体有C个部位亲和力向量场。得到集合J和L后使用Greedy algorithm算法将人体骨骼关节点信息找出。将提取到的攀爬骨架特征信息制作成所需识别动作的正负样本数据集,正样本为人有攀爬动作,负样本为人没有攀爬动作。骨骼关键点预测如图1所示。Step S1: Collect the feature information of human skeleton joint points related to climbing when people are climbing walls, climbing ladders, etc.: OpenPose predicts the position reliability S of the key nodes of the climbing human body through a feedforward network, and at the same time predicts the position of the body. The affinity vector field L (the relationship of each joint of the human skeleton), the set S=(S1, S2, SJ)J indicates that each skeleton joint point has J body part position information maps. The set L(L1, L2, LC) has C site affinity vector fields for each limb. After the sets J and L are obtained, the Greedy algorithm is used to find out the joint point information of the human skeleton. The extracted climbing skeleton feature information is made into a positive and negative sample data set for the required recognition action. Skeletal keypoint predictions are shown in Figure 1.
步骤S2:将步骤S1中制作的攀爬正负样本数据集作为输入,训练基于ResNet-56动作分类器模型。在ResNet-56模型中如图2所示,使用Softmax损失进行分类,目标函数如下:Step S2: Use the climbing positive and negative sample data set made in Step S1 as input, and train a ResNet-56-based action classifier model. In the ResNet-56 model shown in Figure 2, the Softmax loss is used for classification, and the objective function is as follows:
N为样本个数,λ是回归损失权重,i是批次中提议窗口的索引,P是预测概率。N is the number of samples, λ is the regression loss weight, i is the index of the proposed window in the batch, and P is the predicted probability.
神经络的输入是x,期望输出是H(x),残差网络F(x)=H(x)-x,训练ResNet-56模型直至F(x)接近0则模型训练结束。此时的动作模型可以用于识别视频流中的攀爬行为,但是精度还有待提高。The input of the neural network is x, the expected output is H(x), the residual network F(x)=H(x)-x, and the ResNet-56 model is trained until F(x) is close to 0. The model training ends. The action model at this time can be used to identify the climbing behavior in the video stream, but the accuracy needs to be improved.
步骤S3.1:在本专利中,YOLOv4的物体检测是官方权重和自建数据集权重合并使用。所以第一步是攀爬相关物体数据集的建立:攀爬是需要测测的物体为围墙、门、梯子、电线杆等,将这些物品的图片进行分类收集,并将其制作成标准VOC数据集。Step S3.1: In this patent, the object detection of YOLOv4 is a combination of official weights and self-built dataset weights. Therefore, the first step is to establish a dataset of climbing-related objects: climbing is to measure the objects such as walls, doors, ladders, telephone poles, etc., classify and collect the pictures of these objects, and make them into standard VOC data set.
步骤S3.2:将上述步骤S3.1中制作好的物品VOC数据集输入到标准的YOLOv4模型中进行训练如图3所示。Step S3.2: Input the item VOC data set prepared in the above step S3.1 into the standard YOLOv4 model for training, as shown in Figure 3.
在本专利使用的模型YOLOv4模型中,拥有多个卷积层和一个全连接层,大量运用了卷积的级联结构,网络的输入层设计为448*448。卷积级联结构用于图像的特征提取,全连接层用于预测类概率和边框。模型中逻辑回归的代价函数为:In the model YOLOv4 model used in this patent, there are multiple convolution layers and one fully connected layer, and the convolution cascade structure is widely used. The input layer of the network is designed to be 448*448. The convolution cascade structure is used for image feature extraction, and the fully connected layer is used to predict class probabilities and bounding boxes. The cost function of logistic regression in the model is:
其中hθ是sigmoid函数,作为激活函数在网络中。当物品检测的准确率达到95%以上时,模型训练完毕,随即可以用于检测视频流中的物品和行人信息并记录下来。where h θ is the sigmoid function as the activation function in the network. When the accuracy rate of item detection reaches more than 95%, the model is trained and can be used to detect and record the information of items and pedestrians in the video stream.
步骤S4.1:将步骤S3中识别出的物品信息标签划分为正负标签,和攀爬相关的物品如门、梯子、电线杆等视为阳性样本,将手机、耳机、食物等视为阴性样本,其余识别到的物体按照构造softmax分类器来正确的对正负样本进行分类:Step S4.1: Divide the item information tags identified in step S3 into positive and negative tags, and items related to climbing such as doors, ladders, telephone poles, etc. are regarded as positive samples, and mobile phones, earphones, food, etc. are regarded as negative samples Samples, and the rest of the recognized objects are classified according to the construction of a softmax classifier to correctly classify positive and negative samples:
分母项包括一个正数和n-1个阴性样本;The denominator term includes a positive number and n-1 negative samples;
步骤S4.2:将物品信息分为阳性、阴性样本后,这两种物品信息会传入ResNet-56动作分类器模型中,在经过自监督的学习下将动作识别的精度提高。Step S4.2: After the item information is divided into positive and negative samples, the two types of item information will be passed into the ResNet-56 action classifier model, and the accuracy of action recognition will be improved under self-supervised learning.
模型训练好后,就可将训练模型用于检测视频流中是否出现攀爬行为,精度相对于没有经过自监督的模型会有较大的提升。After the model is trained, the trained model can be used to detect whether there is climbing behavior in the video stream, and the accuracy will be greatly improved compared to the model without self-supervision.
Claims (4)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010894661.9A CN111967433A (en) | 2020-08-31 | 2020-08-31 | Action identification method based on self-supervision learning network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010894661.9A CN111967433A (en) | 2020-08-31 | 2020-08-31 | Action identification method based on self-supervision learning network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111967433A true CN111967433A (en) | 2020-11-20 |
Family
ID=73400782
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010894661.9A Pending CN111967433A (en) | 2020-08-31 | 2020-08-31 | Action identification method based on self-supervision learning network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111967433A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112560649A (en) * | 2020-12-09 | 2021-03-26 | 广州云从鼎望科技有限公司 | Behavior action detection method, system, equipment and medium |
CN112560618A (en) * | 2020-12-06 | 2021-03-26 | 复旦大学 | Behavior classification method based on skeleton and video feature fusion |
CN112668492A (en) * | 2020-12-30 | 2021-04-16 | 中山大学 | Behavior identification method for self-supervised learning and skeletal information |
CN113191228A (en) * | 2021-04-20 | 2021-07-30 | 上海东普信息科技有限公司 | Express item casting identification method, device, equipment and storage medium |
CN113935407A (en) * | 2021-09-29 | 2022-01-14 | 光大科技有限公司 | Abnormal behavior recognition model determining method and device |
CN114758294A (en) * | 2022-04-08 | 2022-07-15 | 上海东普信息科技有限公司 | Method, device, equipment and storage medium for monitoring dangerous behaviors of distribution center |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110135319A (en) * | 2019-05-09 | 2019-08-16 | 广州大学 | A kind of abnormal behavior detection method and system |
CN110852303A (en) * | 2019-11-21 | 2020-02-28 | 中科智云科技有限公司 | Eating behavior identification method based on OpenPose |
CN111178251A (en) * | 2019-12-27 | 2020-05-19 | 汇纳科技股份有限公司 | Pedestrian attribute identification method and system, storage medium and terminal |
-
2020
- 2020-08-31 CN CN202010894661.9A patent/CN111967433A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110135319A (en) * | 2019-05-09 | 2019-08-16 | 广州大学 | A kind of abnormal behavior detection method and system |
CN110852303A (en) * | 2019-11-21 | 2020-02-28 | 中科智云科技有限公司 | Eating behavior identification method based on OpenPose |
CN111178251A (en) * | 2019-12-27 | 2020-05-19 | 汇纳科技股份有限公司 | Pedestrian attribute identification method and system, storage medium and terminal |
Non-Patent Citations (1)
Title |
---|
李宾皑 等: "目标检测、人体姿态估计算法叠加的监控视频分析方法", 《电子技术与软件工程》 * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112560618A (en) * | 2020-12-06 | 2021-03-26 | 复旦大学 | Behavior classification method based on skeleton and video feature fusion |
CN112560649A (en) * | 2020-12-09 | 2021-03-26 | 广州云从鼎望科技有限公司 | Behavior action detection method, system, equipment and medium |
CN112668492A (en) * | 2020-12-30 | 2021-04-16 | 中山大学 | Behavior identification method for self-supervised learning and skeletal information |
CN112668492B (en) * | 2020-12-30 | 2023-06-20 | 中山大学 | A behavior recognition method based on self-supervised learning and skeletal information |
CN113191228A (en) * | 2021-04-20 | 2021-07-30 | 上海东普信息科技有限公司 | Express item casting identification method, device, equipment and storage medium |
CN113935407A (en) * | 2021-09-29 | 2022-01-14 | 光大科技有限公司 | Abnormal behavior recognition model determining method and device |
CN114758294A (en) * | 2022-04-08 | 2022-07-15 | 上海东普信息科技有限公司 | Method, device, equipment and storage medium for monitoring dangerous behaviors of distribution center |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111967433A (en) | Action identification method based on self-supervision learning network | |
CN111783576B (en) | Pedestrian re-identification method based on improved YOLOv3 network and feature fusion | |
CN109446923B (en) | Deeply supervised convolutional neural network behavior recognition method based on training feature fusion | |
CN111259786B (en) | Pedestrian re-identification method based on synchronous enhancement of appearance and motion information of video | |
Özyer et al. | Human action recognition approaches with video datasets—A survey | |
Misra et al. | Shuffle and learn: unsupervised learning using temporal order verification | |
CN106203318B (en) | Pedestrian recognition method based on multi-level deep feature fusion in camera network | |
CN111898736B (en) | An Efficient Pedestrian Re-identification Method Based on Attribute Awareness | |
CN108765394B (en) | Target identification method based on quality evaluation | |
WO2020182121A1 (en) | Expression recognition method and related device | |
CN110991340B (en) | A Method of Human Motion Analysis Based on Image Compression | |
CN110070067A (en) | The training method of video classification methods and its model, device and electronic equipment | |
CN112784763A (en) | Expression recognition method and system based on local and overall feature adaptive fusion | |
CN110575663B (en) | Physical education auxiliary training method based on artificial intelligence | |
CN113361549B (en) | A model updating method and related device | |
CN106909938B (en) | Perspective-independent behavior recognition method based on deep learning network | |
CN110807434A (en) | Pedestrian re-identification system and method based on combination of human body analysis and coarse and fine particle sizes | |
CN112001347A (en) | Motion recognition method based on human skeleton shape and detection target | |
CN110321862B (en) | A Pedestrian Re-Identification Method Based on Compact Ternary Loss | |
CN109446897B (en) | Scene recognition method and device based on image context information | |
CN110674483A (en) | An identity recognition method based on multimodal information | |
CN107992854A (en) | Forest Ecology man-machine interaction method based on machine vision | |
CN101354787A (en) | A method for extracting features of target motion trajectory in intelligent visual surveillance retrieval | |
CN114419391A (en) | Target image recognition method and device, electronic device and readable storage medium | |
CN111723600A (en) | A feature descriptor for person re-identification based on multi-task learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20201120 |