CN113569627A - Human body posture prediction model training method, human body posture prediction method and device - Google Patents
Human body posture prediction model training method, human body posture prediction method and device Download PDFInfo
- Publication number
- CN113569627A CN113569627A CN202110658308.5A CN202110658308A CN113569627A CN 113569627 A CN113569627 A CN 113569627A CN 202110658308 A CN202110658308 A CN 202110658308A CN 113569627 A CN113569627 A CN 113569627A
- Authority
- CN
- China
- Prior art keywords
- human body
- body posture
- posture prediction
- generator
- loss value
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000012549 training Methods 0.000 title claims abstract description 136
- 238000000034 method Methods 0.000 title claims abstract description 64
- 238000002372 labelling Methods 0.000 claims abstract description 27
- 238000001514 detection method Methods 0.000 claims description 28
- 238000004590 computer program Methods 0.000 claims description 8
- 238000003860 storage Methods 0.000 claims description 6
- 238000005457 optimization Methods 0.000 claims description 5
- 230000036544 posture Effects 0.000 description 193
- 238000012545 processing Methods 0.000 description 10
- 238000005516 engineering process Methods 0.000 description 8
- 238000013473 artificial intelligence Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 6
- 230000009471 action Effects 0.000 description 4
- 238000004891 communication Methods 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 4
- 210000000707 wrist Anatomy 0.000 description 4
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000003796 beauty Effects 0.000 description 2
- 238000004422 calculation algorithm Methods 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000010801 machine learning Methods 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000012795 verification Methods 0.000 description 2
- 101000742346 Crotalus durissus collilineatus Zinc metalloproteinase/disintegrin Proteins 0.000 description 1
- 101000872559 Hediste diversicolor Hemerythrin Proteins 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 238000009529 body temperature measurement Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Medical Informatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Image Analysis (AREA)
Abstract
The application provides a human body posture prediction model training method, a human body posture prediction method and a human body posture prediction device. The method comprises the following steps: acquiring a labeled training set and an unlabeled training set, wherein the labeled training set comprises a plurality of first human body images containing labeled data, and the labeled data is used for representing real posture information in the first human body images; inputting the first human body image into a generator in a human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeling data and the first human body posture prediction result; inputting the second human body image into a generator to obtain a corresponding second human body posture prediction result; calculating a second loss value corresponding to the discriminator in the human body posture prediction model according to the first human body image, the annotation data, the second human body image and the second human body posture prediction result; and optimizing the generator and the discriminator according to the first loss value and the second loss value to obtain a human body posture prediction model.
Description
Technical Field
The application relates to the technical field of artificial intelligence, in particular to a human posture prediction model training method, a human posture prediction method and a human posture prediction device.
Background
The human body posture estimation means that the human body posture and key points in the image are predicted. In essence, human pose estimation abstracts the position of various parts of the human body in an image into a set of structured coordinates. The human body posture estimation technology has important application in the fields of human-computer interaction, image retrieval, abnormity detection, action prediction and the like.
Most of the existing human body posture estimation methods rely on a large amount of labeled data. The data size of the labeled data has great influence on the final effect of the human body posture estimation. Although the data volume of the body posture estimation data set is greatly improved compared with the prior art, it is still very difficult to construct a large body posture estimation data set. For example, the MPII human pose estimation dataset contains 2 million 5 thousand images, and around 4 million human poses, which is much less than the million magnitude dataset required for image classification and image detection. This is because the labeling of the human pose estimation data set is more demanding, more elaborate, and more complex, and therefore requires a significant expenditure of labor and time.
Deep learning has gained a lot of attention since birth and has also gained application in the field of human posture estimation. In contrast to deep learning, supervised training using a large amount of training data has become the mainstream of human posture estimation technology. However, relying too much on labeled training data has hindered the advancement of human body posture estimation techniques to some extent. The technologies such as internet and the like are rapidly growing and bring a large amount of data, and it is not imaginable to use manpower to label such a large amount of data. Therefore, how to estimate the human body posture by using massive non-labeled data is also an urgent problem to be solved in the prior art.
Disclosure of Invention
An object of the embodiments of the present application is to provide a human body posture prediction model training method, a human body posture prediction method, and a device, which can combine a plurality of unlabeled second human body images to perform countermeasure training, thereby improving accuracy and reducing the amount of demand for labeled first human body images.
In a first aspect, an embodiment of the present application provides a human body posture prediction model training method, including: acquiring a labeled training set and an unlabeled training set, wherein the labeled training set comprises a plurality of first human body images containing labeled data, and the labeled data is used for representing real posture information in the first human body images; the unmarked training set comprises a plurality of second human body images which do not contain marking data; inputting the first human body image into a generator in a human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeling data and the first human body posture prediction result; inputting the second human body image into the generator to obtain a corresponding second human body posture prediction result; calculating a second loss value corresponding to a discriminator in the human body posture prediction model according to the first human body image, the labeled data, the second human body image and the second human body posture prediction result; and optimizing the generator and the discriminator according to the first loss value and the second loss value to obtain the human body posture prediction model.
Optionally, in the method for training a human body posture prediction model according to the embodiment of the present application, the inputting the first human body image into a generator in the human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeled data and the first human body posture prediction result includes: inputting the first human body image into the generator to obtain a predicted human body posture heat map of a corresponding multi-channel; wherein the predicted human posture heat map of each channel predicts a human body key point position; generating a corresponding reference human body posture heat map based on the annotation data corresponding to the first human body image; calculating the first loss value from the predicted human pose heat map and the reference human pose heat map.
Optionally, in the method for training a human body posture prediction model according to the embodiment of the present application, the calculating a second loss value corresponding to a discriminator in the human body posture prediction model according to the first human body image, the annotation data, the second human body image, and the second human body posture prediction result includes: inputting the first human body image and the reference human body posture heat map into the discriminator as a true data sequence, and inputting the second human body image and the second human body posture prediction result into the discriminator as a false data sequence to respectively obtain discrimination results output by the discriminator; and calculating the second loss value according to the judgment result.
Optionally, in the human body posture prediction model training method according to the embodiment of the present application, the optimizing the generator and the discriminator according to the first loss value and the second loss value includes: and updating the network parameters of the generator according to the first loss value and the second loss value, and updating the network parameters of the discriminator according to the second loss value.
Optionally, in the human body posture detection model training method according to the embodiment of the present application, the obtaining a labeled training set and an unlabeled training set includes: acquiring an original labeling training set containing labeling data and an original unlabeled training set not containing labeling data; respectively carrying out human body detection on the images in the original labeled training set and the images in the original unlabeled training set by using a pre-trained human body detection model to obtain the first human body images in the labeled training set and the second human body images in the unlabeled training set; wherein the first human body image and the second human body image are both single person images.
In a second aspect, an embodiment of the present application provides a method for predicting a human body posture, including: acquiring a third human body image; and inputting the third human body image into a generator in the human body posture prediction model obtained by adopting the human body posture prediction model training method in the first aspect to obtain a third human body posture prediction result.
In a third aspect, an embodiment of the present application provides a human body posture prediction model training device, including: the system comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for acquiring a labeled training set and an unlabeled training set, the labeled training set comprises a plurality of first human body images containing labeled data, and the labeled data is used for representing real posture information in the first human body images; the unmarked training set comprises a plurality of second human body images which do not contain marking data; the first input module is used for inputting the first human body image into a generator in a human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeling data and the first human body posture prediction result; the second input module is used for inputting the second human body image into the generator to obtain a corresponding second human body posture prediction result; a calculating module, configured to calculate a second loss value corresponding to a discriminator in the human body posture prediction model according to the first human body image, the annotation data, the second human body image, and the second human body posture prediction result; and the optimization module is used for optimizing the generator and the discriminator according to the first loss value and the second loss value to obtain the human body posture prediction model.
Optionally, in the human body posture detection model training device according to the embodiment of the present application, the first input module is specifically configured to: inputting the first human body image into the generator to obtain a predicted human body posture heat map of a corresponding multi-channel; wherein the predicted human posture heat map of each channel predicts a human body key point position; generating a corresponding reference human body posture heat map based on the annotation data corresponding to the first human body image; calculating the first loss value from the predicted human pose heat map and the reference human pose heat map.
Optionally, in the human body posture detection model training apparatus according to the embodiment of the present application, the calculation module is specifically configured to: inputting the first human body image and the reference human body posture heat map into the discriminator as a true data sequence, and inputting the second human body image and the second human body posture prediction result into the discriminator as a false data sequence to respectively obtain discrimination results output by the discriminator; and calculating the second loss value according to the judgment result.
Optionally, in the human body posture detection model training device according to the embodiment of the present application, the optimization module is specifically configured to: and updating the network parameters of the generator according to the first loss value and the second loss value, and updating the network parameters of the discriminator according to the second loss value.
Optionally, in the human body posture detection model training device according to the embodiment of the present application, the first obtaining module is specifically configured to: the acquiring of the labeled training set and the unlabeled training set includes: acquiring an original labeling training set containing labeling data and an original unlabeled training set not containing labeling data; respectively carrying out human body detection on the images in the original labeled training set and the images in the original unlabeled training set by using a pre-trained human body detection model to obtain the first human body images in the labeled training set and the second human body images in the unlabeled training set; wherein the first human body image and the second human body image are both single person images.
In a fourth aspect, an embodiment of the present application provides a human body posture prediction apparatus, including: the second acquisition module is used for acquiring a third human body image; and the prediction module is used for inputting the third human body image into a generator in the human body posture prediction model obtained by adopting the human body posture prediction model training method in the first aspect to obtain a third human body posture prediction result.
In a fifth aspect, embodiments of the present application provide an electronic device, including a processor and a memory, where the memory stores computer-readable instructions that, when executed by the processor, perform the method as in the first aspect or the method as in the second aspect.
In a sixth aspect, embodiments of the present application provide a storage medium having a computer program stored thereon, where the computer program is executed by a processor to perform the method as described in the first aspect or the method as described in the second aspect.
As can be seen from the above, the human body posture prediction model training method, the human body posture prediction method and the device provided by the embodiment of the present application acquire a labeled training set and an unlabeled training set, where the labeled training set includes a plurality of first human body images containing labeled data, and the labeled data is used to represent real posture information in the first human body images; the unmarked training set comprises a plurality of second human body images which do not contain marking data; inputting the first human body image into a generator in a human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeling data and the first human body posture prediction result; inputting the second human body image into the generator to obtain a corresponding second human body posture prediction result; calculating a second loss value corresponding to a discriminator in the human body posture prediction model according to the first human body image, the labeled data, the second human body image and the second human body posture prediction result; and optimizing the generator and the discriminator according to the first loss value and the second loss value to obtain the human body posture prediction model. Therefore, due to the fact that the unmarked training set is applied in the training process, the anti-confrontation training is performed by combining the plurality of unmarked second human body images under the condition that the marked first human body images are insufficient, the accuracy can be improved, and the demand for the marked first human body images is reduced.
Additional features and advantages of the present application will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by the practice of the embodiments of the present application. The objectives and other advantages of the application may be realized and attained by the structure particularly pointed out in the written description and claims hereof as well as the appended drawings.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings that are required to be used in the embodiments of the present application will be briefly described below, it should be understood that the following drawings only illustrate some embodiments of the present application and therefore should not be considered as limiting the scope, and that those skilled in the art can also obtain other related drawings based on the drawings without inventive efforts.
Fig. 1 is a flowchart of a human body posture prediction model training method according to an embodiment of the present disclosure;
fig. 2 is a flowchart of a human body posture prediction method according to an embodiment of the present application;
fig. 3 is a schematic structural diagram of a human body posture prediction model training device according to an embodiment of the present application;
fig. 4 is a schematic structural diagram of a human body posture prediction apparatus according to an embodiment of the present application;
fig. 5 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. The components of the embodiments of the present application, generally described and illustrated in the figures herein, can be arranged and designed in a wide variety of different configurations. Thus, the following detailed description of the embodiments of the present application, presented in the accompanying drawings, is not intended to limit the scope of the claimed application, but is merely representative of selected embodiments of the application. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present application without making any creative effort, shall fall within the protection scope of the present application.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, it need not be further defined and explained in subsequent figures. Meanwhile, in the description of the present application, the terms "first", "second", and the like are used only for distinguishing the description, and are not to be construed as indicating or implying relative importance.
In recent years, technical research based on artificial intelligence, such as computer vision, deep learning, machine learning, image processing, and image recognition, has been actively developed. Artificial Intelligence (AI) is an emerging scientific technology for studying and developing theories, methods, techniques and application systems for simulating and extending human Intelligence. The artificial intelligence subject is a comprehensive subject and relates to various technical categories such as chips, big data, cloud computing, internet of things, distributed storage, deep learning, machine learning and neural networks. Computer vision is used as an important branch of artificial intelligence, particularly a machine is used for identifying the world, and the computer vision technology generally comprises the technologies of face identification, living body detection, fingerprint identification and anti-counterfeiting verification, biological feature identification, face detection, pedestrian detection, target detection, pedestrian identification, image processing, image identification, image semantic understanding, image retrieval, character identification, video processing, video content identification, behavior identification, three-dimensional reconstruction, virtual reality, augmented reality, synchronous positioning and map construction (SLAM), computational photography, robot navigation and positioning and the like. With the research and progress of artificial intelligence technology, the technology is applied to various fields, such as security, city management, traffic management, building management, park management, face passage, face attendance, logistics management, warehouse management, robots, intelligent marketing, computational photography, mobile phone images, cloud services, smart homes, wearable equipment, unmanned driving, automatic driving, smart medical treatment, face payment, face unlocking, fingerprint unlocking, testimony verification, smart screens, smart televisions, cameras, mobile internet, live webcasts, beauty treatment, medical beauty treatment, intelligent temperature measurement and the like.
Referring to fig. 1, fig. 1 is a flowchart of a human body posture prediction model training method according to an embodiment of the present disclosure. The human body posture prediction model training method can comprise the following steps:
s101, acquiring a labeled training set and an unlabeled training set.
S102, inputting the first human body image into a generator in the human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeling data and the first human body posture prediction result.
And S103, inputting the second human body image into the generator to obtain a corresponding second human body posture prediction result.
And S104, calculating a second loss value corresponding to the discriminator in the human body posture prediction model according to the first human body image, the annotation data, the second human body image and the second human body posture prediction result.
And S105, optimizing the generator and the discriminator according to the first loss value and the second loss value to obtain a human body posture prediction model.
Specifically, in the step S101, the annotation training set includes a plurality of first human body images including annotation data. As an embodiment, a plurality of first human body images containing annotation data can be obtained from a public database; as another embodiment, a human body image not containing the annotation data may be obtained from the network, and then a plurality of first human body images containing the annotation data may be obtained from the human body image not containing the annotation data.
The annotation data is used for representing real posture information in the first human body image, and the real posture information can include human body key node information (the human body key points include positions of the top of the head, the neck, the left shoulder, the left elbow, the left wrist, the right shoulder, the right elbow, the right wrist and the like), human body posture information and the like.
It will be appreciated that there are a number of ways to derive the first human image from a human image that does not contain annotation data, for example: processing the human body image by utilizing the existing human body posture prediction model; or, labeling the human body image manually, and the like, which is not specifically limited in the embodiment of the present application.
Similarly, in the step S101, the unlabeled training set includes a plurality of second human body images that do not include the labeled data, that is, the second human body images are original images that only include the human body postures that have not been subjected to the labeling processing. It is understood that the second human body image can be obtained through various ways such as disclosing an unlabeled portion (e.g., COCO unlabeled) in the dataset.
Further, both the first human body image and the second human body image may be human body images including only one person. Therefore, as for the scene images with a plurality of persons, the scene images can be preprocessed through a pre-trained human body detection model, then the images of each person are scratched out to obtain the corresponding human body images, and the specific steps are as follows:
the method comprises the following steps of firstly, obtaining an original labeled training set containing labeled data and an original unlabeled training set containing no labeled data.
And secondly, respectively carrying out human body detection on the images in the original labeled training set and the images in the original unlabeled training set by using a pre-trained human body detection model to obtain a first human body image in the labeled training set and a second human body image in the unlabeled training set.
In steps S102 to S105, the body posture prediction model provided in the embodiment of the present application may be a generated countermeasure network (GAN), and the body posture prediction model may include two parts, i.e., a generator and a discriminator.
In step S102, for the first human body image input to the generator, the generator outputs a corresponding first human body posture prediction result. As an embodiment, the first human pose prediction output by the generator may comprise a plurality of channels of human pose heat maps. The body pose heat map for each channel may be used to represent a body keypoint location predicted by the generator, and thus, the body pose heat maps for multiple channels are used to represent different body keypoint locations predicted by the generator, respectively. It will be appreciated that on the human pose heat map, the probability that the location is a predicted human keypoint location may be represented in different colors, for example: the closer the color on the human pose heat map is to the red, the more likely this location is to be the predicted corresponding human keypoint location.
After the generator outputs the first human body posture prediction result, a first loss value of the generator can be calculated according to the first human body posture prediction result and the annotation data contained in the first human body image. The step S102 may specifically include the following steps:
the first step is to input the first human body image into a generator to obtain a predicted human body posture heat map corresponding to multiple channels.
And secondly, generating a corresponding reference human body posture heat map based on the annotation data corresponding to the first human body image.
And thirdly, calculating a first loss value according to the predicted human body posture heat map and the reference human body posture heat map.
In the above steps, first, a first human body image may be input to the generator to obtain a predicted human body posture heat map; then, directly mapping the annotation data corresponding to the first human body image into a real human body posture heat map through an algorithm; finally, a first loss value of the generator may be calculated based on the predicted human pose heat map and the real human pose heat map, wherein the first loss value characterizes a difference between the predicted human pose heat map and the real human pose heat map.
In step S103, for the second human body image input to the generator, the generator outputs a corresponding second human body posture prediction result. Similar to the first human pose prediction, as an embodiment, the second human pose prediction output by the generator may also include a plurality of channels of human pose heatmaps. And the second human body image does not contain the annotation data, so the confidence of the human body posture heat map corresponding to the second human body posture prediction result output by the generator is lower.
In step S104, the step S104 may specifically include the following steps:
the method comprises the following steps that firstly, a first human body image and a reference human body posture heat map are used as true data sequences to be input into a discriminator, and a second human body image and a second human body posture prediction result are used as false data sequences to be input into the discriminator to respectively obtain discrimination results output by the discriminator;
and secondly, calculating a second loss value according to the judgment result.
In the above steps, the discriminator is a key module for assigning a field to the second human body image not containing the labeling data. Similar to the existing confrontation training network model, the input of the discriminator provided by the embodiment of the application comprises two parts of true and false. The first human body image containing the annotation data and the reference human body posture heat map generated according to the corresponding annotation data form the input of a 'true' end; the second human body image not containing the labeling data and the corresponding second human body posture prediction result output by the generator form a false end input. The discriminator makes a true and false discrimination on the input data and outputs a probability between 0 and 1 of true and false, thereby calculating a second loss value corresponding to the discriminator.
In this step S105, the corresponding total loss value can be calculated by using the first loss value and the second loss value obtained in the above embodiments. Then, the generator and the discriminator in the human body posture prediction model provided by the embodiment of the application can be optimized by adopting the total loss value, so that the trained human body posture prediction model is obtained.
As an embodiment, the second penalty value is optimized separately for the arbiter and the generator, while the first penalty value is optimized only for the generator. The step S105 may specifically include the following steps:
and updating the network parameters of the generator according to the first loss value and the second loss value, and updating the network parameters of the discriminator according to the second loss value.
In the above steps, when the discriminator is optimized, the second loss value will force the discriminator to make correct true and false discrimination, and when the input is "true" data, the second loss value will make the output of the discriminator as close to 1 as possible; when the input is "false" data, the second penalty value will bring the output of the discriminator as close to 0 as possible.
When the generator is optimized, the first loss value and the second loss value enable the human posture heat map predicted by the generator to conform to the real human posture heat map distribution (as real as possible).
The first loss value and the second loss value enable the two arbiter and generator modules to compete with each other and jointly advance by alternately updating the network parameters of the arbiter and generator.
Further, in some embodiments, the step of obtaining the label training set in step S101 may include the following sub-steps:
the method comprises the steps of firstly, obtaining a plurality of first original images with annotation data; each of the first raw images includes pose data of a human body.
And secondly, carrying out scaling processing on the plurality of first original images to obtain first human body images with the same size and specification.
In the above steps, the accuracy rate of the subsequent training of the generator can be improved and the loss can be reduced by setting the plurality of first human body images to be the same size specification. Of course, for the unlabeled training set, the same processing manner as that of the labeled training set may be adopted, and details are not described here.
As can be seen from the above, the human body posture detection model training method provided in the embodiment of the present application obtains the labeled training set and the unlabeled training set, where the labeled training set includes a plurality of first human body images containing labeled data, and the labeled data is used to represent real posture information in the first human body images; the unmarked training set comprises a plurality of second human body images which do not contain marking data; inputting the first human body image into a generator in a human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeling data and the first human body posture prediction result; inputting the second human body image into a generator to obtain a corresponding second human body posture prediction result; calculating a second loss value corresponding to the discriminator in the human body posture prediction model according to the first human body image, the annotation data, the second human body image and the second human body posture prediction result; and optimizing the generator and the discriminator according to the first loss value and the second loss value to obtain a human body posture prediction model. Therefore, due to the fact that the unmarked training set is applied in the training process, the anti-confrontation training is performed by combining the plurality of unmarked second human body images under the condition that the marked first human body images are insufficient, the accuracy can be improved, and the demand for the marked first human body images is reduced.
Referring to fig. 2, fig. 2 is a flowchart of a human body posture prediction method according to an embodiment of the present application. The human body posture prediction method can comprise the following steps:
s201, acquiring a third human body image.
S202, inputting the third human body image into a generator in the human body posture prediction model obtained by adopting a human body posture prediction model training method to obtain a third human body posture prediction result.
Specifically, in step S201, a third human body image to be predicted may be obtained, where the third human body image may be an original image that does not include annotation data, and the posture of the human body in the third human body image may be predicted by the human body posture prediction method provided in the embodiment of the present application.
In step S202, the third human body image acquired in step S201 may be input into a generator of a human body posture prediction model trained in advance, where the human body posture prediction model may be obtained by using the human body posture prediction model training method in the foregoing embodiment. The generator can output a third human body posture prediction result corresponding to the third human body image according to the third human body image, and the human body posture in the third human body image can be predicted.
Referring to fig. 3, fig. 3 is a schematic structural diagram of a human body posture prediction model training device 300 according to the present application, including: a first obtaining module 301, a first input module 302, a second input module 303, a calculating module 304, and an optimizing module 305.
The first obtaining module 301 is configured to obtain a labeled training set and an unlabeled training set.
The annotation training set includes a plurality of first human images containing annotation data. As an embodiment, a plurality of first human body images containing annotation data can be obtained from a public database; as another embodiment, a human body image not containing the annotation data may be obtained from the network, and then a plurality of first human body images containing the annotation data may be obtained from the human body image not containing the annotation data.
The annotation data is used for representing real posture information in the first human body image, and the real posture information can include human body key node information (the human body key points include positions of the top of the head, the neck, the left shoulder, the left elbow, the left wrist, the right shoulder, the right elbow, the right wrist and the like), human body posture information and the like.
It will be appreciated that there are a number of ways to derive the first human image from a human image that does not contain annotation data, for example: processing the human body image by utilizing the existing human body posture prediction model; or, labeling the human body image manually, and the like, which is not specifically limited in the embodiment of the present application.
Similarly, the unlabeled training set includes a plurality of second human body images that do not include the labeled data, that is, the second human body images are original images that only include the human body poses that are not subjected to the labeling processing. It is understood that the second human body image can be obtained through various ways such as disclosing an unlabeled portion (e.g., COCO unlabeled) in the dataset.
Further, both the first human body image and the second human body image may be human body images including only one person. Therefore, as for the scene images with a plurality of persons, the scene images can be preprocessed through a pre-trained human body detection model, then the images of each person are scratched out to obtain the corresponding human body images, and the specific steps are as follows:
the method comprises the following steps of firstly, obtaining an original labeled training set containing labeled data and an original unlabeled training set containing no labeled data.
And secondly, respectively carrying out human body detection on the images in the original labeled training set and the images in the original unlabeled training set by using a pre-trained human body detection model to obtain a first human body image in the labeled training set and a second human body image in the unlabeled training set.
The first input module 302 is configured to input a first human body image into a generator in a human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculate a first loss value of the generator according to the annotation data and the first human body posture prediction result.
For a first human body image input to the generator, the generator outputs a corresponding first human body posture prediction result. As an embodiment, the first human pose prediction output by the generator may comprise a plurality of channels of human pose heat maps. The body pose heat map for each channel may be used to represent a body keypoint location predicted by the generator, and thus, the body pose heat maps for multiple channels are used to represent different body keypoint locations predicted by the generator, respectively. It will be appreciated that the closer the color on the human pose heat map is to the red, the more likely it is that the location is the predicted corresponding human keypoint location.
After the generator outputs the first human body posture prediction result, a first loss value of the generator can be calculated according to the first human body posture prediction result and the annotation data contained in the first human body image.
In the human body posture detection model training apparatus 300 according to the embodiment of the present application, the first input module 302 is specifically configured to: inputting the first human body image into a generator to obtain a predicted human body posture heat map corresponding to multiple channels; wherein the predicted human posture heat map of each channel predicts a human body key point position; generating a corresponding reference human body posture heat map based on the annotation data corresponding to the first human body image; a first loss value is calculated from the predicted human pose heat map and the reference human pose heat map.
First, a first human body image may be input to a generator to obtain a predicted human body posture heat map; then, directly mapping the annotation data corresponding to the first human body image into a real human body posture heat map through an algorithm; finally, based on the predicted human pose heat map and the real human pose heat map, a first loss value for the generator may be calculated.
The second input module 303 is configured to input a second human body image into the generator to obtain a corresponding second human body posture prediction result.
For the second human body image input into the generator, the generator outputs a corresponding second human body posture prediction result. Similar to the first human pose prediction, as an embodiment, the second human pose prediction output by the generator may also include a plurality of channels of human pose heatmaps. And the second human body image does not contain the annotation data, so the confidence of the human body posture heat map corresponding to the second human body posture prediction result output by the generator is lower.
The calculating module 304 is configured to calculate a second loss value corresponding to the discriminator in the human body posture prediction model according to the first human body image, the labeled data, the second human body image, and the second human body posture prediction result.
In the human body posture detection model training device 300 according to the embodiment of the present application, the calculation module 304 is specifically configured to: inputting the first human body image and the reference human body posture heat map into a discriminator as a true data sequence, and inputting the second human body image and the second human body posture prediction result into the discriminator as a false data sequence to respectively obtain discrimination results output by the discriminator; and calculating a second loss value according to the judgment result.
In the above steps, the discriminator is a key module for assigning a field to the second human body image not containing the labeling data. Similar to the existing confrontational training network model, the input of the discriminator comprises two parts of true and false. The first human body image containing the annotation data and the reference human body posture heat map generated by the corresponding annotation data form the input of a 'true' end; the second human body image which does not contain the labeling data and the corresponding second human body posture prediction result form a false end input. The discriminator determines whether the input data is true or false, and outputs a probability of true or false between 0 and 1, thereby calculating the resistance loss function.
The optimization module 305 is configured to optimize the generator and the discriminator according to the first loss value and the second loss value to obtain a human body posture prediction model.
The first loss value and the second loss value obtained in the above embodiments may be used to calculate a corresponding total loss function. Then, the generator and the discriminator in the human body posture prediction model provided by the embodiment of the application can be optimized by adopting the total loss function, so that the trained human body posture prediction model is obtained.
As an embodiment, the second penalty value is optimized separately for the arbiter and the generator, while the first penalty value is optimized only for the generator. In the human body posture detection model training device 300 according to the embodiment of the present application, the optimization module 305 is specifically configured to: and updating the network parameters of the generator according to the first loss value and the second loss value, and updating the network parameters of the discriminator according to the second loss value.
When the discriminator is optimized, the second loss value forces the discriminator to make correct true and false discrimination, and when the input data is 'true' data, the output of the discriminator is close to 1 as much as possible; when the input is "false" data, the second penalty value will bring the output of the discriminator as close to 0 as possible.
When the generator is optimized, the first loss value and the second loss value enable the human posture heat map predicted by the generator to conform to the real human posture heat map distribution (as real as possible).
The first loss value and the second loss value enable the two arbiter and generator modules to compete with each other and jointly advance by alternately updating the network parameters of the arbiter and generator.
Further, in some embodiments, the first obtaining module 301 is specifically configured to: acquiring a plurality of first original images with annotation data; each first original image comprises posture data of a human body; and carrying out scaling processing on the plurality of first original images to obtain first human body images with the same size specification.
The accuracy rate of the plurality of first human body images during subsequent training of the generator can be improved by setting the plurality of first human body images to be in the same size specification, and loss is reduced. Of course, correspondingly, for unlabeled training sets.
As can be seen from the above, the human body posture detection model training device 300 provided in the embodiment of the present application obtains a labeled training set and an unlabeled training set, where the labeled training set includes a plurality of first human body images containing labeled data, and the labeled data is used to represent real posture information in the first human body images; the unmarked training set comprises a plurality of second human body images which do not contain marking data; inputting the first human body image into a generator in a human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeling data and the first human body posture prediction result; inputting the second human body image into a generator to obtain a corresponding second human body posture prediction result; calculating a second loss value corresponding to the discriminator in the human body posture prediction model according to the first human body image, the annotation data, the second human body image and the second human body posture prediction result; and optimizing the generator and the discriminator according to the first loss value and the second loss value to obtain a human body posture prediction model. Therefore, due to the fact that the unmarked training set is applied in the training process, the anti-confrontation training is performed by combining the plurality of unmarked second human body images under the condition that the marked first human body images are insufficient, the accuracy can be improved, and the demand for the marked first human body images is reduced.
Referring to fig. 4, fig. 4 is a schematic structural diagram of a human body posture prediction apparatus 400 according to an embodiment of the present application, including: a second acquisition module 401 and a prediction module 402.
The second obtaining module 401 is configured to obtain a third human body image.
The method for predicting the human body posture can be used for predicting the posture of the human body in the third human body image.
The prediction module 402 is configured to input the third human body image into a generator in a human body posture prediction model obtained by using a human body posture prediction model training method, so as to obtain a third human body posture prediction result.
The third human body image acquired in step S201 may be input into a generator of a human body posture prediction model trained in advance, where the human body posture prediction model may be obtained by using the human body posture prediction model training method in the foregoing embodiment. The generator can output a third human body posture prediction result corresponding to the third human body image according to the third human body image, and the human body posture in the third human body image can be predicted.
Referring to fig. 5, fig. 5 is a schematic structural diagram of an electronic device according to an embodiment of the present disclosure, which provides an electronic device 500, including: the processor 501 and the memory 502, the processor 501 and the memory 502 being interconnected and communicating with each other via a communication bus 503 and/or other form of connection mechanism (not shown), the memory 502 storing a computer program executable by the processor 501, the computer program being executable by the processor 501 when the computing device is running, the processor 501 executing the computer program to perform the method of any of the alternative implementations of the embodiments described above.
The embodiment of the present application provides a storage medium, and when being executed by a processor, the computer program performs the method in any optional implementation manner of the above embodiment.
The storage medium may be implemented by any type of volatile or nonvolatile storage device or combination thereof, such as a Static Random Access Memory (SRAM), an Electrically Erasable Programmable Read-Only Memory (EEPROM), an Erasable Programmable Read-Only Memory (EPROM), a Programmable Read-Only Memory (PROM), a Read-Only Memory (ROM), a magnetic Memory, a flash Memory, a magnetic disk, or an optical disk.
In the embodiments provided in the present application, it should be understood that the disclosed apparatus and method may be implemented in other ways. The above-described embodiments of the apparatus are merely illustrative, and for example, the division of the units is only one logical division, and there may be other divisions when actually implemented, and for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection of devices or units through some communication interfaces, and may be in an electrical, mechanical or other form.
In addition, units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
Furthermore, the functional modules in the embodiments of the present application may be integrated together to form an independent part, or each module may exist separately, or two or more modules may be integrated to form an independent part.
In this document, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions.
The above description is only an example of the present application and is not intended to limit the scope of the present application, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present application shall be included in the protection scope of the present application.
Claims (10)
1. A human posture prediction model training method is characterized by comprising the following steps:
acquiring a labeled training set and an unlabeled training set, wherein the labeled training set comprises a plurality of first human body images containing labeled data, and the labeled data is used for representing real posture information in the first human body images; the unmarked training set comprises a plurality of second human body images which do not contain marking data;
inputting the first human body image into a generator in a human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeling data and the first human body posture prediction result;
inputting the second human body image into the generator to obtain a corresponding second human body posture prediction result;
calculating a second loss value corresponding to a discriminator in the human body posture prediction model according to the first human body image, the labeled data, the second human body image and the second human body posture prediction result;
and optimizing the generator and the discriminator according to the first loss value and the second loss value to obtain the human body posture prediction model.
2. The method for training the human body posture prediction model according to claim 1, wherein the inputting the first human body image into a generator in the human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeled data and the first human body posture prediction result comprises:
inputting the first human body image into the generator to obtain a predicted human body posture heat map of a corresponding multi-channel; wherein the predicted human posture heat map of each channel predicts a human body key point position;
generating a corresponding reference human body posture heat map based on the annotation data corresponding to the first human body image;
calculating the first loss value from the predicted human pose heat map and the reference human pose heat map.
3. The method for training the human body posture prediction model according to claim 2, wherein the calculating a second loss value corresponding to a discriminator in the human body posture prediction model according to the first human body image, the annotation data, the second human body image and the second human body posture prediction result comprises:
inputting the first human body image and the reference human body posture heat map into the discriminator as a true data sequence, and inputting the second human body image and the second human body posture prediction result into the discriminator as a false data sequence to respectively obtain discrimination results output by the discriminator;
and calculating the second loss value according to the judgment result.
4. The human pose prediction model training method of any one of claims 1-3, wherein the optimizing the generator and the discriminator according to the first loss value and the second loss value comprises:
and updating the network parameters of the generator according to the first loss value and the second loss value, and updating the network parameters of the discriminator according to the second loss value.
5. The human body posture prediction model training method according to any one of claims 1-4, wherein the obtaining of the labeled training set and the unlabeled training set comprises:
acquiring an original labeling training set containing labeling data and an original unlabeled training set not containing labeling data;
respectively carrying out human body detection on the images in the original labeled training set and the images in the original unlabeled training set by using a pre-trained human body detection model to obtain the first human body images in the labeled training set and the second human body images in the unlabeled training set; wherein the first human body image and the second human body image are both single person images.
6. A human posture prediction method is characterized by comprising the following steps:
acquiring a third human body image;
inputting the third human body image into a generator of the human body posture prediction model obtained by the human body posture prediction model training method of any one of claims 1 to 5 to obtain a third human body posture prediction result.
7. A human body posture prediction model training device is characterized by comprising:
the system comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for acquiring a labeled training set and an unlabeled training set, the labeled training set comprises a plurality of first human body images containing labeled data, and the labeled data is used for representing real posture information in the first human body images; the unmarked training set comprises a plurality of second human body images which do not contain marking data;
the first input module is used for inputting the first human body image into a generator in a human body posture prediction model to obtain a corresponding first human body posture prediction result, and calculating a first loss value of the generator according to the labeling data and the first human body posture prediction result;
the second input module is used for inputting the second human body image into the generator to obtain a corresponding second human body posture prediction result;
a calculating module, configured to calculate a second loss value corresponding to a discriminator in the human body posture prediction model according to the first human body image, the annotation data, the second human body image, and the second human body posture prediction result;
and the optimization module is used for optimizing the generator and the discriminator according to the first loss value and the second loss value to obtain the human body posture prediction model.
8. A human body posture prediction apparatus, comprising:
the second acquisition module is used for acquiring a third human body image;
a prediction module, configured to input the third human body image into a generator in a human body posture prediction model obtained by using the human body posture prediction model training method according to any one of claims 1 to 5, so as to obtain a third human body posture prediction result.
9. An electronic device comprising a processor and a memory, the memory storing computer readable instructions that, when executed by the processor, perform the method of any one of claims 1-5 or the method of claim 6.
10. A storage medium having a computer program stored thereon, wherein the computer program, when executed by a processor, performs the method of any one of claims 1-5 or the method of claim 6.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110658308.5A CN113569627B (en) | 2021-06-11 | 2021-06-11 | Human body posture prediction model training method, human body posture prediction method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110658308.5A CN113569627B (en) | 2021-06-11 | 2021-06-11 | Human body posture prediction model training method, human body posture prediction method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113569627A true CN113569627A (en) | 2021-10-29 |
CN113569627B CN113569627B (en) | 2024-06-14 |
Family
ID=78162053
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110658308.5A Active CN113569627B (en) | 2021-06-11 | 2021-06-11 | Human body posture prediction model training method, human body posture prediction method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113569627B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114926709A (en) * | 2022-05-26 | 2022-08-19 | 成都极米科技股份有限公司 | Data labeling method and device and electronic equipment |
CN116311519A (en) * | 2023-03-17 | 2023-06-23 | 北京百度网讯科技有限公司 | Action recognition method, model training method and device |
WO2024037546A1 (en) * | 2022-08-19 | 2024-02-22 | 北京字跳网络技术有限公司 | Method and apparatus for recognizing human body action, and device and medium |
Citations (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130250050A1 (en) * | 2012-03-23 | 2013-09-26 | Objectvideo, Inc. | Video surveillance systems, devices and methods with improved 3d human pose and shape modeling |
CN108090902A (en) * | 2017-12-30 | 2018-05-29 | 中国传媒大学 | A kind of non-reference picture assessment method for encoding quality based on multiple dimensioned generation confrontation network |
CN108875510A (en) * | 2017-11-28 | 2018-11-23 | 北京旷视科技有限公司 | Method, apparatus, system and the computer storage medium of image procossing |
CN108875732A (en) * | 2018-01-11 | 2018-11-23 | 北京旷视科技有限公司 | Model training and example dividing method, device and system and storage medium |
CN109711329A (en) * | 2018-12-25 | 2019-05-03 | 北京迈格威科技有限公司 | Attitude estimation and network training method, device and system and storage medium |
CN109859296A (en) * | 2019-02-01 | 2019-06-07 | 腾讯科技(深圳)有限公司 | Training method, server and the storage medium of SMPL parametric prediction model |
CN109934165A (en) * | 2019-03-12 | 2019-06-25 | 南方科技大学 | Joint point detection method and device, storage medium and electronic equipment |
CN110008835A (en) * | 2019-03-05 | 2019-07-12 | 成都旷视金智科技有限公司 | Sight prediction technique, device, system and readable storage medium storing program for executing |
CN110163082A (en) * | 2019-04-02 | 2019-08-23 | 腾讯科技(深圳)有限公司 | A kind of image recognition network model training method, image-recognizing method and device |
CN110188633A (en) * | 2019-05-14 | 2019-08-30 | 广州虎牙信息科技有限公司 | Human body posture index prediction technique, device, electronic equipment and storage medium |
WO2019228317A1 (en) * | 2018-05-28 | 2019-12-05 | 华为技术有限公司 | Face recognition method and device, and computer readable medium |
CN110598554A (en) * | 2019-08-09 | 2019-12-20 | 中国地质大学(武汉) | Multi-person posture estimation method based on counterstudy |
US20200019799A1 (en) * | 2018-07-10 | 2020-01-16 | DeepScale, Inc. | Automated annotation techniques |
CN111311729A (en) * | 2020-01-18 | 2020-06-19 | 西安电子科技大学 | Natural scene three-dimensional human body posture reconstruction method based on bidirectional projection network |
CN111428650A (en) * | 2020-03-26 | 2020-07-17 | 北京工业大学 | Pedestrian re-identification method based on SP-PGGAN style migration |
CN111523422A (en) * | 2020-04-15 | 2020-08-11 | 北京华捷艾米科技有限公司 | Key point detection model training method, key point detection method and device |
CN112016501A (en) * | 2020-09-04 | 2020-12-01 | 平安科技(深圳)有限公司 | Training method and device of face recognition model and computer equipment |
CN112132172A (en) * | 2020-08-04 | 2020-12-25 | 绍兴埃瓦科技有限公司 | Model training method, device, equipment and medium based on image processing |
CN112149645A (en) * | 2020-11-10 | 2020-12-29 | 西北工业大学 | Human body posture key point identification method based on generation of confrontation learning and graph neural network |
WO2021007859A1 (en) * | 2019-07-18 | 2021-01-21 | 华为技术有限公司 | Method and apparatus for estimating pose of human body |
CN112464895A (en) * | 2020-12-14 | 2021-03-09 | 深圳市优必选科技股份有限公司 | Posture recognition model training method and device, posture recognition method and terminal equipment |
CN112529073A (en) * | 2020-12-07 | 2021-03-19 | 北京百度网讯科技有限公司 | Model training method, attitude estimation method and apparatus, and electronic device |
US20210097718A1 (en) * | 2019-09-27 | 2021-04-01 | Martin Adrian FISCH | Methods and apparatus for orientation keypoints for complete 3d human pose computerized estimation |
CN112700408A (en) * | 2020-12-28 | 2021-04-23 | 中国银联股份有限公司 | Model training method, image quality evaluation method and device |
CN112801895A (en) * | 2021-01-15 | 2021-05-14 | 南京邮电大学 | Two-stage attention mechanism-based GAN network image restoration algorithm |
-
2021
- 2021-06-11 CN CN202110658308.5A patent/CN113569627B/en active Active
Patent Citations (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130250050A1 (en) * | 2012-03-23 | 2013-09-26 | Objectvideo, Inc. | Video surveillance systems, devices and methods with improved 3d human pose and shape modeling |
CN108875510A (en) * | 2017-11-28 | 2018-11-23 | 北京旷视科技有限公司 | Method, apparatus, system and the computer storage medium of image procossing |
CN108090902A (en) * | 2017-12-30 | 2018-05-29 | 中国传媒大学 | A kind of non-reference picture assessment method for encoding quality based on multiple dimensioned generation confrontation network |
CN108875732A (en) * | 2018-01-11 | 2018-11-23 | 北京旷视科技有限公司 | Model training and example dividing method, device and system and storage medium |
WO2019228317A1 (en) * | 2018-05-28 | 2019-12-05 | 华为技术有限公司 | Face recognition method and device, and computer readable medium |
US20200019799A1 (en) * | 2018-07-10 | 2020-01-16 | DeepScale, Inc. | Automated annotation techniques |
CN109711329A (en) * | 2018-12-25 | 2019-05-03 | 北京迈格威科技有限公司 | Attitude estimation and network training method, device and system and storage medium |
CN109859296A (en) * | 2019-02-01 | 2019-06-07 | 腾讯科技(深圳)有限公司 | Training method, server and the storage medium of SMPL parametric prediction model |
CN110008835A (en) * | 2019-03-05 | 2019-07-12 | 成都旷视金智科技有限公司 | Sight prediction technique, device, system and readable storage medium storing program for executing |
CN109934165A (en) * | 2019-03-12 | 2019-06-25 | 南方科技大学 | Joint point detection method and device, storage medium and electronic equipment |
CN110163082A (en) * | 2019-04-02 | 2019-08-23 | 腾讯科技(深圳)有限公司 | A kind of image recognition network model training method, image-recognizing method and device |
CN110188633A (en) * | 2019-05-14 | 2019-08-30 | 广州虎牙信息科技有限公司 | Human body posture index prediction technique, device, electronic equipment and storage medium |
WO2021007859A1 (en) * | 2019-07-18 | 2021-01-21 | 华为技术有限公司 | Method and apparatus for estimating pose of human body |
CN110598554A (en) * | 2019-08-09 | 2019-12-20 | 中国地质大学(武汉) | Multi-person posture estimation method based on counterstudy |
US20210097718A1 (en) * | 2019-09-27 | 2021-04-01 | Martin Adrian FISCH | Methods and apparatus for orientation keypoints for complete 3d human pose computerized estimation |
CN111311729A (en) * | 2020-01-18 | 2020-06-19 | 西安电子科技大学 | Natural scene three-dimensional human body posture reconstruction method based on bidirectional projection network |
CN111428650A (en) * | 2020-03-26 | 2020-07-17 | 北京工业大学 | Pedestrian re-identification method based on SP-PGGAN style migration |
CN111523422A (en) * | 2020-04-15 | 2020-08-11 | 北京华捷艾米科技有限公司 | Key point detection model training method, key point detection method and device |
CN112132172A (en) * | 2020-08-04 | 2020-12-25 | 绍兴埃瓦科技有限公司 | Model training method, device, equipment and medium based on image processing |
CN112016501A (en) * | 2020-09-04 | 2020-12-01 | 平安科技(深圳)有限公司 | Training method and device of face recognition model and computer equipment |
CN112149645A (en) * | 2020-11-10 | 2020-12-29 | 西北工业大学 | Human body posture key point identification method based on generation of confrontation learning and graph neural network |
CN112529073A (en) * | 2020-12-07 | 2021-03-19 | 北京百度网讯科技有限公司 | Model training method, attitude estimation method and apparatus, and electronic device |
CN112464895A (en) * | 2020-12-14 | 2021-03-09 | 深圳市优必选科技股份有限公司 | Posture recognition model training method and device, posture recognition method and terminal equipment |
CN112700408A (en) * | 2020-12-28 | 2021-04-23 | 中国银联股份有限公司 | Model training method, image quality evaluation method and device |
CN112801895A (en) * | 2021-01-15 | 2021-05-14 | 南京邮电大学 | Two-stage attention mechanism-based GAN network image restoration algorithm |
Non-Patent Citations (1)
Title |
---|
刘一敏;蒋建国;齐美彬;刘皓;周华捷;: "融合生成对抗网络和姿态估计的视频行人再识别方法", 自动化学报, vol. 46, no. 03, pages 576 - 584 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114926709A (en) * | 2022-05-26 | 2022-08-19 | 成都极米科技股份有限公司 | Data labeling method and device and electronic equipment |
WO2024037546A1 (en) * | 2022-08-19 | 2024-02-22 | 北京字跳网络技术有限公司 | Method and apparatus for recognizing human body action, and device and medium |
CN116311519A (en) * | 2023-03-17 | 2023-06-23 | 北京百度网讯科技有限公司 | Action recognition method, model training method and device |
CN116311519B (en) * | 2023-03-17 | 2024-04-19 | 北京百度网讯科技有限公司 | Action recognition method, model training method and device |
Also Published As
Publication number | Publication date |
---|---|
CN113569627B (en) | 2024-06-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110414432B (en) | Training method of object recognition model, object recognition method and corresponding device | |
CN110175527B (en) | Pedestrian re-identification method and device, computer equipment and readable medium | |
CN113569627B (en) | Human body posture prediction model training method, human body posture prediction method and device | |
CN114758362B (en) | Clothing changing pedestrian re-identification method based on semantic perception attention and visual shielding | |
CN114821014B (en) | Multi-mode and countermeasure learning-based multi-task target detection and identification method and device | |
CN111368672A (en) | Construction method and device for genetic disease facial recognition model | |
CN114998934B (en) | Clothes-changing pedestrian re-identification and retrieval method based on multi-mode intelligent perception and fusion | |
CN111553267A (en) | Image processing method, image processing model training method and device | |
CN115565238B (en) | Face-changing model training method, face-changing model training device, face-changing model training apparatus, storage medium, and program product | |
CN110866469B (en) | Facial five sense organs identification method, device, equipment and medium | |
CN112801236B (en) | Image recognition model migration method, device, equipment and storage medium | |
CN116129473B (en) | Identity-guide-based combined learning clothing changing pedestrian re-identification method and system | |
CN115115969A (en) | Video detection method, apparatus, device, storage medium and program product | |
CN114937293A (en) | Agricultural service management method and system based on GIS | |
CN114219971A (en) | Data processing method, data processing equipment and computer readable storage medium | |
CN116453226A (en) | Human body posture recognition method and device based on artificial intelligence and related equipment | |
CN116091596A (en) | Multi-person 2D human body posture estimation method and device from bottom to top | |
CN113673308B (en) | Object identification method, device and electronic system | |
CN111368761A (en) | Shop business state recognition method and device, readable storage medium and equipment | |
CN115577768A (en) | Semi-supervised model training method and device | |
CN114385846A (en) | Image classification method, electronic device, storage medium and program product | |
CN113705293A (en) | Image scene recognition method, device, equipment and readable storage medium | |
Gonzalez-Soler et al. | Semi-synthetic data generation for tattoo segmentation | |
CN112529116B (en) | Scene element fusion processing method, device and equipment and computer storage medium | |
CN115841605A (en) | Target detection network training and target detection method, electronic device and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |