CN111368663A - Method, device, medium and equipment for recognizing static facial expressions in natural scene - Google Patents
Method, device, medium and equipment for recognizing static facial expressions in natural scene Download PDFInfo
- Publication number
- CN111368663A CN111368663A CN202010115562.6A CN202010115562A CN111368663A CN 111368663 A CN111368663 A CN 111368663A CN 202010115562 A CN202010115562 A CN 202010115562A CN 111368663 A CN111368663 A CN 111368663A
- Authority
- CN
- China
- Prior art keywords
- feature
- local
- feature extraction
- extraction process
- picture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 64
- 230000008921 facial expression Effects 0.000 title claims abstract description 50
- 230000003068 static effect Effects 0.000 title claims abstract description 27
- 238000000605 extraction Methods 0.000 claims abstract description 58
- 239000013598 vector Substances 0.000 claims abstract description 42
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 25
- 238000001514 detection method Methods 0.000 claims abstract description 25
- 238000007781 pre-processing Methods 0.000 claims abstract description 22
- 239000011159 matrix material Substances 0.000 claims abstract description 12
- 238000007499 fusion processing Methods 0.000 claims abstract description 8
- 230000008569 process Effects 0.000 claims abstract description 8
- 230000014509 gene expression Effects 0.000 claims description 21
- 230000004927 fusion Effects 0.000 claims description 8
- 238000004590 computer program Methods 0.000 claims description 3
- 230000000694 effects Effects 0.000 abstract description 5
- 238000011160 research Methods 0.000 description 7
- 230000008451 emotion Effects 0.000 description 5
- 238000012549 training Methods 0.000 description 5
- 238000004364 calculation method Methods 0.000 description 4
- 238000011161 development Methods 0.000 description 4
- 230000003993 interaction Effects 0.000 description 4
- 238000012360 testing method Methods 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000011840 criminal investigation Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 201000010099 disease Diseases 0.000 description 1
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 1
- 230000002996 emotional effect Effects 0.000 description 1
- 210000001097 facial muscle Anatomy 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000007935 neutral effect Effects 0.000 description 1
- 230000036544 posture Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/169—Holistic features and representations, i.e. based on the facial image taken as a whole
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
- G06V40/175—Static expression
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Human Computer Interaction (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Probability & Statistics with Applications (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention provides a method, a device, a medium and equipment for identifying static facial expressions in a natural scene; the method comprises a pretreatment process, a global feature extraction process, a local feature extraction process and a feature fusion process which are sequentially executed; the preprocessing procedure is used for removing the background of a natural scene in the picture and only reserving a face area; the global feature extraction process is to adopt a convolutional neural network to extract the features of the whole picture after the preprocessing process and convert the features into feature vectors; the local feature extraction process is to extract a local area with the most information quantity of the picture by a target detection method and convert the local area into a feature vector; the feature fusion process is to integrate the feature vectors extracted in the global feature extraction process and the feature vectors extracted in the local feature extraction process into a feature matrix, and then obtain the facial expression probability through a full connection layer and softmax. The facial expression recognition method is high in accuracy rate, high in recognition efficiency and good in recognition effect.
Description
Technical Field
The invention relates to the technical field of computer vision, in particular to a method, a device, a medium and equipment for recognizing static facial expressions in natural scenes.
Background
With the development of artificial intelligence technology, the application of facial expression recognition attracts a wide range of concerns, such as: the facial expression recognition method can be applied to the fields of fatigue driving, life entertainment, criminal investigation, medical diagnosis, virtual reality, intelligent education and the like. Particularly, with the development of emotion calculation research, the research on facial expressions is more hot, and according to the research of related researchers, the information contained in the facial expressions can reach 55%, so that the research on facial expression recognition is of great importance to emotion calculation.
For a long time, most researchers have studied facial expression data sets under laboratory conditions; with the development of research and the publication of data sets under natural conditions, the current research focus gradually turns to the recognition of facial expressions under natural conditions. Compared with the expression image under the natural condition, the expression image under the natural condition has the problems of multiple postures, shielding and the like under the laboratory condition, so that the traditional manual characteristic method and the shallow convolutional neural network method are difficult to obtain better identification accuracy on the expression data set under the natural condition. Because the expressions are formed by combining the movements of facial muscles, the characteristics of the expressions are finer compared with other images, the expressions of the same category have larger difference, and the different expressions are often local characteristic changes, so that a model capable of better extracting the characteristics with discrimination and the local characteristics is required to be used, and the accuracy of facial expression recognition can be improved.
Disclosure of Invention
The invention aims to overcome the defects in the prior art, and provides a method, a device, a medium and equipment for identifying static facial expressions in natural scenes, which have the advantages of high accuracy and high identification efficiency and have good identification effect.
In order to achieve the purpose, the invention is realized by the following technical scheme: a method for recognizing static facial expressions in natural scenes is characterized by comprising the following steps: the method comprises a pretreatment process, a global feature extraction process, a local feature extraction process and a feature fusion process which are sequentially executed;
the preprocessing procedure is used for removing the background of a natural scene in the picture and only reserving a face area;
the global feature extraction process is to extract the features of the whole picture after the preprocessing process by adopting a convolutional neural network and convert the features into feature vectors;
the local feature extraction process is to extract a local area with the most information quantity of the picture by a target detection method and convert the local area into a feature vector;
the feature fusion process is to integrate the feature vectors extracted by the global feature extraction process and the feature vectors extracted by the local feature extraction process into a feature matrix, and then obtain the facial expression probability through a full connection layer and softmax.
Preferably, the preprocessing procedure refers to performing face detection, face alignment and face cropping operations on the picture.
Preferably, the local feature extraction process includes the following sub-steps:
s1, generating a plurality of target detection frames by adopting a target detection method;
s2, converting the local area feature map of the target detection box in the S1 into a score of a corresponding expression category by using a convolutional neural network, and expressing the score as follows by using a formula:
S:A→[0,1]
wherein S represents a score set of local regions, A is a set of all local regions, and R ∈ A, the score of the local region is set as C (R), the information content of the local region is set as I (R), and the information content of all local regions is ordered:
R1,R2∈A,ifC(R1)>C(R2),I(R1)>I(R2)
s3, selecting the first K local areas with the largest information quantity;
and S4, extracting K local area features screened in the S3 by adopting a convolutional neural network, and converting the K local area features into feature vectors.
Preferably, in steps S2 and S4 of the global feature extraction process and the local feature extraction process, the convolutional neural network uses resnet50 as a base network.
Preferably, in step S4 of the global feature extraction step and the local feature extraction step, the feature vectors are both M × 1-sized vectors;
in the feature fusion step, the size of the feature matrix is M × (K + 1).
An apparatus for recognizing facial expressions based on static states in a natural scene, comprising:
the preprocessing module is used for removing the background of a natural scene in the picture and only reserving a face area;
the global feature extraction module is used for extracting the features of the whole picture processed by the preprocessing module by adopting a convolutional neural network and converting the features into feature vectors;
the local feature extraction module is used for extracting a local area with the most information quantity of the picture by a target detection method and converting the local area into a feature vector;
and the feature fusion module is used for integrating the feature vectors extracted by the global feature extraction module and the feature vectors extracted by the local feature extraction module into a feature matrix, and then obtaining the facial expression probability through a full connection layer and softmax.
A storage medium, wherein the storage medium stores a computer program that, when executed by a processor, causes the processor to execute the above-described static facial expression recognition method in a natural scene.
A computing device comprising a processor and a memory for storing a program executable by the processor, wherein the processor implements the method for recognizing static facial expressions in natural scenes when executing the program stored in the memory.
Compared with the prior art, the invention has the following advantages and beneficial effects:
according to the method, a data set in a natural scene is preprocessed, and global and local features are extracted to finally fuse the global and local features to obtain a model for recognizing the expression, so that the static facial expression is accurately recognized; the facial expression recognition accuracy is high, the recognition efficiency is high, and a good recognition effect is achieved; the method can be used for human emotion calculation and a human-computer interaction system, and can improve the working efficiency of a facial expression correlation system.
Drawings
FIG. 1 is a schematic flow chart of a method for recognizing static facial expressions in a natural scene according to the present invention;
fig. 2 is a flow chart of local feature extraction in the static facial expression recognition method in natural scenes according to the present invention.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and specific embodiments.
Example one
As shown in fig. 1, the method for recognizing static facial expressions in a natural scene in the present embodiment is characterized in that: the method comprises a preprocessing procedure, a global feature extraction procedure, a local feature extraction procedure and a feature fusion procedure which are sequentially executed.
The preprocessing procedure is used for removing the background of a natural scene in the picture and only reserving a face area; the method refers to the operations of face detection, face alignment and face cutting of a picture.
The global feature extraction process is to extract the features of the whole picture after the preprocessing process by adopting a convolutional neural network and convert the features into feature vectors.
The local feature extraction process is to extract a local area with the most information quantity of the picture by a target detection method and convert the local area into a feature vector.
The local feature extraction process includes the following sub-steps, as shown in fig. 2:
s1, generating a plurality of target detection frames by adopting a target detection method;
s2, converting the local area feature map of the target detection box in the S1 into a score of a corresponding expression category by using a convolutional neural network, and expressing the score as follows by using a formula:
S:A→[0,1]
wherein S represents a score set of local regions, A is a set of all local regions, and R ∈ A, the score of the local region is set as C (R), the information content of the local region is set as I (R), and the information content of all local regions is ordered:
R1,R2∈A,ifC(R1)>C(R2),I(R1)>I(R2)
s3, selecting the first K local areas with the largest information quantity;
and S4, extracting K local area features screened in the S3 by adopting a convolutional neural network, and converting the K local area features into feature vectors.
When the score corresponding to the local area feature is obtained, the feature needs to be extracted through a resnet50 network, a final score value is obtained through a classifier, the area with the larger score value represents the area with the most information quantity, the area is screened out to be used as a local area module, the specific area number K is used as a hyper-parameter of the model, and manual adjustment is needed.
The feature fusion process is to integrate the feature vectors extracted in the global feature extraction process and the feature vectors extracted in the local feature extraction process into a feature matrix, and then obtain the facial expression probability through a full connection layer and softmax.
In steps S2 and S4 of the global feature extraction step and the local feature extraction step, the convolutional neural network uses resnet50 as a base network.
In step S4 of the global feature extraction step and the local feature extraction step, the feature vectors are all M × 1 vectors, and in the feature fusion step, the feature matrix is M × (K + 1).
Convolutional neural network training may employ existing techniques. Specifically, the convolutional neural network is essentially function fitting, for example, simple y ═ wx, a weight w is obtained by training a model, first, w is initialized, after data is input each time, y is obtained, a difference between y and an expected target value is calculated, y is continuously approximated, and finally w is obtained, that is, the convolutional neural network.
The following description will be made of specific embodiments.
Based on the public facial expression data RAF-DB, a static facial expression recognition method in a natural scene is provided, an accurate solution is provided for facial expression recognition in the natural scene, the working efficiency of a facial expression related system is improved, and an emotion man-machine interaction system is perfected.
In this embodiment, the pictures of the data set are all processed into pictures with the size of 224 × 224 through a preprocessing procedure, and are classified into expressions, and the output categories of the classifier are mainly classified into seven categories such as anger (anger), depression (dismost), fear (fear), happy (happy), neutral (netural), sadness (sad), surprise (surpride) and the like according to the labels of the data set, so that the number of the output categories of the classifier is set to be 7.
The method for recognizing the static facial expression in the natural scene comprises the following steps:
the method comprises the following steps of preprocessing, wherein the preprocessing comprises the operations of face detection, face alignment, face cutting and the like, five key point coordinates in a face are detected by adopting a multitask cascade (MTCNN) face detection model, the left eye coordinates and the right eye coordinates are connected to calculate the included angle theta between the left eye coordinates and the right eye coordinates and the horizontal direction, the image is rotated clockwise by the angle theta according to the left eye coordinates as an axis to obtain a new image and new key point coordinates, the image is cut according to the new key point coordinates, and the cut image is stored as a 224-size face image with the size of 224 × 224.
And a global feature extraction process, wherein a convolutional neural network is adopted to extract the features of the whole picture after the preprocessing process and convert the features into feature vectors, and the convolutional neural network adopts the resnet50 without a full connection layer as a feature extraction network and converts the feature matrix into the feature vectors of 2048 × 1.
The local feature extraction process mainly adopts an extraction frame in target detection to extract the features of local areas, and selects an area containing the maximum information amount from all the local areas as the local area features of expression recognition, and the detailed steps are as follows:
s1, generating a plurality of target detection frames by adopting a target detection method;
s2, converting the local area feature map of the target detection box in the S1 into a score of a corresponding expression category by using a convolutional neural network, and expressing the score as follows by using a formula:
S:A→[0,1]
wherein S represents the fraction of the local area, A is the set of all the local areas, and R ∈ A, the fraction of the local area is set as C (R), the information content of the local area is set as I (R), the information content of all the local areas is ordered:
R1,R2∈A,ifC(R1)>C(R2),I(R1)>I(R2)
s3, selecting the first K local areas with the largest information quantity;
and S4, extracting K local region features screened in S3 by adopting a convolutional neural network, and converting the K local region features into K2048 × 1 feature vectors.
And a feature fusion process, namely integrating the feature vectors extracted in the global feature extraction process and the feature vectors extracted in the local feature extraction process into a 2048 × (K +1) feature matrix, and then obtaining the facial expression probability through a full connection layer and softmax.
The method is implemented by using a public data set RAF-DB, a training set and a test set of the RAF-DB are preprocessed according to the preprocessing mode, then a local feature and global feature fusion model is constructed, the training set training model is adopted, finally, the test set is only used for testing, and finally, the expression recognition accuracy rate obtained by the model is obtained.
The effect of the method of the invention on the RAF-DB data set is compared with that of other methods through the following table, and the comparison shows that the method of the invention has better identification effect compared with other methods.
Method of producing a composite material | Year of year | Rate of accuracy |
DLP-CNN(baseline) | 2019 | 84.13 |
gACNN | 2018 | 85.07 |
The method of the invention | 2020 | 85.75 |
The facial expression recognition method mainly comprises the steps of extracting similar features of the same type of expressions and constructing a model for distinguishing different expressions according to different features of different types of expressions. By combining related theories of psychology, graphics and information science, practical and valuable references are provided for human emotion calculation and human-computer interaction system research. With the development of artificial intelligence technology, the method is widely applied to the fields of psychological disease treatment, emotional interaction and the like. However, at present, the expression recognition in a natural scene is a problem with a larger challenge, in the embodiment, the disclosed expression data set in the natural scene is preprocessed, the local features and the global features are extracted, and finally the global features and the local features are fused to obtain a model for the expression recognition, so that the accurate recognition of the static facial expression is realized.
Example two
In order to implement the method for recognizing static facial expression in natural scene according to the first embodiment, the present embodiment provides a device for recognizing static facial expression in natural scene, including:
the preprocessing module is used for removing the background of a natural scene in the picture and only reserving a face area;
the global feature extraction module is used for extracting the features of the whole picture processed by the preprocessing module by adopting a convolutional neural network and converting the features into feature vectors;
the local feature extraction module is used for extracting a local area with the most information quantity of the picture by a target detection method and converting the local area into a feature vector;
and the feature fusion module is used for integrating the feature vectors extracted by the global feature extraction module and the feature vectors extracted by the local feature extraction module into a feature matrix, and then obtaining the facial expression probability through a full connection layer and softmax.
EXAMPLE III
The present embodiment is a storage medium storing a computer program, which when executed by a processor causes the processor to execute the method for recognizing a static facial expression in a natural scene according to the first embodiment.
Example four
The embodiment of the invention relates to a computing device, which comprises a processor and a memory for storing a program executable by the processor, and is characterized in that when the processor executes the program stored in the memory, the method for recognizing the static facial expression in the natural scene, as described in the first embodiment, is implemented.
The above embodiments are preferred embodiments of the present invention, but the present invention is not limited to the above embodiments, and any other changes, modifications, substitutions, combinations, and simplifications which do not depart from the spirit and principle of the present invention should be construed as equivalents thereof, and all such changes, modifications, substitutions, combinations, and simplifications are intended to be included in the scope of the present invention.
Claims (8)
1. A method for recognizing static facial expressions in natural scenes is characterized by comprising the following steps: the method comprises a pretreatment process, a global feature extraction process, a local feature extraction process and a feature fusion process which are sequentially executed;
the preprocessing procedure is used for removing the background of a natural scene in the picture and only reserving a face area;
the global feature extraction process is to extract the features of the whole picture after the preprocessing process by adopting a convolutional neural network and convert the features into feature vectors;
the local feature extraction process is to extract a local area with the most information quantity of the picture by a target detection method and convert the local area into a feature vector;
the feature fusion process is to integrate the feature vectors extracted by the global feature extraction process and the feature vectors extracted by the local feature extraction process into a feature matrix, and then obtain the facial expression probability through a full connection layer and softmax.
2. The method of recognizing static facial expressions in natural scenes according to claim 1, characterized in that: the preprocessing procedure refers to the operations of face detection, face alignment and face cutting of the picture.
3. The method of recognizing static facial expressions in natural scenes according to claim 1, characterized in that: the local feature extraction process comprises the following sub-steps:
s1, generating a plurality of target detection frames by adopting a target detection method;
s2, converting the local area feature map of the target detection box in S1 into scores of corresponding expression classes by using a convolutional neural network, and formulating as follows:
S:A→[0,1]
wherein S represents a score set of local regions, A is a set of all local regions, and R ∈ A, the score of the local region is set as C (R), the information content of the local region is set as I (R), and the information content of all local regions is ordered:
R1,R2∈A,ifC(R1)>C(R2),I(R1)>I(R2)
s3, selecting the first K local areas with the largest information quantity;
s4, extracting K local area features screened in S3 by adopting a convolutional neural network, and converting the K local area features into feature vectors.
4. A method of recognizing static facial expressions in natural scenes according to claim 3, characterized in that: in steps S2 and S4 of the global feature extraction step and the local feature extraction step, the convolutional neural network uses resnet50 as a base network.
5. The method of claim 3, wherein in step S4 of the global feature extraction step and the local feature extraction step, the feature vectors are all M × 1 vectors;
in the feature fusion step, the size of the feature matrix is M × (K + 1).
6. An apparatus for recognizing facial expressions based on static states in a natural scene, comprising:
the preprocessing module is used for removing the background of a natural scene in the picture and only reserving a face area;
the global feature extraction module is used for extracting the features of the whole picture processed by the preprocessing module by adopting a convolutional neural network and converting the features into feature vectors;
the local feature extraction module is used for extracting a local area with the most information quantity of the picture by a target detection method and converting the local area into a feature vector;
and the feature fusion module is used for integrating the feature vectors extracted by the global feature extraction module and the feature vectors extracted by the local feature extraction module into a feature matrix, and then obtaining the facial expression probability through a full connection layer and softmax.
7. A storage medium, wherein the storage medium stores a computer program which, when executed by a processor, causes the processor to execute the method of static facial expression recognition in natural scenes according to any one of claims 1 to 5.
8. A computing device comprising a processor and a memory for storing a processor-executable program, wherein the processor, when executing the program stored in the memory, implements the method of static facial expression recognition in natural scenes according to any one of claims 1 to 5.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010115562.6A CN111368663B (en) | 2020-02-25 | 2020-02-25 | Method, device, medium and equipment for recognizing static facial expression in natural scene |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010115562.6A CN111368663B (en) | 2020-02-25 | 2020-02-25 | Method, device, medium and equipment for recognizing static facial expression in natural scene |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111368663A true CN111368663A (en) | 2020-07-03 |
CN111368663B CN111368663B (en) | 2024-02-20 |
Family
ID=71206437
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010115562.6A Active CN111368663B (en) | 2020-02-25 | 2020-02-25 | Method, device, medium and equipment for recognizing static facial expression in natural scene |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111368663B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111832669A (en) * | 2020-09-21 | 2020-10-27 | 首都师范大学 | Method and device for establishing learning participation degree recognition network model |
CN113591718A (en) * | 2021-07-30 | 2021-11-02 | 北京百度网讯科技有限公司 | Target object identification method and device, electronic equipment and storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102902980A (en) * | 2012-09-13 | 2013-01-30 | 中国科学院自动化研究所 | Linear programming model based method for analyzing and identifying biological characteristic images |
CN107729835A (en) * | 2017-10-10 | 2018-02-23 | 浙江大学 | A kind of expression recognition method based on face key point region traditional characteristic and face global depth Fusion Features |
CN108491835A (en) * | 2018-06-12 | 2018-09-04 | 常州大学 | Binary channels convolutional neural networks towards human facial expression recognition |
CN110580461A (en) * | 2019-08-29 | 2019-12-17 | 桂林电子科技大学 | Facial expression recognition algorithm combined with multilevel convolution characteristic pyramid |
-
2020
- 2020-02-25 CN CN202010115562.6A patent/CN111368663B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102902980A (en) * | 2012-09-13 | 2013-01-30 | 中国科学院自动化研究所 | Linear programming model based method for analyzing and identifying biological characteristic images |
CN107729835A (en) * | 2017-10-10 | 2018-02-23 | 浙江大学 | A kind of expression recognition method based on face key point region traditional characteristic and face global depth Fusion Features |
CN108491835A (en) * | 2018-06-12 | 2018-09-04 | 常州大学 | Binary channels convolutional neural networks towards human facial expression recognition |
CN110580461A (en) * | 2019-08-29 | 2019-12-17 | 桂林电子科技大学 | Facial expression recognition algorithm combined with multilevel convolution characteristic pyramid |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111832669A (en) * | 2020-09-21 | 2020-10-27 | 首都师范大学 | Method and device for establishing learning participation degree recognition network model |
CN113591718A (en) * | 2021-07-30 | 2021-11-02 | 北京百度网讯科技有限公司 | Target object identification method and device, electronic equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN111368663B (en) | 2024-02-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US12039454B2 (en) | Microexpression-based image recognition method and apparatus, and related device | |
US20220189142A1 (en) | Ai-based object classification method and apparatus, and medical imaging device and storage medium | |
CN110532900B (en) | Facial expression recognition method based on U-Net and LS-CNN | |
Yu et al. | Image quality classification for DR screening using deep learning | |
CN112132197B (en) | Model training, image processing method, device, computer equipment and storage medium | |
WO2020133636A1 (en) | Method and system for intelligent envelope detection and warning in prostate surgery | |
CN109034210A (en) | Object detection method based on super Fusion Features Yu multi-Scale Pyramid network | |
CN110464366A (en) | A kind of Emotion identification method, system and storage medium | |
CN109920538B (en) | Zero sample learning method based on data enhancement | |
CN113486700A (en) | Facial expression analysis method based on attention mechanism in teaching scene | |
CN108596256A (en) | One kind being based on RGB-D object identification grader building methods | |
CN117935339A (en) | Micro-expression recognition method based on multi-modal fusion | |
Mamdouh et al. | A New Model for Image Segmentation Based on Deep Learning. | |
CN111368663A (en) | Method, device, medium and equipment for recognizing static facial expressions in natural scene | |
CN114550270A (en) | Micro-expression identification method based on double-attention machine system | |
He et al. | Dual multi-task network with bridge-temporal-attention for student emotion recognition via classroom video | |
Lu et al. | Image recognition algorithm based on improved AlexNet and shared parameter transfer learning | |
CN112560668A (en) | Human behavior identification method based on scene prior knowledge | |
CN117633558A (en) | Multi-excitation fusion zero-sample lesion detection method based on visual language model | |
CN110674675A (en) | Pedestrian face anti-fraud method | |
CN115719497A (en) | Student concentration degree identification method and system | |
Boroujerdi et al. | Deep interactive region segmentation and captioning | |
CN115588217A (en) | Face attribute detection method based on deep self-attention network | |
Park et al. | Ensuring Visual Commonsense Morality for Text-to-Image Generation | |
CN112906539B (en) | Object identification method based on EEG data |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |