Nothing Special   »   [go: up one dir, main page]

CN110580461A - Facial expression recognition algorithm combined with multilevel convolution characteristic pyramid - Google Patents

Facial expression recognition algorithm combined with multilevel convolution characteristic pyramid Download PDF

Info

Publication number
CN110580461A
CN110580461A CN201910806700.2A CN201910806700A CN110580461A CN 110580461 A CN110580461 A CN 110580461A CN 201910806700 A CN201910806700 A CN 201910806700A CN 110580461 A CN110580461 A CN 110580461A
Authority
CN
China
Prior art keywords
network
feature
facial expression
feature extraction
level
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910806700.2A
Other languages
Chinese (zh)
Inventor
符强
昌涛
孙希延
纪元法
任风华
严素清
付文涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guilin University of Electronic Technology
Original Assignee
Guilin University of Electronic Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guilin University of Electronic Technology filed Critical Guilin University of Electronic Technology
Priority to CN201910806700.2A priority Critical patent/CN110580461A/en
Publication of CN110580461A publication Critical patent/CN110580461A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a facial expression recognition algorithm combined with a multilevel convolution characteristic pyramid, which comprises the following steps: inputting the original facial expression picture into a first feature extraction network for feature extraction, and extracting global features; inputting the cut and amplified facial expression picture into a second feature extraction network for feature extraction, and extracting local features; constructing a feature pyramid network in the first feature extraction network and the second feature extraction network; carrying out regional positioning on the local features in the second feature extraction network by using an attention regional positioning network; and fusing the global features and the local features of the human face expression pictures by using the feature fusion network, and classifying the human face expression pictures through a full connection layer. The aim of improving the accuracy of the facial expression recognition technology on the facial expression recognition is achieved.

Description

Facial expression recognition algorithm combined with multilevel convolution characteristic pyramid
Technical Field
the invention relates to the field of image processing, in particular to a facial expression recognition algorithm combined with a multilevel convolution characteristic pyramid.
Background
The human face expression is not only a kind of emotion of people, but also can transmit abundant human behavior information. The facial expression recognition technology is a main form for researching human emotion, and the accurate recognition of the facial expression can effectively judge the emotional state of a human. The facial expression recognition can be applied to many fields, such as the field of safe driving, the field of intelligent human-computer interaction, the field of medical monitoring, lie detection in criminal investigation, psychotherapy and the like. However, the accuracy of the existing facial expression recognition technology for facial expression recognition is low.
Disclosure of Invention
the invention aims to provide a facial expression recognition algorithm combined with a multilevel convolution characteristic pyramid, and aims to solve the technical problem that the facial expression recognition technology in the prior art is low in accuracy rate of facial expression recognition.
In order to achieve the above object, the facial expression recognition algorithm of the invention, which combines a multilevel convolution feature pyramid, comprises the following steps:
Inputting the original facial expression picture into a first feature extraction network for feature extraction, and extracting global features;
inputting the cut and amplified facial expression picture into a second feature extraction network for feature extraction, and extracting local features;
Constructing a feature pyramid network in the first feature extraction network and the second feature extraction network;
carrying out regional positioning on the local features in the second feature extraction network by using an attention regional positioning network;
and fusing the global features and the local features of the human face expression pictures by using the feature fusion network, and classifying the human face expression pictures through a full connection layer.
The input sample of the second-level feature extraction network is obtained by cutting and amplifying an input sample picture in the first-level feature extraction network.
The first-level feature extraction network and the second-level feature extraction network are both full convolution structure networks.
Wherein, the convolution kernel with the size of 3 x 3 is used in the first convolution layer of the global feature extraction and the local feature extraction.
the method comprises the following steps of inputting an original facial expression picture into a first-level feature extraction network, extracting features of the facial expression picture through a convolutional neural network, and then outputting categories and probability values of each category, wherein the probability value of each category is obtained by the following method:
C(i)=f(Wi⊙Xi)
p(i)=f(Wi⊙Xi)
Wherein, the input facial expression picture is X, Xiand Wirepresents the input and weight parameters of the i-th convolutional neural network, indicates the convolution, pooling, activation, or normalization operation performed during convolution, f (-) indicates the feature extraction network, C(i)(i ═ 1,2) denotes an output class label of the i-th convolutional neural network, p(i)(i ═ 1,2) denotes the probability value of the i-th order convolutional neural network for outputting each class.
Wherein, each level of network generates a feature map about the input facial expression picture for userepresenting a set of generated feature maps of each stage of the network, wherein m represents the number of network stages contained in the model, wherein m is equal to (1,2), and n represents the final output of each stage of the networkthe input image X is represented by the following sub-formula:
the feature graph generated by the convolutional layer at the later layer is transposed and convolved, and then the two feature graphs are added to unify the scales of the feature graphs.
Wherein the attention area positioning network maps the input feature map into (D)x,Dy) Is a center, DαThe attention area is a square attention area with side length, then the attention area is mapped to an original facial expression picture, and the original facial expression picture is cut and amplified and then sent to a next-level network.
the second feature extraction network can automatically activate the salient region of the facial expression during recognition and extraction of the facial expression, and firstly, accurate positioning must be realized on the region with the largest response in the feature map, and then the accurately positioned regions are mapped to the original facial expression picture to obtain the important local region of the original facial expression, and feature extraction is carried out on the local region to realize local region recognition of the picture.
When initializing, all feature maps generated by the last layer of the first-level feature extraction network are added together, then an area with a larger numerical value in the added feature maps is positioned, the positioned area is fitted into a square, then the central coordinate and the side length of the square are obtained and set as initialization parameters of the attention network, and the initialization parameters are expressed as:
Wherein F represents a feature map generated by the last convolutional layer of the first feature extraction network, d represents the total number of feature maps generated by the network, F is the total feature map obtained by adding each corresponding point of each feature map, h and w represent the height and width of the feature map, and p represents the pixel mean value of the total feature map.
the invention relates to a facial expression recognition algorithm combined with a multilevel convolution feature pyramid, which is characterized in that an original facial expression picture is input into a first feature extraction network for feature extraction, and global features are extracted; inputting the cut and amplified facial expression picture into a second feature extraction network for feature extraction, and extracting local features; constructing a feature pyramid network in the first feature extraction network and the second feature extraction network; carrying out regional positioning on the local features in the second feature extraction network by using an attention regional positioning network; and fusing the global features and the local features of the human face expression pictures by using the feature fusion network, and classifying the human face expression pictures through a full connection layer. The human face features are transferred from the whole to the local by adopting a cascade structure of two levels of human face feature extraction networks, and a feature pyramid network is added into each level of human face feature extraction network, so that the robustness of the algorithm to the size of the human face target is improved. Therefore, the effect of improving the accuracy of the facial expression recognition technology on the facial expression recognition is obtained.
drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a step diagram of the facial expression recognition algorithm of the present invention incorporating a multi-level convolution feature pyramid.
FIG. 2 is a flow chart of the facial expression recognition algorithm of the present invention incorporating a multi-level convolution feature pyramid.
Detailed Description
Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the drawings are illustrative and intended to be illustrative of the invention and are not to be construed as limiting the invention.
in the description of the present invention, it is to be understood that the terms "length", "width", "upper", "lower", "front", "rear", "left", "right", "vertical", "horizontal", "top", "bottom", "inner", "outer", and the like, indicate orientations or positional relationships based on the orientations or positional relationships illustrated in the drawings, and are used merely for convenience in describing the present invention and for simplicity in description, and do not indicate or imply that the devices or elements referred to must have a particular orientation, be constructed in a particular orientation, and be operated, and thus, are not to be construed as limiting the present invention. Further, in the description of the present invention, "a plurality" means two or more unless specifically defined otherwise.
referring to fig. 1 and fig. 2, the present invention provides a facial expression recognition algorithm combining a multilevel convolution feature pyramid, which includes the following steps:
S100: inputting the original facial expression picture into a first feature extraction network for feature extraction, and extracting global features;
S200: inputting the cut and amplified facial expression picture into a second feature extraction network for feature extraction, and extracting local features;
s300: constructing a feature pyramid network in the first feature extraction network and the second feature extraction network;
S400: carrying out regional positioning on the local features in the second feature extraction network by using an attention regional positioning network;
s500: and fusing the global features and the local features of the human face expression pictures by using the feature fusion network, and classifying the human face expression pictures through a full connection layer.
In this embodiment, the first-level feature extraction network can perform feature extraction on 448 × 448 pixels of an original facial expression picture, and extract global features; the second-level feature extraction network is used for extracting features of the cut and amplified human face expression picture and extracting local features; the input sample of the second-level feature extraction network is obtained by cutting and amplifying an input sample picture in the first-level feature extraction network, and the first feature extraction network and the second feature extraction network are both full convolution structure networks. The structure of the network is similar to that of the VGG19 network, each module of the network is composed of a convolution layer, a BatchNorm layer, a relu layer and an average pooling layer, and 16 convolution modules are provided. The first convolutional layer in both global feature extraction and local feature extraction uses a convolution kernel of size 3 x 3.
inputting the original facial expression picture into a first-level feature extraction network, extracting the features of the facial expression picture through a convolutional neural network, and then outputting the category and the probability value of each category, which are expressed by the following sub-expressions:
C(i)=f(Wi⊙Xi)
p(i)=f(Wi⊙Xi)
In the above formula, the input facial expression picture is X, XiAnd Wirepresents the input and weight parameters of the i-th convolutional neural network, indicates the convolution, pooling, activation, or normalization operation performed during convolution, f (-) indicates the feature extraction network, C(i)(i ═ 1,2) denotes an output class label of the i-th convolutional neural network, p(i)(i ═ 1,2) denotes the probability value of the i-th order convolutional neural network for outputting each class.
Each level of network generates a feature map about the input facial expression pictureRepresenting that each stage of network generates a set of feature maps, m represents the number of network stages contained in the model, wherein m is (1,2), n represents the number of the feature maps finally output by each stage of network, and the input image X is represented by the following formula:
The convolutional neural network extracts features of an input picture layer by layer through convolutional cores of each layer, feature graphs of different convolutional layers correspond to different regions and targets of the input picture, in order to solve the problem of face size diversification, and in order to better utilize shallow convolutional layers of the convolutional neural network to extract face expression information, a feature pyramid network is constructed among convolutional layers of the convolutional neural network, the feature graph scale generated by the shallow convolutional layers in the convolutional neural network is large, the feature graph scale generated by the convolutional layers at the body of the convolutional neural network is small, in order to unify the scale of the feature graphs, firstly, feature graphs generated by the convolutional layers of the next layer are subjected to transposition convolution operation, then, the two feature graphs are added, and the generated feature graph is the final representation of the whole face expression picture.
in order to realize transition and transfer of facial expressions from global features to local features, the attention area positioning network is arranged between feature extraction networks and is responsible for positioning local feature areas (such as eyebrows, eyes, a nose and a mouth) in an original facial expression picture, cutting the corresponding areas and sending the areas to a next-level network. The trained facial expression recognition network can automatically activate the salient region of the facial expression (namely, the region with important influence on the facial expression recognition effect) when the facial expression is recognized, so that the region with the largest response in the feature map must be accurately positioned, and then the accurately positioned regions are mapped to the original facial expression picture, so that the very important local region of the original facial expression can be obtained, and the local region is subjected to feature extraction to realize the local region recognition of the picture. The regions having important influence on facial expression recognition are automatically located by adopting the step-by-step location attention network.
the attention area positioning network maps the input feature map into (D)x,Dy) Is a center, DaThe attention area is a square attention area with side length, the attention area is mapped to an original facial expression picture, and the original facial expression picture is cut and amplified and then sent to a next-level network, so that the facial expression picture can be finely identified from a global feature picture to a local feature area.
the key point of mapping the strong response area in the feature map obtained through the convolutional neural network to the original facial expression picture is to obtain the coordinates of the strong response area in the feature map, and then the feature map can be automatically positioned through the full-connection layer network by designing an attention area positioning network of a 3-layer full-connection neural network. In order to accelerate the convergence speed of the attention area positioning network, all feature maps generated by the last layer of the first-stage feature extraction network are added together during initialization, then an area with a larger numerical value in the added feature maps is positioned, the positioned area is fitted into a square, then the central coordinate and the side length of the square are obtained and set as initialization parameters of the attention area positioning network, and the specific implementation mode is as follows:
in the above formula, F represents a feature map generated by the last convolutional layer of the first feature extraction network, d represents the total number of feature maps generated by the network, F is the total feature map obtained by adding each corresponding point of each feature map, h and w represent the height and width of the feature map, p represents the pixel mean value of the total feature map, then each pixel point of the total feature map is compared with the pixel mean value, which is equivalent to the threshold value taking the pixel mean value as the total feature map, the pixel points larger than the threshold value are set to be 1, the pixel points smaller than or equal to the threshold value are set to be 0, then the longest edge of the maximum connected region set to be 1 is selected to be the edge length of the square, and finally the central coordinate and the edge length of the square are initialized to the parameters of the attention localization network. The attention area positioning network can automatically position the area with the largest response in the feature map, and then cuts and enlarges the attention area after positioning, wherein the attention area is supposed to be gradually increased from left to right and from bottom to top by taking the lower left corner of the picture as the origin and the horizontal direction as the x axis. According to the obtained coordinates of the center point of the square attention area and the side length of the square attention area, four vertexes of the square attention area can be calculated, so that automatic clipping of the attention area is achieved, and finally the attention area picture is amplified by utilizing a bilinear interpolation value. To obtain the picture we need.
Each level of feature extraction network of the algorithm can extract features of the facial expression picture, the first feature extraction network extracts global features of the facial expression picture, the second network extracts local detail features of the facial expression picture, pixel points corresponding to feature pictures of each channel are added, feature dimension reduction is conducted through convolution with the size of 1 x 1, then feature fusion and pooling are conducted, and finally category labels of the facial expression are output through a full connection layer.
After the network is designed according to the method, in order to simultaneously optimize the feature extraction network and the attention area positioning network, two loss functions are adopted to optimize the network, the weighting sum of the cross entropy loss functions of the two networks and the cross entropy loss function of the fusion network is used as the integral loss function of the network, so that the effect of distinguishing different facial expressions is good, and meanwhile, in order to restrict the difference of similar facial expressions, the punishment verification loss of the two networks of the feature extraction network and the attention area positioning network is adopted, so that the recognition performance between similar facial expressions is greatly enhanced.
In summary, the following steps: the algorithm can automatically position the area with larger distinguishing degree of the facial expression picture, and the global and local characteristics of the facial expression picture are fused, so that the accurate recognition of the facial expression picture is realized. The first feature extraction network and the second feature extraction network are responsible for carrying out feature extraction on the facial expression pictures, the attention area positioning network is responsible for positioning the local feature areas of the facial expression pictures, expressing and positioning the local feature areas of the facial expression pictures, and for the facial expression pictures with high similarity, such as anger and disgust, the algorithm realizes the area positioning of the local features of the facial expression with high discrimination, and comprehensively utilizes the superficial layer high-resolution face structure information and the deep layer low-resolution semantic information of the convolutional neural network to obtain better positioning of the local feature areas of the facial expression, so that the recognition rate of the facial expression pictures with high similarity is greatly improved.
The feature fusion network is responsible for fusing the global features and the local features of the facial expressions and fusing the global feature information of the facial expression pictures and the local feature information of the facial expression pictures. The local feature information of the facial expression pictures can improve the distinguishing capability of facial expressions with larger similarity, and meanwhile, the global structure information of the facial expression pictures is kept, so that the features extracted by all levels of facial expression feature extraction networks are fused by adopting a feature fusion network, more comprehensive feature information of the facial expression pictures is obtained, and the accuracy of facial expression recognition is improved. Because the single-pole facial expression feature extraction network cannot simultaneously obtain the global and local features of the facial expression pictures, the two-stage facial expression feature extraction network cascade structure is adopted to realize the transfer of the features from the global to the local, and a feature pyramid network is constructed between the feature images of each stage of facial expression feature extraction network aiming at the characteristic that the facial expression pictures have scale changes, so that the feature description capacity of the network is improved, and the identification accuracy of the facial expression pictures with scale changes, the facial expression pictures with shelters and the facial expression pictures with similarity is greatly improved.
While the invention has been described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (10)

1. A facial expression recognition algorithm combined with a multilevel convolution feature pyramid is characterized by comprising the following steps:
Inputting the original facial expression picture into a first feature extraction network for feature extraction, and extracting global features;
Inputting the cut and amplified facial expression picture into a second feature extraction network for feature extraction, and extracting local features;
Constructing a feature pyramid network in the first feature extraction network and the second feature extraction network;
Carrying out regional positioning on the local features in the second feature extraction network by using an attention regional positioning network;
And fusing the global features and the local features of the human face expression pictures by using the feature fusion network, and classifying the human face expression pictures through a full connection layer.
2. The algorithm for facial expression recognition in combination with the multi-level pyramid of convolution features of claim 1,
The input sample of the second-level feature extraction network is obtained by cutting and amplifying an input sample picture in the first-level feature extraction network.
3. The algorithm for facial expression recognition in combination with the multi-level pyramid of convolution features of claim 1,
The first level feature extraction network and the second level feature extraction network are both full convolution structure networks.
4. The algorithm for facial expression recognition in combination with the multi-level pyramid of convolution features of claim 3,
The first convolutional layer in both global feature extraction and local feature extraction uses a convolution kernel of size 3 x 3.
5. The algorithm for facial expression recognition in combination with the pyramid of multilevel convolution features of claim 4,
Inputting an original facial expression picture into a first-level feature extraction network, extracting features of the facial expression picture through a convolutional neural network, and then outputting categories and probability values of each category, wherein the probability value of each category is obtained by the following steps:
C(i)=f(Wi⊙Xi)
p(i)=f(Wi⊙Xi)
Wherein, the input facial expression picture is X, XiAnd Wirepresents the input and weight parameters of the i-th convolutional neural network, indicates the convolution, pooling, activation, or normalization operation performed during convolution, f (-) indicates the feature extraction network, C(i)(i ═ 1,2) denotes an output class label of the i-th convolutional neural network, p(i)(i ═ 1,2) denotes the probability value of the i-th order convolutional neural network for outputting each class.
6. The algorithm for facial expression recognition in combination with the pyramid of multilevel convolution features of claim 5,
Each level of network generates a feature map about the input facial expression picturerepresenting a set of generated feature maps of each level of network, wherein m represents the number of network levels contained in the model, wherein m is equal to (1,2), n represents the number of the feature maps finally output by each level of network, and an input image X is represented as:
7. The algorithm for facial expression recognition in combination with the multi-level pyramid of convolution features of claim 6,
The feature graph generated by the shallow convolutional layer in the convolutional neural network is large in scale, the feature graph generated by the deep convolutional layer in the convolutional neural network is small in scale, the feature graph generated by the next convolutional layer is transposed and convolved, and then the two feature graphs are added to unify the scales of the feature graphs.
8. the algorithm for facial expression recognition in combination with the multi-level pyramid of convolution features of claim 7,
the attention area positioning network maps the input feature map into (D)x,Dy) Is a center, Dαthe attention area is a square attention area with side length, then the attention area is mapped to an original facial expression picture, and the original facial expression picture is cut and amplified and then sent to a next-level network.
9. The algorithm for facial expression recognition in combination with the pyramid of multilevel convolution features of claim 8,
The second feature extraction network can automatically activate the salient regions of the facial expressions in the process of identifying and extracting the facial expressions, firstly, accurate positioning must be achieved on the regions with the largest response in the feature images, then the accurately positioned regions are mapped to the original facial expression images, important local regions of the original facial expressions are obtained, and feature extraction is conducted on the local regions to achieve local region identification of the images.
10. The algorithm for facial expression recognition in combination with the multi-level pyramid of convolution features of claim 9,
When initializing, all feature maps generated by the last layer of the first-level feature extraction network are added together, then an area with a larger numerical value in the added feature maps is positioned, the positioned area is fitted into a square, then the central coordinate and the side length of the square are obtained and set as initialization parameters of the attention network, and the initialization parameters are expressed as:
Wherein F represents a feature map generated by the last convolutional layer of the first feature extraction network, d represents the total number of feature maps generated by the network, F is the total feature map obtained by adding each corresponding point of each feature map, h and w represent the height and width of the feature map, and p represents the pixel mean value of the total feature map.
CN201910806700.2A 2019-08-29 2019-08-29 Facial expression recognition algorithm combined with multilevel convolution characteristic pyramid Pending CN110580461A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910806700.2A CN110580461A (en) 2019-08-29 2019-08-29 Facial expression recognition algorithm combined with multilevel convolution characteristic pyramid

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910806700.2A CN110580461A (en) 2019-08-29 2019-08-29 Facial expression recognition algorithm combined with multilevel convolution characteristic pyramid

Publications (1)

Publication Number Publication Date
CN110580461A true CN110580461A (en) 2019-12-17

Family

ID=68812404

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910806700.2A Pending CN110580461A (en) 2019-08-29 2019-08-29 Facial expression recognition algorithm combined with multilevel convolution characteristic pyramid

Country Status (1)

Country Link
CN (1) CN110580461A (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111368663A (en) * 2020-02-25 2020-07-03 华南理工大学 Method, device, medium and equipment for recognizing static facial expressions in natural scene
CN112036260A (en) * 2020-08-10 2020-12-04 武汉星未来教育科技有限公司 Expression recognition method and system for multi-scale sub-block aggregation in natural environment
CN112200065A (en) * 2020-10-09 2021-01-08 福州大学 Micro-expression classification method based on action amplification and self-adaptive attention area selection
CN112364827A (en) * 2020-11-30 2021-02-12 腾讯科技(深圳)有限公司 Face recognition method and device, computer equipment and storage medium
CN112464865A (en) * 2020-12-08 2021-03-09 北京理工大学 Facial expression recognition method based on pixel and geometric mixed features
CN112800914A (en) * 2021-01-20 2021-05-14 桂林电子科技大学 Face recognition method based on expert prior knowledge and LTP complex illumination
CN113361493A (en) * 2021-07-21 2021-09-07 天津大学 Facial expression recognition method for robustness of different image resolutions
CN113569732A (en) * 2021-07-27 2021-10-29 厦门理工学院 Face attribute recognition method and system based on parallel sharing multitask network
CN115205986A (en) * 2022-08-09 2022-10-18 山东省人工智能研究院 False video detection method based on knowledge distillation and transformer
WO2023029678A1 (en) * 2022-04-06 2023-03-09 江苏商贸职业学院 Gis-based agricultural service management method and system
CN115909455A (en) * 2022-11-16 2023-04-04 航天恒星科技有限公司 Expression recognition method integrating multi-scale feature extraction and attention mechanism
WO2023173646A1 (en) * 2022-03-17 2023-09-21 深圳须弥云图空间科技有限公司 Expression recognition method and apparatus
WO2024001095A1 (en) * 2022-06-27 2024-01-04 闻泰通讯股份有限公司 Facial expression recognition method, terminal device and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109409222A (en) * 2018-09-20 2019-03-01 中国地质大学(武汉) A kind of multi-angle of view facial expression recognizing method based on mobile terminal
CN109815924A (en) * 2019-01-29 2019-05-28 成都旷视金智科技有限公司 Expression recognition method, apparatus and system

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109409222A (en) * 2018-09-20 2019-03-01 中国地质大学(武汉) A kind of multi-angle of view facial expression recognizing method based on mobile terminal
CN109815924A (en) * 2019-01-29 2019-05-28 成都旷视金智科技有限公司 Expression recognition method, apparatus and system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
梁华刚等: "多级卷积特征金字塔的细粒度食物图片识别", 《中国图象图形学报》 *

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111368663B (en) * 2020-02-25 2024-02-20 华南理工大学 Method, device, medium and equipment for recognizing static facial expression in natural scene
CN111368663A (en) * 2020-02-25 2020-07-03 华南理工大学 Method, device, medium and equipment for recognizing static facial expressions in natural scene
CN112036260A (en) * 2020-08-10 2020-12-04 武汉星未来教育科技有限公司 Expression recognition method and system for multi-scale sub-block aggregation in natural environment
CN112200065B (en) * 2020-10-09 2022-08-09 福州大学 Micro-expression classification method based on action amplification and self-adaptive attention area selection
CN112200065A (en) * 2020-10-09 2021-01-08 福州大学 Micro-expression classification method based on action amplification and self-adaptive attention area selection
CN112364827A (en) * 2020-11-30 2021-02-12 腾讯科技(深圳)有限公司 Face recognition method and device, computer equipment and storage medium
CN112364827B (en) * 2020-11-30 2023-11-10 腾讯科技(深圳)有限公司 Face recognition method, device, computer equipment and storage medium
CN112464865A (en) * 2020-12-08 2021-03-09 北京理工大学 Facial expression recognition method based on pixel and geometric mixed features
CN112800914A (en) * 2021-01-20 2021-05-14 桂林电子科技大学 Face recognition method based on expert prior knowledge and LTP complex illumination
CN113361493B (en) * 2021-07-21 2022-05-20 天津大学 Facial expression recognition method robust to different image resolutions
CN113361493A (en) * 2021-07-21 2021-09-07 天津大学 Facial expression recognition method for robustness of different image resolutions
CN113569732A (en) * 2021-07-27 2021-10-29 厦门理工学院 Face attribute recognition method and system based on parallel sharing multitask network
CN113569732B (en) * 2021-07-27 2023-06-06 厦门理工学院 Face attribute identification method and system based on parallel sharing multitasking network
WO2023173646A1 (en) * 2022-03-17 2023-09-21 深圳须弥云图空间科技有限公司 Expression recognition method and apparatus
WO2023029678A1 (en) * 2022-04-06 2023-03-09 江苏商贸职业学院 Gis-based agricultural service management method and system
WO2024001095A1 (en) * 2022-06-27 2024-01-04 闻泰通讯股份有限公司 Facial expression recognition method, terminal device and storage medium
CN115205986A (en) * 2022-08-09 2022-10-18 山东省人工智能研究院 False video detection method based on knowledge distillation and transformer
CN115205986B (en) * 2022-08-09 2023-05-19 山东省人工智能研究院 Dummy video detection method based on knowledge distillation and transformation
CN115909455B (en) * 2022-11-16 2023-09-19 航天恒星科技有限公司 Expression recognition method integrating multi-scale feature extraction and attention mechanism
CN115909455A (en) * 2022-11-16 2023-04-04 航天恒星科技有限公司 Expression recognition method integrating multi-scale feature extraction and attention mechanism

Similar Documents

Publication Publication Date Title
CN110580461A (en) Facial expression recognition algorithm combined with multilevel convolution characteristic pyramid
CN111340814B (en) RGB-D image semantic segmentation method based on multi-mode self-adaptive convolution
Yang et al. Towards real-time traffic sign detection and classification
CN111639544B (en) Expression recognition method based on multi-branch cross-connection convolutional neural network
CN105528575B (en) Sky detection method based on Context Reasoning
CN112132197B (en) Model training, image processing method, device, computer equipment and storage medium
CN112801015B (en) Multi-mode face recognition method based on attention mechanism
CN111639564B (en) Video pedestrian re-identification method based on multi-attention heterogeneous network
CN112766229B (en) Human face point cloud image intelligent identification system and method based on attention mechanism
CN112016601B (en) Network model construction method based on knowledge graph enhanced small sample visual classification
Yang et al. Facial expression recognition based on dual-feature fusion and improved random forest classifier
CN109165658B (en) Strong negative sample underwater target detection method based on fast-RCNN
CN115019274B (en) Road surface disease identification method integrating tracking and searching algorithms
CN112836625A (en) Face living body detection method and device and electronic equipment
CN112396036B (en) Method for re-identifying blocked pedestrians by combining space transformation network and multi-scale feature extraction
CN113378675A (en) Face recognition method for simultaneous detection and feature extraction
CN111368775A (en) Complex scene dense target detection method based on local context sensing
CN113011253A (en) Face expression recognition method, device, equipment and storage medium based on ResNeXt network
CN115880397A (en) Topological semantic map construction method for blind assisting scene
CN110414431B (en) Face recognition method and system based on elastic context relation loss function
CN114529842A (en) Figure interaction detection method based on two-way attention mechanism under knowledge guidance
CN111881803B (en) Face recognition method based on improved YOLOv3
CN116665265A (en) Facial living body detection method and system based on facial feature segmentation
CN116311518A (en) Hierarchical character interaction detection method based on human interaction intention information
CN111898473B (en) Driver state real-time monitoring method based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20191217