CN111488856B - Multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning - Google Patents
Multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning Download PDFInfo
- Publication number
- CN111488856B CN111488856B CN202010347655.1A CN202010347655A CN111488856B CN 111488856 B CN111488856 B CN 111488856B CN 202010347655 A CN202010347655 A CN 202010347655A CN 111488856 B CN111488856 B CN 111488856B
- Authority
- CN
- China
- Prior art keywords
- map
- orthogonal
- feature
- expression recognition
- method based
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
A multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning relates to the technical field of computer vision. The method utilizes face point cloud data to generate three attribute maps which are respectively a depth map, a direction map and an elevation map, wherein the depth map, the direction map and the elevation map are combined into a three-channel RGB map, and the RGB map is used as the input of a certain branch in a network, so that the parameter quantity of a model is reduced. By the multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning, the complexity of a deep learning network is reduced, the redundancy among features extracted from different branches in the network is inhibited, and good economic and social benefits are generated.
Description
Technical Field
The invention relates to the technical field of computer vision, in particular to a multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning.
Background
With the rapid development of deep learning, multimodal 2D and 3D Facial Expression Recognition (FER) are receiving wide attention in the field of computer vision. The methods based on deep learning are that a plurality of 3D attribute graphs are extracted by utilizing 3D point cloud data, the attribute graphs and a 2D face graph are used as input and are respectively sent into each characteristic extraction branch of a CNN network, and finally, the extracted characteristics of each branch are fused to be used as the input of a classifier. However, since the 2D color map and the 3D attribute map are both from the same sample, there may be redundancy in the learned features of each branch, which is not conducive to direct feature fusion, and in addition, one branch is used for each Zhang Shuxing map to extract features, which greatly increases the complexity of the model.
Disclosure of Invention
The invention aims to provide a multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning, aiming at the defects and shortcomings of the prior art, and the method can be used for reducing the complexity of a deep learning network and inhibiting the redundancy among the features extracted from different branches in the network.
In order to achieve the purpose, the invention adopts the following technical scheme: three attribute maps, namely a depth map, a direction map and an elevation map, are generated by using face point cloud data, the depth map, the direction map and the elevation map are combined into a three-channel RGB map, and the RGB map is used as the input of a certain branch in a network, so that the parameter quantity of the model is reduced.
The multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning introduces an orthogonal module to ensure that the features are orthogonal during feature fusion.
According to the multi-mode 2D and 3D facial expression recognition method based on orthogonal guide learning, the feature extraction part uses two network branches with different structures to respectively extract features of a 2D face graph and a 3D attribute graph, the features are respectively defined as FE2DNet and FE3DNet, the FE2DNet is a VGG network deformation, and the FE3DNet is a Resnet derivation.
In the multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning, a Global Weighted Pooling (GWP) layer is adopted to replace a GAP layer, which is different from a general object, in the facial expression recognition task, images input into a CNN network are aligned through key points, so that in a deep feature map, each pixel represents a certain specific area of the input image and contains fixed semantic information. Important areas such as mouth, nose, eyes, etc. play a crucial role in the correct classification of expressions, and additional attention needs to be paid to semantic information of these areas. Using GAP directly, averaging all pixels directly, the semantic information of these key regions is likely to be ignored.
According to the multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning, each feature map is provided with a weight map with the same size, the weight in the weight map can be updated by gradient descent, the output feature vector is obtained by calculating the dot product of the feature map and the weight map, and the calculation formula is shown as follows:(wherein x k ,y k ,w k Respectively representing the values of the feature map, the weight map and the corresponding feature vector elements), after training with a large amount of face data, the weight map will focus more on a specific spatial region, and a larger weight in the weight map indicates that the spatial region contributes more to the final classification result.
The multimode 2D and 3D facial expression recognition method based on orthogonal guide learning is characterized in that input images of two channels are 2D gray level images and 3D attribute images from the same face, and feature vectors V extracted by a feature extractor 1 And V 2 There may be redundancy, let V before feature fusion 1 And V 2 Passing through an orthogonal guide module to output a feature vector F 1 And F 2 Orthogonal, removing the redundant part between the two vectors.The orthogonal guide module is composed of a full connection layer and a Relu layer.
The multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning,
orthogonal steering modules are respectively provided with V 1 And V 2 As input, it is converted by the full connection layer and two orthogonal features F are output 1 And F 2 Designing a quadrature loss function L orth To supervise updating of orthogonal bootstrap module weights to ensure F 1 And F 2 Orthogonality therebetween. L is orth Is defined as follows:wherein θ is F 1 And F 2 The included angle therebetween. When loss function L orth The closer to 0, the closer to 90 degrees the angle θ is represented, and F is 1 And F 2 The more orthogonal, i.e. uncorrelated, there is.
The working principle of the invention is as follows: a multi-mode 2D and 3D face expression recognition method based on orthogonal guide learning utilizes face point cloud data to generate three attribute maps which are a depth map, a direction map and an elevation map respectively, the depth map, the direction map and the elevation map are combined into a three-channel RGB map, an orthogonal module is introduced to ensure that features are orthogonal during feature fusion, and before feature fusion, V is firstly led to 1 And V 2 Passing through an orthogonal guide module to output a feature vector F 1 And F 2 Orthogonal, removing the redundant part between the two vectors.
After the technical scheme is adopted, the invention has the beneficial effects that: by the multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning, the complexity of a deep learning network is reduced, the redundancy among features extracted from different branches in the network is inhibited, and good economic and social benefits are generated.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
FIG. 1 is a schematic diagram of the network architecture of the present invention and its flow chart;
FIG. 2 is a schematic diagram of the network structure of the FE2DNet and FE3DNet of the present invention;
figure 3 is a schematic flow diagram of the GWP operation structure of the present invention;
fig. 4 is a flow chart of the structure of the orthogonal guide module of the present invention.
Detailed Description
Referring to fig. 1 to 4, the technical solution adopted by the present embodiment is: the method utilizes face point cloud data to generate three attribute maps which are respectively a depth map, a direction map and an elevation map, wherein the depth map, the direction map and the elevation map are combined into a three-channel RGB map, and the RGB map is used as the input of a certain branch in a network, so that the parameter quantity of a model is reduced.
Furthermore, the multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning introduces an orthogonal module to ensure that the features are orthogonal during feature fusion.
Further, in the method for recognizing the multi-modal 2D and 3D facial expressions based on the orthogonal guide learning, the feature extraction part uses two network branches with different structures to respectively extract features of the 2D face map and the 3D attribute map, which are respectively defined as FE2DNet and FE3DNet, where FE2DNet is a variant of the VGG network and FE3DNet is a derivative of Resnet.
Further, in the multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning, a Global Weighted Pooling (GWP) layer is used to replace a GAP layer, which is different from a general object, in the facial expression recognition task, images input to the CNN network are aligned through key points, so that in a deep feature map, each pixel represents a specific area of the input image and contains fixed semantic information. Important areas such as mouth, nose, eyes, etc. play a crucial role in the correct classification of expressions, and additional attention needs to be paid to semantic information of these areas. Using GAP directly, averaging all pixels directly, the semantic information of these key regions is likely to be ignored.
Furthermore, each feature map is provided with a weight map with the same size, the weight in the weight map can be updated by gradient descent, the output feature vector is obtained by calculating the dot product of the feature map and the weight map, and the calculation formula is as follows:(wherein x k ,y k ,w k Respectively representing the values of the feature map, the weight map and the corresponding feature vector elements), after training with a large amount of face data, the weight map will focus more on a specific spatial region, and a larger weight in the weight map indicates that the spatial region contributes more to the final classification result.
Further, according to the multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning, input images of two channels are both a 2D gray level image and a 3D attribute image of the same face, and a feature vector V extracted by the feature extractor 1 And V 2 There may be redundancy, so V is left before feature fusion takes place 1 And V 2 Passing through an orthogonal guide module to output a feature vector F 1 And F 2 Orthogonal, removing the redundant part between the two vectors. The orthogonal guide module is composed of a full connection layer and a Relu layer.
Further, in the multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning, the orthogonal guide modules respectively adopt V 1 And V 2 As input, it is converted by the full connection layer and two orthogonal features F are output 1 And F 2 Design a quadrature loss function L orth To supervise updating of orthogonal bootstrap module weights to ensure F 1 And F 2 Orthogonality therebetween. L is orth The formula of (c) is defined as follows:wherein θ is F 1 And F 2 The included angle therebetween. When loss function L orth The closer to 0, the closer to 90 degrees the angle θ is represented, in which case F 1 And F 2 The more orthogonal, i.e. uncorrelated, there is.
The working principle of the invention is as follows: a multi-mode 2D and 3D face expression recognition method based on orthogonal guide learning utilizes face point cloud data to generate three attribute maps which are a depth map, a direction map and an elevation map respectively, the depth map, the direction map and the elevation map are combined into a three-channel RGB map, an orthogonal module is introduced to ensure that features are orthogonal during feature fusion, and before feature fusion, V is firstly led to 1 And V 2 Passing through an orthogonal guide module to output a feature vector F 1 And F 2 Orthogonal, removing the redundant part between the two vectors.
After the technical scheme is adopted, the invention has the beneficial effects that: the multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning reduces the complexity of a deep learning network and inhibits the redundancy among the features extracted from different branches in the network, thereby generating good economic benefit and social benefit.
The above description is only for the purpose of illustrating the technical solutions of the present invention and not for the purpose of limiting the same, and other modifications or equivalent substitutions made by those skilled in the art to the technical solutions of the present invention should be covered within the scope of the claims of the present invention without departing from the spirit and scope of the technical solutions of the present invention.
Claims (3)
1. A multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning is characterized in that: the method comprises the steps of generating three attribute maps which are respectively a depth map, a direction map and an elevation map by using face point cloud data, synthesizing the depth map, the direction map and the elevation map into a three-channel RGB map, using the RGB map as the input of a branch in a network, and using two attribute maps as a feature extraction partIn the human face expression recognition task, images input into a CNN network are aligned through key points, each pixel in a deep feature map represents a specific area of an input image, input images of two channels are 2D gray level maps and 3D attribute maps from the same human face, and feature vectors V extracted by a feature extractor are respectively 2D face maps and 3D attribute maps 1 And V 2 Having redundancy, let V before feature fusion 1 And V 2 Passing through an orthogonal guide module to output a feature vector F 1 And F 2 Orthogonal to eliminate redundant part between two vectors, and the orthogonal guide module is composed of a layer of full connection layer and Relu layer, and the orthogonal guide modules are respectively divided into V 1 And V 2 As input, it is converted by the full connection layer and two orthogonal features F are output 1 And F 2 Designing a quadrature loss function L orth To supervise updating of orthogonal bootstrap module weights to ensure F 1 And F 2 Orthogonality therebetween.
2. The multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning of claim 1, wherein: an orthogonality module is introduced to ensure that the features are orthogonal when the features are fused.
3. The multi-modal 2D and 3D facial expression recognition method based on orthogonal guide learning of claim 1, wherein: each feature map is provided with a weight map with the same size, the weight in the weight map is updated by gradient descent, and the output feature vector is calculated by the dot product of the feature map and the weight map.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010347655.1A CN111488856B (en) | 2020-04-28 | 2020-04-28 | Multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010347655.1A CN111488856B (en) | 2020-04-28 | 2020-04-28 | Multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111488856A CN111488856A (en) | 2020-08-04 |
CN111488856B true CN111488856B (en) | 2023-04-18 |
Family
ID=71796623
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010347655.1A Active CN111488856B (en) | 2020-04-28 | 2020-04-28 | Multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111488856B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112052834B (en) * | 2020-09-29 | 2022-04-08 | 支付宝(杭州)信息技术有限公司 | Face recognition method, device and equipment based on privacy protection |
CN113408462B (en) * | 2021-06-29 | 2023-05-02 | 西南交通大学 | Landslide remote sensing information extraction method based on convolutional neural network and class thermodynamic diagram |
CN113642467B (en) * | 2021-08-16 | 2023-12-01 | 江苏师范大学 | Facial expression recognition method based on improved VGG network model |
Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102043943A (en) * | 2009-10-23 | 2011-05-04 | 华为技术有限公司 | Method and device for obtaining human face pose parameter |
WO2016110005A1 (en) * | 2015-01-07 | 2016-07-14 | 深圳市唯特视科技有限公司 | Gray level and depth information based multi-layer fusion multi-modal face recognition device and method |
CN106778468A (en) * | 2016-11-14 | 2017-05-31 | 深圳奥比中光科技有限公司 | 3D face identification methods and equipment |
CN107392190A (en) * | 2017-09-07 | 2017-11-24 | 南京信息工程大学 | Color face recognition method based on semi-supervised multi views dictionary learning |
JP2018055470A (en) * | 2016-09-29 | 2018-04-05 | 国立大学法人神戸大学 | Facial expression recognition method, facial expression recognition apparatus, computer program, and advertisement management system |
CN108510573A (en) * | 2018-04-03 | 2018-09-07 | 南京大学 | A method of the multiple views human face three-dimensional model based on deep learning is rebuild |
CN108573284A (en) * | 2018-04-18 | 2018-09-25 | 陕西师范大学 | Deep learning facial image extending method based on orthogonal experiment analysis |
CN109299702A (en) * | 2018-10-15 | 2019-02-01 | 常州大学 | A kind of Human bodys' response method and system based on depth space-time diagram |
CN109344909A (en) * | 2018-10-30 | 2019-02-15 | 咪付(广西)网络技术有限公司 | A kind of personal identification method based on multichannel convolutive neural network |
WO2019196308A1 (en) * | 2018-04-09 | 2019-10-17 | 平安科技(深圳)有限公司 | Device and method for generating face recognition model, and computer-readable storage medium |
CN110516616A (en) * | 2019-08-29 | 2019-11-29 | 河南中原大数据研究院有限公司 | A kind of double authentication face method for anti-counterfeit based on extensive RGB and near-infrared data set |
CN110717916A (en) * | 2019-09-29 | 2020-01-21 | 华中科技大学 | Pulmonary embolism detection system based on convolutional neural network |
CN114638283A (en) * | 2022-02-11 | 2022-06-17 | 华南理工大学 | Orthogonal convolution neural network image identification method based on tensor optimization space |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11216541B2 (en) * | 2018-09-07 | 2022-01-04 | Qualcomm Incorporated | User adaptation for biometric authentication |
CN109815785A (en) * | 2018-12-05 | 2019-05-28 | 四川大学 | A kind of face Emotion identification method based on double-current convolutional neural networks |
-
2020
- 2020-04-28 CN CN202010347655.1A patent/CN111488856B/en active Active
Patent Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102043943A (en) * | 2009-10-23 | 2011-05-04 | 华为技术有限公司 | Method and device for obtaining human face pose parameter |
WO2016110005A1 (en) * | 2015-01-07 | 2016-07-14 | 深圳市唯特视科技有限公司 | Gray level and depth information based multi-layer fusion multi-modal face recognition device and method |
JP2018055470A (en) * | 2016-09-29 | 2018-04-05 | 国立大学法人神戸大学 | Facial expression recognition method, facial expression recognition apparatus, computer program, and advertisement management system |
CN106778468A (en) * | 2016-11-14 | 2017-05-31 | 深圳奥比中光科技有限公司 | 3D face identification methods and equipment |
CN107392190A (en) * | 2017-09-07 | 2017-11-24 | 南京信息工程大学 | Color face recognition method based on semi-supervised multi views dictionary learning |
CN108510573A (en) * | 2018-04-03 | 2018-09-07 | 南京大学 | A method of the multiple views human face three-dimensional model based on deep learning is rebuild |
WO2019196308A1 (en) * | 2018-04-09 | 2019-10-17 | 平安科技(深圳)有限公司 | Device and method for generating face recognition model, and computer-readable storage medium |
CN108573284A (en) * | 2018-04-18 | 2018-09-25 | 陕西师范大学 | Deep learning facial image extending method based on orthogonal experiment analysis |
CN109299702A (en) * | 2018-10-15 | 2019-02-01 | 常州大学 | A kind of Human bodys' response method and system based on depth space-time diagram |
CN109344909A (en) * | 2018-10-30 | 2019-02-15 | 咪付(广西)网络技术有限公司 | A kind of personal identification method based on multichannel convolutive neural network |
CN110516616A (en) * | 2019-08-29 | 2019-11-29 | 河南中原大数据研究院有限公司 | A kind of double authentication face method for anti-counterfeit based on extensive RGB and near-infrared data set |
CN110717916A (en) * | 2019-09-29 | 2020-01-21 | 华中科技大学 | Pulmonary embolism detection system based on convolutional neural network |
CN114638283A (en) * | 2022-02-11 | 2022-06-17 | 华南理工大学 | Orthogonal convolution neural network image identification method based on tensor optimization space |
Non-Patent Citations (6)
Title |
---|
orthogonalization guided feature fusion network for multimodal 2D +3D facial expression recognition;Lin,SS等;《IEEE TRANSACTION ON MULTIMEDIA》;第23卷;第1581-1591页 * |
Towards Reading Beyond Faces for Sparsity-aware 3D/4D Affect Recognition;MuzammilBehzad等;《Neurocomputing》;第458卷;第297-307页 * |
一种基于两步降维和并行特征融合的表情识别方法;杨勇等;《重庆邮电大学学报(自然科学版)》;第27卷(第3期);第377-385页 * |
人脸表情识别关键技术的研究;李宏菲;《中国优秀硕士学位论文全文数据库信息科技辑》(第8期);第I138-1019页 * |
基于卷积神经网络的人脸表情识别研究;李思泉等;《软件导刊》;第17卷(第1期);第28-31页 * |
基于多视图核鉴别相关与正交分析的图像分类;朱震宇;《中国优秀硕士学位论文全文数据库信息科技辑》(第2期);第I138-3605页 * |
Also Published As
Publication number | Publication date |
---|---|
CN111488856A (en) | 2020-08-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111488856B (en) | Multimodal 2D and 3D facial expression recognition method based on orthogonal guide learning | |
WO2022151535A1 (en) | Deep learning-based face feature point detection method | |
CN110852182B (en) | Depth video human body behavior recognition method based on three-dimensional space time sequence modeling | |
CN110175248B (en) | Face image retrieval method and device based on deep learning and Hash coding | |
CN113269089A (en) | Real-time gesture recognition method and system based on deep learning | |
CN113111751B (en) | Three-dimensional target detection method capable of adaptively fusing visible light and point cloud data | |
US12080098B2 (en) | Method and device for training multi-task recognition model and computer-readable storage medium | |
CN115018999B (en) | Multi-robot collaboration dense point cloud map construction method and device | |
CN112036260A (en) | Expression recognition method and system for multi-scale sub-block aggregation in natural environment | |
CN113591928B (en) | Vehicle re-identification method and system based on multi-view and convolution attention module | |
Sun et al. | IRDCLNet: Instance segmentation of ship images based on interference reduction and dynamic contour learning in foggy scenes | |
CN111862030A (en) | Face synthetic image detection method and device, electronic equipment and storage medium | |
CN112241959A (en) | Attention mechanism generation semantic segmentation method based on superpixels | |
CN111274946B (en) | Face recognition method, system and equipment | |
Kang et al. | Real-time eye tracking for bare and sunglasses-wearing faces for augmented reality 3D head-up displays | |
Mizukami et al. | CUDA implementation of deformable pattern recognition and its application to MNIST handwritten digit database | |
CN113902753A (en) | Image semantic segmentation method and system based on dual-channel and self-attention mechanism | |
CN113822114A (en) | Image processing method, related equipment and computer readable storage medium | |
CN114639132A (en) | Feature extraction model processing method, device and equipment in face recognition scene | |
CN112434576A (en) | Face recognition method and system based on depth camera | |
CN114937153B (en) | Visual characteristic processing system and method based on neural network in weak texture environment | |
CN116959073A (en) | Direction-adaptive multi-pose point cloud face recognition method and system | |
CN116246305A (en) | Pedestrian retrieval method based on hybrid component transformation network | |
Bai et al. | DHRNet: A Dual-Branch Hybrid Reinforcement Network for Semantic Segmentation of Remote Sensing Images | |
CN113963374A (en) | Pedestrian attribute identification method based on multi-level features and identity information assistance |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |