WO2020125623A1 - Procédé et dispositif de détection de corps vivant, support d'informations et dispositif électronique - Google Patents
Procédé et dispositif de détection de corps vivant, support d'informations et dispositif électronique Download PDFInfo
- Publication number
- WO2020125623A1 WO2020125623A1 PCT/CN2019/125957 CN2019125957W WO2020125623A1 WO 2020125623 A1 WO2020125623 A1 WO 2020125623A1 CN 2019125957 W CN2019125957 W CN 2019125957W WO 2020125623 A1 WO2020125623 A1 WO 2020125623A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- dimensional color
- image
- depth
- face
- color image
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/40—Spoof detection, e.g. liveness detection
- G06V40/45—Detection of the body part being alive
Definitions
- the present application relates to the technical field of face recognition, in particular to a living body detection method, device, storage medium and electronic equipment.
- electronic devices use relevant face recognition technology to not only distinguish between individual users, but also perform live detection on users. For example, electronic devices obtain user faces (such as photos taken through a depth camera such as a structured light camera or a time-of-flight camera). The RGB-D image of the user's face image) can determine whether the user's face is a living face.
- user faces such as photos taken through a depth camera such as a structured light camera or a time-of-flight camera.
- the RGB-D image of the user's face image can determine whether the user's face is a living face.
- the embodiments of the present application provide a living body detection method, device, storage medium, and electronic equipment, which can reduce the hardware cost of the electronic equipment for living body detection.
- an embodiment of the present application provides a living body detection method, which is applied to an electronic device, the electronic device includes a monocular camera, and the living body detection method includes:
- the two-dimensional color image and the depth image are input into a pre-trained living body detection model to obtain a detection result.
- an embodiment of the present application provides a living body detection device, which is applied to an electronic device, the electronic device includes a monocular camera, and the living body detection device includes:
- a color image acquisition module configured to shoot the face to be detected through the monocular camera to obtain a two-dimensional color image of the face to be detected
- a depth image acquisition module configured to input the two-dimensional color image into a pre-trained depth estimation model to obtain a depth image corresponding to the two-dimensional color image;
- the living body face detection module inputs the two-dimensional color image and the depth image into a pre-trained living body detection model to obtain a detection result.
- an embodiment of the present application provides a storage medium on which a computer program is stored, and when the computer program runs on a computer, causes the computer to execute:
- the two-dimensional color image and the depth image are input into a pre-trained living body detection model to obtain a detection result.
- an embodiment of the present application provides an electronic device including a processor, a memory, and a monocular camera.
- the memory has a computer program
- the processor is used to execute the computer program by calling the computer program:
- the two-dimensional color image and the depth image are input into a pre-trained living body detection model to obtain a detection result.
- FIG. 1 is a schematic flowchart of a living body detection method provided by an embodiment of the present application.
- FIG. 2 is a schematic diagram of the living body detection performed by the electronic device through the living body detection model in the embodiment of the present application.
- FIG. 3 is another schematic flowchart of the living body detection method provided by the embodiment of the present application.
- FIG. 4 is a schematic diagram of constructing a training sample set in an embodiment of the present application.
- FIG. 5 is a schematic structural diagram of a living body detection device provided by an embodiment of the present application.
- FIG. 6 is a schematic structural diagram of an electronic device provided by an embodiment of the present application.
- FIG 7 is another schematic structural diagram of an electronic device provided by an embodiment of the present application.
- the related art proposes a living body detection technology based on a depth camera such as a structured light camera or a time-of-flight camera.
- a depth camera such as a structured light camera or a time-of-flight camera.
- its implementation requires that the electronic device be equipped with an additional depth camera, which increases The cost of biopsy.
- the embodiments of the present application firstly provide a living body detection method, which realizes living body detection based on a monocular camera commonly configured in electronic devices, without increasing the hardware cost of the electronic devices.
- the execution subject of the living body detection method may be the living body detection device provided in the embodiment of the present application, or an electronic device integrated with the living body detection device.
- the living body detection device may be implemented by hardware or software, and the electronic device may be intelligent Mobile phones, tablet computers, PDAs, notebook computers, or desktop computers are equipped with processors and have processing capabilities.
- An embodiment of the present application provides a living body detection method, including:
- the two-dimensional color image and the depth image are input into a pre-trained living body detection model for living body detection, and a detection result is obtained.
- the living body detection model is a convolutional neural network model, including a convolutional layer, a pooling layer, and a fully connected layer connected in sequence, and the inputting of the two-dimensional color image and the depth image in advance Trained living body detection model to obtain the detection results, including:
- the inputting the two-dimensional color image and the depth image into the convolution layer for feature extraction to obtain the joint global features of the two-dimensional color image and the depth image includes:
- the preprocessing the two-dimensional color image to obtain the face area image in the two-dimensional color image includes:
- An ellipse template, a circular template or a rectangular template is used to extract the face area image from the two-dimensional color image.
- the method before capturing the face to be detected through the monocular camera to obtain a two-dimensional color image of the face to be detected, the method further includes:
- a plurality of different live human face images are captured by the monocular camera to obtain multiple two-dimensional color live human face sample images, and a depth image corresponding to each of the two-dimensional color live human face sample images is obtained to obtain multiple A depth image;
- a plurality of different non-living human faces are photographed through the monocular camera to obtain a plurality of two-dimensional color non-living human face sample images, and a depth image corresponding to each of the two-dimensional color non-living human face sample images is obtained to obtain Multiple second depth images;
- a convolutional neural network is used to perform model training on the training sample set to obtain the convolutional neural network model.
- the convolutional neural network is used to perform model training on the training sample set, and before the convolutional neural network model is obtained, the method further includes:
- the acquiring depth images corresponding to each of the two-dimensional color live human face sample images to obtain multiple first depth images includes:
- a depth image corresponding to each two-dimensional color live human face sample image is generated to obtain a plurality of first depth images.
- the living body detection method further includes:
- each of the two-dimensional color live human face sample images and each of the two-dimensional color non-live human face sample images as training inputs, and using the first depth image corresponding to each of the two-dimensional color live human face sample images and each location
- the second depth image corresponding to the two-dimensional color non-living human face sample image is output as the target, and the supervised model training is performed to obtain the depth estimation model.
- the method before inputting the two-dimensional color image into a pre-trained depth estimation model for depth estimation, the method further includes:
- FIG. 1 is a schematic flowchart of a living body detection method provided by an embodiment of the present application.
- the flow of the living body detection method provided by the embodiment of the present application may be as follows:
- a face to be detected is photographed through a monocular camera to obtain a two-dimensional color image of the face to be detected.
- the electronic device can treat the detected person through the configured monocular camera when receiving an operation that requires face recognition for user identity detection, such as an unlock operation based on face recognition or a payment operation based on face recognition The face is photographed. Since the monocular camera is only sensitive to two-dimensional color information, a two-dimensional color image of the face to be detected will be captured.
- a front monocular camera also commonly known as a front camera
- a rear monocular camera also commonly known as a rear camera
- the imaging capability of the rear monocular camera is higher than the imaging capability of the front monocular camera, so that when the electronic device shoots the face to be detected through the monocular camera, it can default to perform the shooting through the front monocular camera Operation to shoot the face to be detected; the shooting operation can also be performed by the rear monocular camera by default to shoot the face to be detected; the front monocular camera and the rear can also be predicted based on the real-time pose information
- the monocular camera facing the face to be detected in the monocular camera so that the shooting operation is automatically performed by the monocular camera facing the face to be detected in the front monocular camera and the rear monocular camera, and the face to be detected is photographed .
- the current unlocking method adopted by the electronic device is “face unlocking”.
- the electronic device receives the trigger operation for unlocking the face, by default the front monocular camera is used to shoot the face to be detected, thereby obtaining the person to be detected Two-dimensional color image of the face.
- the payment method currently adopted by the electronic device is "face-swapping payment"
- the electronic device receives the trigger operation of the face-swapping payment
- the face to be detected is photographed by the front monocular camera by default, thereby obtaining the pending detection Two-dimensional color image of human face.
- the captured two-dimensional color image is input to a pre-trained depth estimation model to perform depth estimation to obtain a depth image corresponding to the two-dimensional color image.
- a depth estimation model for depth estimation is pre-trained, where the depth estimation model may be stored locally in the electronic device or may be stored in a remote server.
- the electronic device calls the pre-trained depth estimation model locally or calls the pre-trained depth estimation model from the remote server, and transfers the person to be detected
- the two-dimensional color image of the face is input to a pre-trained depth estimation model, and the depth estimation model is used to perform depth estimation on the two-dimensional color image to obtain a depth image corresponding to the two-dimensional color image.
- the resolution of the estimated depth image is the same as the resolution of the two-dimensional color image.
- the pixel value of each pixel in the depth image is used to describe the corresponding pixel in the two-dimensional color image to the aforementioned monocular camera (That is, the distance of a monocular camera that captures a two-dimensional color image).
- the electronic device After obtaining the two-dimensional color image of the face to be detected through the front monocular camera, the electronic device calls a locally stored, pre-trained depth estimation model, and uses the depth estimation model to perform depth estimation on the two-dimensional color image. Get a depth image corresponding to a two-dimensional color image.
- a two-dimensional color image and its corresponding depth image are input into a pre-trained living body detection model for living body detection, and a detection result is obtained.
- the living body detection model for living body detection is also pre-trained, where the living body detection model may be stored locally in the electronic device , Can also be stored in a remote server.
- the pre-trained living body detection model is called locally or from a remote
- the server at the end calls the pre-trained living body detection model, and inputs the previously acquired two-dimensional color image and its corresponding depth image to the pre-trained living body detection model.
- the living body detection model is based on the input two-dimensional color image and The corresponding depth image performs live detection on the face to be detected to obtain a detection result that the face to be detected is a living face, or a detection result that the face to be detected is a non-living face.
- the electronic device calls a locally stored, pre-trained depth estimation model, and uses the depth estimation model for the two-dimensional color Perform depth estimation on the image to obtain a depth image corresponding to the two-dimensional color image, then call the locally stored, pre-trained living body detection model, and input the previously obtained two-dimensional color image and its corresponding depth image to the living body detection model for living body Detection, and the detection result is obtained, wherein, if the detection result of the face to be detected is a living face, it means that the face to be detected is a real face of a person with vital signs, and if the face to be detected is a non-living face
- the detection result indicates that the face to be detected is not the real face of the person with vital signs, and may be a face image or a face video captured in advance.
- the electronic device in the embodiment of the present application can first obtain the two-dimensional color image of the face to be detected by the configured monocular camera, and then input the obtained two-dimensional color image into the pre-trained depth estimation model Perform depth estimation to obtain a depth image corresponding to a two-dimensional color image, and finally input the previously obtained two-dimensional color image and its corresponding depth image into a pre-trained living body detection model for living body detection to obtain a detection result.
- the electronic device can realize the living body detection without using an additional depth camera, but using a generally configured monocular camera, which reduces the hardware cost of the electronic device for living body detection.
- FIG. 3 is another schematic flowchart of the living body detection method provided by the embodiment of the present application.
- the living body detection method may be applied to an electronic device, and the flow of the living body detection method may include:
- the electronic device is trained with a machine learning algorithm to obtain a depth estimation model and a living body detection model, where the living body detection model is a convolutional neural network model.
- the electronic device uses a machine learning algorithm to train in advance to obtain a depth estimation model and a living body detection model. It should be noted that after the trained depth estimation model and the living body detection model, the electronic device may store the depth estimation model and the living body detection model locally in the electronic device, or may store the depth estimation model and the living body detection model on a remote The server may also store one of the depth estimation model and the living body detection model locally on the electronic device and the other on a remote server.
- machine learning algorithms can include: decision tree models, logistic regression models, Bayesian models, neural network models, clustering models, and so on.
- machine learning algorithms can be divided according to various situations. For example, based on the learning method, machine learning algorithms can be divided into: supervised learning algorithms, unsupervised learning algorithms, semi-supervised learning algorithms, reinforcement learning algorithms, and so on.
- training data Under supervised learning, the input data is called “training data”, and each set of training data has a clear identification or result, such as “spam” and “non-spam” in anti-spam systems, and recognition of handwritten digits. "1,2,3,4" and so on.
- Common application scenarios of supervised learning are classification problems and regression problems.
- Common algorithms are Logistic Regression and Backward Propagation Neural Network.
- the model In unsupervised learning, the data is not specifically identified, the model is to infer some internal structure of the data.
- Common application scenarios include association rule learning and clustering.
- Common algorithms include Apriori algorithm and k-Means algorithm.
- Semi-supervised learning algorithm In this learning mode, the input data is partially identified.
- This learning model can be used for type recognition, but the model first needs to learn the internal structure of the data in order to reasonably organize the data for prediction.
- Application scenarios include classification and regression.
- the algorithm includes some extensions to commonly used supervised learning algorithms. These algorithms first attempt to model unlabeled data, and then predict the labeled data on this basis. Such as graph theory inference algorithm (Graph Inference) or Laplacian support vector machine (Laplacian SVM), etc.
- Reinforcement learning algorithm In this learning mode, the input data is used as feedback to the model. Unlike the supervised model, the input data is only used as a way to check whether the model is right or wrong. Under reinforcement learning, the input data is directly fed back to the model. The model must make adjustments immediately.
- Common application scenarios include dynamic systems and robot control.
- Common algorithms include Q-Learning and time difference learning (Temporal learning).
- Regression algorithms common regression algorithms include: least squares (Ordinary Least Square), logistic regression (Logistic Regression), stepwise regression (Stepwise Regression), multiple adaptive regression spline (Multivariate Adaptive Regression Splines) and local scatter smoothing Estimate (Locally Estimated Scatterplot Smoothing).
- KNN k-Nearest Neighbor
- LVQ Learning Vector Quantization
- SOM Self-Organizing Map
- Regularization methods common algorithms include: Ridge Regression, Least Absolute Shrinkage and Selection Operator (LASSO), and Elastic Network (Elastic Net).
- LASSO Least Absolute Shrinkage and Selection Operator
- Elastic Net Elastic Net
- Decision tree algorithm common algorithms include: Classification and regression tree (Classification And Regression Tree, CART), ID3 (Iterative Dichotomiser 3), C4.5, Chi-squared Automatic Interaction Detection (CHAID), Decision Stump, Random Forest (Random Forest), Multiple Adaptive Regression Spline (MARS) and Gradient Boosting Machine (Gradient Boosting Machine, GBM).
- Bayesian algorithm including: Naive Bayesian algorithm, Average One-Dependence Estimators (AODE), and Bayesian Belief Network (BBN).
- AODE Average One-Dependence Estimators
- BBN Bayesian Belief Network
- a convolutional neural network is used to train a live detection model, that is, the live detection model is a convolutional neural network model, where the convolutional neural network model includes a convolutional layer, a pooling layer, and a fully connected Floor.
- the electronic device shoots the face to be detected through a monocular camera to obtain a two-dimensional color image of the face to be detected.
- the electronic device may treat the detected person through the configured monocular camera when receiving an operation that requires face recognition for user identity detection, such as an unlock operation based on face recognition or a payment operation based on face recognition
- an operation that requires face recognition for user identity detection such as an unlock operation based on face recognition or a payment operation based on face recognition
- the face is photographed. Since the monocular camera is only sensitive to two-dimensional color information, a two-dimensional color image of the face to be detected will be captured.
- a front monocular camera also commonly known as a front camera
- a rear monocular camera also commonly known as a rear camera
- the imaging capability of the rear monocular camera is higher than the imaging capability of the front monocular camera, so that when the electronic device shoots the face to be detected through the monocular camera, it can default to perform the shooting through the front monocular camera Operation to shoot the face to be detected; the shooting operation can also be performed by the rear monocular camera by default to shoot the face to be detected; the front monocular camera and the rear can also be predicted based on the real-time pose information
- the monocular camera facing the face to be detected in the monocular camera so that the shooting operation is automatically performed by the monocular camera facing the face to be detected in the front monocular camera and the rear monocular camera, and the face to be detected is photographed .
- the current unlocking method adopted by the electronic device is “face unlocking”.
- the electronic device receives the trigger operation for unlocking the face, by default the front monocular camera is used to shoot the face to be detected, thereby obtaining the person to be detected Two-dimensional color image of the face.
- the payment method currently adopted by the electronic device is "face-swapping payment"
- the electronic device receives the trigger operation of the face-swapping payment
- the face to be detected is photographed by the front monocular camera by default, thereby obtaining the pending detection Two-dimensional color image of human face.
- the electronic device inputs the captured two-dimensional color image into a pre-trained depth estimation model to perform depth estimation to obtain a depth image corresponding to the two-dimensional color image.
- the electronic device After acquiring the two-dimensional color image of the face to be detected through the monocular camera, the electronic device calls the pre-trained depth estimation model locally or calls the pre-trained depth estimation model from a remote server, and transfers the person to be detected
- the two-dimensional color image of the face is input to a pre-trained depth estimation model, and the depth estimation model is used to perform depth estimation on the two-dimensional color image to obtain a depth image corresponding to the two-dimensional color image.
- the resolution of the estimated depth image is the same as the resolution of the two-dimensional color image.
- the pixel value of each pixel in the depth image is used to describe the corresponding pixel in the two-dimensional color image to the aforementioned monocular camera (That is, the distance of a monocular camera that captures a two-dimensional color image).
- the electronic device After obtaining the two-dimensional color image of the face to be detected through the front monocular camera, the electronic device calls a locally stored, pre-trained depth estimation model, and uses the depth estimation model to perform depth estimation on the two-dimensional color image. Get a depth image corresponding to a two-dimensional color image.
- the electronic device inputs the aforementioned two-dimensional color image and its corresponding depth image to the convolutional layer of the convolutional neural network model for feature extraction, and obtains the combined global features of the aforementioned two-dimensional color image and the aforementioned depth image.
- the electronic device after inputting the two-dimensional color image captured by the monocular camera into the pre-trained depth estimation model and obtaining the depth image corresponding to the two-dimensional color image, the electronic device locally calls the pre-trained living body detection
- the electronic device inputs the aforementioned two-dimensional color image and its corresponding depth image into the convolutional layer of the convolutional neural network model for feature extraction (feature extraction is to map the original image data to the hidden layer feature space, thereby Get the corresponding global features), get the global features of the two-dimensional color image and the global features of the depth image.
- feature extraction is to map the original image data to the hidden layer feature space, thereby Get the corresponding global features
- the global features of the two-dimensional color image and the global features of the depth image are combined in the convolutional layer to obtain the joint global features of the foregoing two-dimensional color image and the foregoing depth image.
- the electronic device obtains the joint global feature and inputs the pooling layer of the convolutional neural network model to perform feature dimensionality reduction to obtain the joint global feature after the dimensionality reduction.
- the joint global features of the two-dimensional color image and the depth image output by the convolution layer will be input into the pooling layer of the convolutional neural network model.
- Sampling is to retain the salient features of the joint global features and achieve feature dimensionality reduction of the joint global features.
- downsampling can be achieved by means of maximum pooling or mean pooling.
- the joint global feature is subjected to feature dimensionality reduction through the pooling layer to obtain a joint global feature of 10*10 dimensionality reduction.
- the electronic device inputs the joint global features after dimensionality reduction into the fully connected layer of the convolutional neural network model for classification processing, and obtains the detection result that the face to be detected is a living face, or the face to be detected is a non-living body Face detection results.
- the fully connected layer is used to implement the function of the classifier.
- Each node of the fully connected layer is connected to all output nodes of the pooling layer.
- a node of the fully connected layer is called a neuron in the fully connected layer.
- the number of neurons in the connection layer can be determined according to the actual application requirements. For example, the number of neurons in the fully connected layer can be set to 4096, and so on.
- the dimensionality-reduced joint global features output by the pooling layer will be input to the fully connected layer for classification processing to obtain the detection result that the face to be detected is a living face, or the face to be detected is Detection results of non-living human faces.
- the electronic device preprocesses the two-dimensional color image to obtain the face area image in the two-dimensional color image
- the electronic device preprocesses the aforementioned depth image to obtain the face area image in the aforementioned depth image;
- the electronic device inputs the face area image in the two-dimensional color image and the face area image in the depth image to the convolutional layer for feature extraction to obtain the combined global features of the two-dimensional color image and the depth image .
- the electronic device inputs the aforementioned two-dimensional color image and its corresponding depth image into the convolutional layer of the convolutional neural network model for feature extraction, it is not the original two-dimensional color image and the original
- the aforementioned depth image is input to the convolutional layer of the convolutional neural network for feature extraction, but the two-dimensional color image and the depth image are preprocessed separately to obtain the face area image in the two-dimensional color image and the aforementioned Face area image in the depth image.
- the face area can be extracted from the two-dimensional color image and the depth image by using an oval template, a circular template, or a rectangular template, etc., respectively. Image, thereby obtaining the face area image in the aforementioned two-dimensional color image and the face area image in the aforementioned depth image.
- the electronic device shoots a plurality of different living human faces through a monocular camera to obtain a plurality of two-dimensional color living human face sample images, and obtains a depth image corresponding to each two-dimensional color living human face sample image to obtain multiple First depth image
- the electronic device shoots multiple different non-living human faces through a monocular camera to obtain multiple two-dimensional color non-living human face sample images, and obtains depth images corresponding to each two-dimensional color non-living human face sample images, Obtain multiple second depth images;
- the electronic device uses each two-dimensional color live human face sample image and its corresponding first depth image as a positive sample, and each two-dimensional color non-live human face sample image and its corresponding second depth image as a negative sample, Construct training sample set;
- the electronic device adopts a convolutional neural network to perform model training on the training sample set to obtain a convolutional neural network model as a living body detection model.
- the electronic device can shoot the faces of the users with different skin colors, different genders, and different ages (ie, live faces) through the monocular camera configured to obtain multiple two-dimensional color live face sample images
- the electronic device also obtains a depth image corresponding to each two-dimensional color non-living human face sample image to obtain multiple first depth images.
- the electronic device can also be connected to a depth camera.
- the external depth camera is used to shoot simultaneously. In this way, the electronic device will obtain the live face through the monocular camera.
- the two-dimensional color live human face sample image is captured by an external depth camera to obtain the depth image of the live human face, and then the captured depth image and the two-dimensional color live human face sample image are aligned, and the aligned depth image is recorded It is the first depth image of the two-dimensional color live human face sample image.
- the electronic device can also shoot different non-living human faces such as different facial images, facial videos, human face masks and human head models through its configured monocular camera to obtain multiple two-dimensional color non-living human face samples
- the electronic device also obtains depth images corresponding to the two-dimensional color non-living human face sample images to obtain multiple second depth images.
- the electronic device can also be connected with a depth camera.
- the external depth camera is used to shoot simultaneously. In this way, the electronic device will obtain the non-living human through the monocular camera.
- the two-dimensional color non-living face sample image of the face is captured by an external depth camera to obtain the depth image of the non-living face, and then the captured depth image and the two-dimensional color non-living face sample image are aligned and aligned
- the post-depth image is recorded as the second depth image of the two-dimensional color non-living human face sample image.
- each two-dimensional color live human face sample image and its corresponding first depth image are used as positive samples, and each two-dimensional color non-live human face sample image and its corresponding second depth image are used as negative samples to construct a training sample set, such as Figure 4 shows.
- the electronic device After completing the construction of the training sample set, the electronic device uses a convolutional neural network to perform model training on the constructed training sample set to obtain a convolutional neural network model as a living body detection model for living body detection.
- a convolutional neural network when used to perform model training on the constructed training sample set, a supervised learning method or an unsupervised learning method may be used, which can be specifically selected by a person of ordinary skill in the art according to actual needs.
- the convolutional neural network before the convolutional neural network is used to perform model training on the training sample set to obtain a convolutional neural network model, which is used as a living body detection model, it further includes:
- the electronic device performs sample expansion processing on the training sample set according to a preset sample expansion strategy.
- the sample expansion of the training sample set can increase the diversity of the samples, so that the trained convolutional neural network model has stronger robustness.
- the sample expansion strategy may be set to perform one or more of small rotation, scaling, and inversion on the positive samples/negative samples in the training sample set.
- the two-dimensional color live human face sample image and its corresponding first depth image can be Rotate the same amplitude to obtain the rotated two-dimensional color live human face sample image and the rotated first depth image.
- the new two-dimensional color live human face sample image and the rotated first depth image form a new Positive sample.
- the following when acquiring depth images corresponding to each two-dimensional color live human face sample image to obtain multiple first depth images, the following may be performed:
- the electronic device receives the distance from each pixel in the two-dimensional color live human face sample image to the monocular camera;
- the electronic device generates a depth image corresponding to each two-dimensional color live human face sample image according to the distance from each pixel in each two-dimensional color live human face sample image to the monocular camera, and obtains a plurality of first depth images.
- the distance from each pixel point in the two-dimensional color live human face sample image to the monocular camera can be manually calibrated, and the electronic The device generates a depth image corresponding to the two-dimensional color live human face sample image according to the distance between each pixel in the two-dimensional color live human face sample image and the monocular camera, and records it as the first depth image.
- the electronic device can receive the distance from each pixel in the two-dimensional color live human face sample image to the monocular camera, and according to the distance from each pixel in the two-dimensional color live human face sample image to the monocular camera , A depth image corresponding to each two-dimensional color live human face sample image is generated, and multiple first depth images are obtained.
- the following when acquiring depth images corresponding to each two-dimensional color non-living human face sample image to obtain multiple second depth images, the following may be performed:
- the electronic device receives the calibrated distance from each pixel in the two-dimensional color non-living face sample image to the monocular camera;
- the electronic device generates a depth image corresponding to each two-dimensional color non-living human face sample image according to the distance from each pixel in each two-dimensional color non-living human face sample image to the monocular camera, and obtains a plurality of second depth images.
- a machine learning algorithm when used to obtain a depth estimation model, the following may be performed:
- the electronic device uses each two-dimensional color live human face sample image and each two-dimensional color non-live human face sample image as training inputs, and uses the first depth image corresponding to each two-dimensional color live human face sample image and each two-dimensional color non-live body
- the second depth image corresponding to the face sample image is used as the target output, and the supervised model training is performed to obtain the depth estimation model.
- the electronic device uses multiple acquired two-dimensional color live human face sample images and corresponding multiple first depth images, and multiple two-dimensional color non-live human face samples
- multiple acquired two-dimensional color live human face sample images and their corresponding multiple first depth images, and multiple second depth images can also be used
- Dimensional color non-living human face sample images and corresponding multiple second depth images are used to train a depth estimation model.
- the electronic device can directly use each two-dimensional color live human face sample image and each two-dimensional color non-live human face sample image as training inputs, and use the first depth image and each two corresponding to each two-dimensional color live human face sample image
- the second depth image corresponding to the dimensional color non-living face sample image is used as the target output, and the supervised model is trained to obtain the depth estimation model.
- the electronic device uses the two-dimensional color live human face sample image as a training input, and uses the first depth image of the two-dimensional color live human face sample image as a corresponding target output ;
- the electronic device uses the two-dimensional color non-living human face sample image as a training input, and the two-dimensional color non-living human face sample image as the corresponding target output .
- FIG. 5 is a schematic structural diagram of a living body detection device according to an embodiment of the present application.
- the living body detection device is applied to an electronic device, the electronic device includes a monocular camera, the living body detection device includes a color image acquisition module 501, a depth image acquisition module 502, and a living face detection module 503, as follows:
- the color image acquisition module 501 is used to shoot a face to be detected through a monocular camera to obtain a two-dimensional color image of the face to be detected;
- the depth image acquisition module 502 is used to input the captured two-dimensional color image into a pre-trained depth estimation model to perform depth estimation to obtain a depth image corresponding to the two-dimensional color image;
- the living body face detection module 503 is used to input a two-dimensional color image and its corresponding depth image into a pre-trained living body detection model for living body detection to obtain a detection result.
- the living body detection model is a convolutional neural network model, which includes a convolution layer, a pooling layer, and a fully connected layer connected in sequence. After inputting a two-dimensional color image and its corresponding depth image into a pre-trained living body detection The model performs live detection, and when the detection result is obtained, the live face detection module 503 can be used to:
- the joint global feature will be input into the pooling layer for feature dimensionality reduction, and the joint global feature after dimensionality reduction will be obtained;
- the joint global features after dimensionality reduction are input into the fully connected layer for classification processing to obtain the detection result that the face to be detected is a living face, or the detection result that the face to be detected is a non-living face.
- the living face detection module 503 may Used for:
- the face area image in the two-dimensional color image and the face area image in the depth image are input to the convolutional layer for feature extraction to obtain a joint global feature of the two-dimensional color image and the depth image.
- the living body detection device further includes a model training module, which is used to:
- the multiple different live human faces are captured through the monocular camera to obtain multiple two-dimensional color live human face sample images, And obtain a depth image corresponding to each two-dimensional color live human face sample image to obtain multiple first depth images;
- a plurality of different non-living human face images are captured by a monocular camera to obtain multiple two-dimensional color non-living human face sample images, and a depth image corresponding to each two-dimensional color non-living human face sample image is obtained to obtain multiple second Depth image
- each two-dimensional color live human face sample image and its corresponding first depth image as a positive sample
- each two-dimensional color non-live human face sample image and its corresponding second depth image as a negative sample
- a convolutional neural network is used to model the training sample set, and a convolutional neural network model is obtained as a living body detection model.
- the model training module before the convolutional neural network is used to train the training sample set, the model training module:
- the model training module when acquiring depth images corresponding to each two-dimensional color live human face sample image to obtain multiple first depth images, the model training module may be used to:
- a depth image corresponding to each two-dimensional color live human face sample image is generated to obtain a plurality of first depth images.
- the model training module when acquiring depth images corresponding to each two-dimensional color non-living human face sample image to obtain multiple second depth images, the model training module may be used to:
- a depth image corresponding to each two-dimensional color non-living human face sample image is generated to obtain a plurality of second depth images.
- model training module can also be used for:
- each two-dimensional color live human face sample image and each two-dimensional color non-live human face sample image as training input, and use the first depth image corresponding to each two-dimensional color live human face sample image and each two-dimensional color non-live human face image
- the second depth image corresponding to the sample image is output as the target, and the supervised model is trained to obtain the depth estimation model.
- An embodiment of the present application provides a computer-readable storage medium on which a computer program is stored, and when the stored computer program is executed on a computer, causes the computer to perform the steps in the living body detection method provided in this embodiment, or The computer is caused to execute the steps in the model training method provided in this embodiment.
- the storage medium may be a magnetic disk, an optical disk, a read-only memory (Read Only Memory, ROM), or a random access device (Random Access Memory, RAM), and so on.
- An embodiment of the present application also provides an electronic device, including a memory, a processor, and the processor executes the steps in the living body detection method provided in this embodiment by calling a computer program stored in the memory, or executes the model as provided in this embodiment Steps in the training method.
- an electronic device is also provided.
- the electronic device includes a processor 701, a memory 702, and a monocular camera 703.
- the processor 701 is electrically connected to the memory 702 and the monocular camera 703.
- the processor 701 is the control center of the electronic device, and uses various interfaces and lines to connect the various parts of the entire electronic device, executes the electronic device by running or loading the computer program stored in the memory 702, and calling the data stored in the memory 702 Various functions and process data.
- the memory 702 may be used to store software programs and modules.
- the processor 701 runs computer programs and modules stored in the memory 702 to execute various functional applications and data processing.
- the memory 702 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, computer programs required by at least one function (such as a sound playback function, an image playback function, etc.), etc.; the storage data area may store Data created by the use of electronic devices, etc.
- the memory 702 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one magnetic disk storage device, a flash memory device, or other volatile solid-state storage devices. Accordingly, the memory 702 may further include a memory controller to provide the processor 701 with access to the memory 702.
- the monocular camera 703 may include a camera having one or more lenses and an image sensor, capable of capturing external image data.
- the processor 701 in the electronic device loads the instructions corresponding to the process of one or more computer programs into the memory 702 according to the following steps, and the processor 701 runs and stores the instructions in the memory 702 Computer program to achieve various functions as follows:
- the monocular camera 703 shoots the face to be detected to obtain a two-dimensional color image of the face to be detected;
- the two-dimensional color image and the corresponding depth image are input into a pre-trained living body detection model for living body detection, and the detection result is obtained.
- FIG. 7 is another schematic structural diagram of an electronic device provided by an embodiment of the present application.
- the electronic device further includes components such as an input unit 704 and an output unit 705.
- the input unit 704 can be used to receive input numbers, character information, or user characteristic information (such as fingerprints), and generate keyboard, mouse, joystick, optical, or trackball signal inputs related to user settings and function control.
- user characteristic information such as fingerprints
- the output unit 705 may be used to display information input by the user or information provided to the user, such as a screen.
- the processor 701 in the electronic device loads the instructions corresponding to the process of one or more computer programs into the memory 702 according to the following steps, and the processor 701 runs and stores the instructions in the memory 702 Computer program to achieve various functions as follows:
- the monocular camera 703 shoots the face to be detected to obtain a two-dimensional color image of the face to be detected;
- the two-dimensional color image and the corresponding depth image are input into a pre-trained living body detection model for living body detection, and the detection result is obtained.
- the living body detection model is a convolutional neural network model, which includes a convolution layer, a pooling layer, and a fully connected layer connected in sequence. After inputting a two-dimensional color image and its corresponding depth image into a pre-trained living body detection The model performs a living body test, and when the test result is obtained, the processor 701 can execute:
- the joint global feature will be input into the pooling layer for feature dimensionality reduction, and the joint global feature after dimensionality reduction will be obtained;
- the joint global features after dimensionality reduction are input into the fully connected layer for classification processing to obtain the detection result that the face to be detected is a living face, or the detection result that the face to be detected is a non-living face.
- the processor 701 may execute:
- the face area image in the two-dimensional color image and the face area image in the depth image are input to the convolutional layer for feature extraction to obtain a joint global feature of the two-dimensional color image and the depth image.
- the processor 701 may execute:
- the monocular camera 703 is used to photograph the face to be detected to obtain a two-dimensional color image of the face to be detected, the monocular camera 703 is used to photograph multiple different live human faces to obtain multiple two-dimensional color live human face samples Image, and obtain a depth image corresponding to each two-dimensional color live human face sample image to obtain multiple first depth images;
- a plurality of different non-living human face images are captured by the monocular camera 703 to obtain a plurality of two-dimensional color non-living human face sample images, and a depth image corresponding to each two-dimensional color non-living human face sample image is obtained to obtain multiple Two depth images;
- each two-dimensional color live human face sample image and its corresponding first depth image as a positive sample
- each two-dimensional color non-live human face sample image and its corresponding second depth image as a negative sample
- a convolutional neural network is used to model the training sample set, and a convolutional neural network model is obtained as a living body detection model.
- the processor 701 may execute:
- the processor 701 may execute:
- a depth image corresponding to each two-dimensional color live human face sample image is generated to obtain a plurality of first depth images.
- the processor 701 may execute:
- a depth image corresponding to each two-dimensional color non-living human face sample image is generated to obtain a plurality of second depth images.
- processor 701 may also execute:
- each two-dimensional color live human face sample image and each two-dimensional color non-live human face sample image as training input, and use the first depth image corresponding to each two-dimensional color live human face sample image and each two-dimensional color non-live human face image
- the second depth image corresponding to the sample image is output as the target, and the supervised model is trained to obtain the depth estimation model.
- the computer program may be stored in a computer-readable storage medium, such as stored in a memory of an electronic device, and executed by at least one processor in the electronic device, and may include, for example, a living body detection method during execution The process of the embodiment.
- the storage medium may be a magnetic disk, an optical disk, a read-only memory, a random access memory, etc.
- each functional module may be integrated into one processing chip, or each module may exist alone physically, or two or more modules may be integrated into one module.
- the above integrated modules can be implemented in the form of hardware or software function modules. If the integrated module is implemented in the form of a software functional module and sold or used as an independent product, it may also be stored in a computer-readable storage medium, such as a read-only memory, magnetic disk, or optical disk, etc. .
Landscapes
- Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Image Analysis (AREA)
Abstract
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811565579.0A CN109635770A (zh) | 2018-12-20 | 2018-12-20 | 活体检测方法、装置、存储介质及电子设备 |
CN201811565579.0 | 2018-12-20 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020125623A1 true WO2020125623A1 (fr) | 2020-06-25 |
Family
ID=66075992
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/125957 WO2020125623A1 (fr) | 2018-12-20 | 2019-12-17 | Procédé et dispositif de détection de corps vivant, support d'informations et dispositif électronique |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN109635770A (fr) |
WO (1) | WO2020125623A1 (fr) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111797745A (zh) * | 2020-06-28 | 2020-10-20 | 北京百度网讯科技有限公司 | 一种物体检测模型的训练及预测方法、装置、设备及介质 |
CN111914758A (zh) * | 2020-08-04 | 2020-11-10 | 成都奥快科技有限公司 | 一种基于卷积神经网络的人脸活体检测方法及装置 |
CN111985427A (zh) * | 2020-08-25 | 2020-11-24 | 深圳前海微众银行股份有限公司 | 活体检测方法、设备及可读存储介质 |
CN112069936A (zh) * | 2020-08-21 | 2020-12-11 | 深圳市商汤科技有限公司 | 攻破点测试方法及相关装置、电子设备、存储介质 |
CN112183357A (zh) * | 2020-09-29 | 2021-01-05 | 深圳龙岗智能视听研究院 | 一种基于深度学习的多尺度活体检测方法及系统 |
CN112200057A (zh) * | 2020-09-30 | 2021-01-08 | 汉王科技股份有限公司 | 人脸活体检测方法、装置、电子设备及存储介质 |
CN112434647A (zh) * | 2020-12-09 | 2021-03-02 | 浙江光珀智能科技有限公司 | 一种人脸活体检测方法 |
CN112699811A (zh) * | 2020-12-31 | 2021-04-23 | 中国联合网络通信集团有限公司 | 活体检测方法、装置、设备、储存介质及程序产品 |
CN113378715A (zh) * | 2021-06-10 | 2021-09-10 | 北京华捷艾米科技有限公司 | 一种基于彩色人脸图像的活体检测方法及相关设备 |
CN113542527A (zh) * | 2020-11-26 | 2021-10-22 | 腾讯科技(深圳)有限公司 | 一种人脸图像传输方法、装置、电子设备及存储介质 |
CN114550274A (zh) * | 2022-03-16 | 2022-05-27 | 北京人人云图信息技术有限公司 | 一种基于面具人脸检测的人脸欺诈识别方法 |
US20230112452A1 (en) * | 2020-04-16 | 2023-04-13 | Samsung Electronics Co., Ltd. | Method and apparatus for testing liveness |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109635770A (zh) * | 2018-12-20 | 2019-04-16 | 上海瑾盛通信科技有限公司 | 活体检测方法、装置、存储介质及电子设备 |
CN110245645B (zh) * | 2019-06-21 | 2021-06-08 | 北京字节跳动网络技术有限公司 | 人脸活体识别方法、装置、设备及存储介质 |
CN110334628B (zh) * | 2019-06-26 | 2021-07-27 | 华中科技大学 | 一种基于结构化随机森林的室外单目图像深度估计方法 |
CN110674759A (zh) * | 2019-09-26 | 2020-01-10 | 深圳市捷顺科技实业股份有限公司 | 一种基于深度图的单目人脸活体检测方法、装置及设备 |
CN111091063B (zh) * | 2019-11-20 | 2023-12-29 | 北京迈格威科技有限公司 | 活体检测方法、装置及系统 |
CN113642466B (zh) * | 2019-11-27 | 2022-11-01 | 马上消费金融股份有限公司 | 活体检测和模型训练方法、设备及介质 |
CN112861586B (zh) * | 2019-11-27 | 2022-12-13 | 马上消费金融股份有限公司 | 活体检测、图像分类和模型训练方法、装置、设备及介质 |
CN111191521B (zh) * | 2019-12-11 | 2022-08-12 | 智慧眼科技股份有限公司 | 人脸活体检测方法、装置、计算机设备及存储介质 |
CN111046845A (zh) * | 2019-12-25 | 2020-04-21 | 上海骏聿数码科技有限公司 | 活体检测方法、装置及系统 |
TWI722872B (zh) * | 2020-04-17 | 2021-03-21 | 技嘉科技股份有限公司 | 人臉辨識裝置以及人臉辨識方法 |
CN113553887A (zh) * | 2020-04-26 | 2021-10-26 | 华为技术有限公司 | 一种基于单目摄像头的活体检测方法、设备和可读存储介质 |
CN111753658A (zh) * | 2020-05-20 | 2020-10-09 | 高新兴科技集团股份有限公司 | 一种睡岗告警方法、装置和计算机设备 |
CN112036331B (zh) * | 2020-09-03 | 2024-04-09 | 腾讯科技(深圳)有限公司 | 活体检测模型的训练方法、装置、设备及存储介质 |
CN112115831B (zh) * | 2020-09-10 | 2024-03-15 | 深圳印像数据科技有限公司 | 活体检测图像预处理方法 |
CN112270303A (zh) * | 2020-11-17 | 2021-01-26 | 北京百度网讯科技有限公司 | 图像识别方法、装置以及电子设备 |
CN112508812B (zh) * | 2020-12-01 | 2024-08-27 | 厦门美图之家科技有限公司 | 图像色偏校正方法、模型训练方法、装置及设备 |
CN113435408A (zh) * | 2021-07-21 | 2021-09-24 | 北京百度网讯科技有限公司 | 人脸活体检测方法、装置、电子设备及存储介质 |
CN113705428B (zh) * | 2021-08-26 | 2024-07-19 | 北京市商汤科技开发有限公司 | 活体检测方法及装置、电子设备及计算机可读存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180157938A1 (en) * | 2016-12-07 | 2018-06-07 | Samsung Electronics Co., Ltd. | Target detection method and apparatus |
CN108876833A (zh) * | 2018-03-29 | 2018-11-23 | 北京旷视科技有限公司 | 图像处理方法、图像处理装置和计算机可读存储介质 |
CN109034102A (zh) * | 2018-08-14 | 2018-12-18 | 腾讯科技(深圳)有限公司 | 人脸活体检测方法、装置、设备及存储介质 |
CN109635770A (zh) * | 2018-12-20 | 2019-04-16 | 上海瑾盛通信科技有限公司 | 活体检测方法、装置、存储介质及电子设备 |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005215750A (ja) * | 2004-01-27 | 2005-08-11 | Canon Inc | 顔検知装置および顔検知方法 |
JP6027070B2 (ja) * | 2014-09-24 | 2016-11-16 | 富士フイルム株式会社 | 領域検出装置、領域検出方法、画像処理装置、画像処理方法、プログラムおよび記録媒体 |
GB201508074D0 (en) * | 2015-05-12 | 2015-06-24 | Apical Ltd | People detection |
US9691152B1 (en) * | 2015-08-14 | 2017-06-27 | A9.Com, Inc. | Minimizing variations in camera height to estimate distance to objects |
CN107871134A (zh) * | 2016-09-23 | 2018-04-03 | 北京眼神科技有限公司 | 一种人脸检测方法及装置 |
CN108171204B (zh) * | 2018-01-17 | 2019-09-17 | 百度在线网络技术(北京)有限公司 | 检测方法和装置 |
CN108537152B (zh) * | 2018-03-27 | 2022-01-25 | 百度在线网络技术(北京)有限公司 | 用于检测活体的方法和装置 |
CN108764024B (zh) * | 2018-04-09 | 2020-03-24 | 平安科技(深圳)有限公司 | 人脸识别模型的生成装置、方法及计算机可读存储介质 |
CN108960127B (zh) * | 2018-06-29 | 2021-11-05 | 厦门大学 | 基于自适应深度度量学习的遮挡行人重识别方法 |
CN108898112A (zh) * | 2018-07-03 | 2018-11-27 | 东北大学 | 一种近红外人脸活体检测方法及系统 |
CN109003297B (zh) * | 2018-07-18 | 2020-11-24 | 亮风台(上海)信息科技有限公司 | 一种单目深度估计方法、装置、终端和存储介质 |
-
2018
- 2018-12-20 CN CN201811565579.0A patent/CN109635770A/zh active Pending
-
2019
- 2019-12-17 WO PCT/CN2019/125957 patent/WO2020125623A1/fr active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180157938A1 (en) * | 2016-12-07 | 2018-06-07 | Samsung Electronics Co., Ltd. | Target detection method and apparatus |
CN108876833A (zh) * | 2018-03-29 | 2018-11-23 | 北京旷视科技有限公司 | 图像处理方法、图像处理装置和计算机可读存储介质 |
CN109034102A (zh) * | 2018-08-14 | 2018-12-18 | 腾讯科技(深圳)有限公司 | 人脸活体检测方法、装置、设备及存储介质 |
CN109635770A (zh) * | 2018-12-20 | 2019-04-16 | 上海瑾盛通信科技有限公司 | 活体检测方法、装置、存储介质及电子设备 |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20230112452A1 (en) * | 2020-04-16 | 2023-04-13 | Samsung Electronics Co., Ltd. | Method and apparatus for testing liveness |
US11836235B2 (en) * | 2020-04-16 | 2023-12-05 | Samsung Electronics Co., Ltd. | Method and apparatus for testing liveness |
CN111797745A (zh) * | 2020-06-28 | 2020-10-20 | 北京百度网讯科技有限公司 | 一种物体检测模型的训练及预测方法、装置、设备及介质 |
CN111914758A (zh) * | 2020-08-04 | 2020-11-10 | 成都奥快科技有限公司 | 一种基于卷积神经网络的人脸活体检测方法及装置 |
CN112069936A (zh) * | 2020-08-21 | 2020-12-11 | 深圳市商汤科技有限公司 | 攻破点测试方法及相关装置、电子设备、存储介质 |
CN111985427A (zh) * | 2020-08-25 | 2020-11-24 | 深圳前海微众银行股份有限公司 | 活体检测方法、设备及可读存储介质 |
CN112183357A (zh) * | 2020-09-29 | 2021-01-05 | 深圳龙岗智能视听研究院 | 一种基于深度学习的多尺度活体检测方法及系统 |
CN112183357B (zh) * | 2020-09-29 | 2024-03-26 | 深圳龙岗智能视听研究院 | 一种基于深度学习的多尺度活体检测方法及系统 |
CN112200057A (zh) * | 2020-09-30 | 2021-01-08 | 汉王科技股份有限公司 | 人脸活体检测方法、装置、电子设备及存储介质 |
CN112200057B (zh) * | 2020-09-30 | 2023-10-31 | 汉王科技股份有限公司 | 人脸活体检测方法、装置、电子设备及存储介质 |
CN113542527B (zh) * | 2020-11-26 | 2023-08-18 | 腾讯科技(深圳)有限公司 | 一种人脸图像传输方法、装置、电子设备及存储介质 |
CN113542527A (zh) * | 2020-11-26 | 2021-10-22 | 腾讯科技(深圳)有限公司 | 一种人脸图像传输方法、装置、电子设备及存储介质 |
CN112434647A (zh) * | 2020-12-09 | 2021-03-02 | 浙江光珀智能科技有限公司 | 一种人脸活体检测方法 |
CN112699811B (zh) * | 2020-12-31 | 2023-11-03 | 中国联合网络通信集团有限公司 | 活体检测方法、装置、设备、储存介质及程序产品 |
CN112699811A (zh) * | 2020-12-31 | 2021-04-23 | 中国联合网络通信集团有限公司 | 活体检测方法、装置、设备、储存介质及程序产品 |
CN113378715A (zh) * | 2021-06-10 | 2021-09-10 | 北京华捷艾米科技有限公司 | 一种基于彩色人脸图像的活体检测方法及相关设备 |
CN113378715B (zh) * | 2021-06-10 | 2024-01-05 | 北京华捷艾米科技有限公司 | 一种基于彩色人脸图像的活体检测方法及相关设备 |
CN114550274A (zh) * | 2022-03-16 | 2022-05-27 | 北京人人云图信息技术有限公司 | 一种基于面具人脸检测的人脸欺诈识别方法 |
Also Published As
Publication number | Publication date |
---|---|
CN109635770A (zh) | 2019-04-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020125623A1 (fr) | Procédé et dispositif de détection de corps vivant, support d'informations et dispositif électronique | |
US11645506B2 (en) | Neural network for skeletons from input images | |
US11783639B2 (en) | Liveness test method and apparatus | |
Kumar et al. | Face detection techniques: a review | |
US10002313B2 (en) | Deeply learned convolutional neural networks (CNNS) for object localization and classification | |
US20220172450A1 (en) | Depth-based object re-identification | |
WO2021043168A1 (fr) | Procédé d'entraînement de réseau de ré-identification de personnes et procédé et appareil de ré-identification de personnes | |
US20190279052A1 (en) | Image recognition method and apparatus, image verification method and apparatus, learning method and apparatus to recognize image, and learning method and apparatus to verify image | |
WO2019227479A1 (fr) | Procédé et appareil de génération d'image de rotation de visage | |
WO2017088432A1 (fr) | Procédé et dispositif de reconnaissance d'image | |
Yang et al. | Facial expression recognition based on dual-feature fusion and improved random forest classifier | |
CN111183455A (zh) | 图像数据处理系统与方法 | |
WO2019128564A1 (fr) | Procédé de mise au point, appareil, support de stockage, et dispositif électronique | |
CN114339054B (zh) | 拍照模式的生成方法、装置和计算机可读存储介质 | |
JP2019185127A (ja) | 多層ニューラルネットワークの学習装置およびその制御方法 | |
WO2021217919A1 (fr) | Procédé et appareil de reconnaissance d'unité d'action faciale, dispositif électronique et support de stockage | |
Asadi-Aghbolaghi et al. | Supervised spatio-temporal kernel descriptor for human action recognition from RGB-depth videos | |
Liu | RETRACTED ARTICLE: Video Face Detection Based on Deep Learning | |
JP7360217B2 (ja) | ユーザのバイオメトリック特性を有する当該ユーザの物体の画像からデータを取得するための方法 | |
US20220277579A1 (en) | Clustered dynamic graph convolutional neural network (cnn) for biometric three-dimensional (3d) hand recognition | |
Mohana et al. | Facial Expression Recognition Using Machine Learning and Deep Learning Techniques: A Systematic Review | |
Cui et al. | Improving the face recognition system by hybrid image preprocessing | |
WO2023109551A1 (fr) | Procédé et appareil de détection d'un corps vivant et dispositif informatique | |
Srinivas et al. | E-CNN-FFE: An Enhanced Convolutional Neural Network for Facial Feature Extraction and Its Comparative Analysis with FaceNet, DeepID, and LBPH Methods | |
CN113128289B (zh) | 人脸识别的特征提取计算方法及设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19901333 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19901333 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS (EPO FORM 1205A DATED 23.12.2021) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19901333 Country of ref document: EP Kind code of ref document: A1 |