Nothing Special   »   [go: up one dir, main page]

WO2002009024A1 - Identity systems - Google Patents

Identity systems Download PDF

Info

Publication number
WO2002009024A1
WO2002009024A1 PCT/GB2001/003322 GB0103322W WO0209024A1 WO 2002009024 A1 WO2002009024 A1 WO 2002009024A1 GB 0103322 W GB0103322 W GB 0103322W WO 0209024 A1 WO0209024 A1 WO 0209024A1
Authority
WO
WIPO (PCT)
Prior art keywords
facial
subject
facial identification
identification matrix
biometric
Prior art date
Application number
PCT/GB2001/003322
Other languages
French (fr)
Inventor
Michael Leslie Sim
Mark Simon Yadegar
Original Assignee
Bio4 Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Bio4 Limited filed Critical Bio4 Limited
Priority to AU2001272667A priority Critical patent/AU2001272667A1/en
Publication of WO2002009024A1 publication Critical patent/WO2002009024A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/243Image signal generators using stereoscopic image cameras using three or more 2D image sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/221Image signal generators using stereoscopic image cameras using a single 2D image sensor using the relative movement between cameras and objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/239Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/286Image signal generators having separate monoscopic and stereoscopic modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/246Calibration of cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0081Depth or disparity estimation from stereoscopic image signals

Definitions

  • FIM Facial Identification Matrix
  • Biometrics is the science of automatically identifying individuals based on their unique physiological and/or behavioural characteristics. Biometric information based on the unique characteristics of a person's face, iris, voice, fingerprints, signature, palm prints, hand prints, or DNA can all be used to authenticate a person's identity or establish an identity from a database.
  • facial biometric information has a number of advantages. Existing methods have a good level of reliability and the information can be obtained quickly using non-intrusive techniques. Furthermore, the data is not sensitive to superficial facial features such as sunglasses and beards; primarily because the system can undertake training of the data whilst processing the image.
  • One of the major methods of the first generation of facial biometric software based on "neural net” creates a template from a two-dimensional (2D) image of the subjects face; although this can be very accurate, both the angle at which the image is captured and the nature of the lighting are critical. Problems also arise with individuals having low skin contrast levels (e.g. people of Afro-Caribbean origin).
  • the invention allows for images captured to be rendered into 3D models, and depending on the number of cameras used, the image can be accurately rotated through 360 degrees.
  • Our invention makes use of a 3D Biometric Engine [The 3D Biometric Engine includes the following modules; 3D Patch Engine, 3D Graph Engine, 3D Feature Selector Engine, 3D Indexing Engine and a control module which controls the modules and the input and output functions of the system.].
  • This takes data from a 3D camera system which is passed to both a 2D and separate 3D Biometric Engine; the 3D engine enables a correlated index key to be created.
  • the present invention provides a process for producing a FIM [A FIM contains both 2D and 3D templates, indexing data and optionally a copy of the original captured.] which is more accurate than any currently existing biometric identifier.
  • the present invention proposes a methodology for producing a FIM from a subject which includes:
  • the invention provides the ability to detect motion at minute levels, down to the level of 3 to 4 pixels, for example the partial blink of an eye-lid.
  • the invention is more accommodating for poor lighting conditions and weak skin contrast levels, the range of angles through which a successful image can be captured is increased through the ability to correct angular displacement.
  • the use of the two different kinds of biometric data significantly increases the overall accuracy of identification when measured against using either of the biometric methods individually.
  • the accuracy of the matrix-generating process can be improved by obtaining a plurality (verification set) of facial images from the subject, generating the two kinds of biometric information for each successive image, and comparing each new set of information with the information previously generated until a predetermined correlation level is achieved.
  • the FIM may be stored together with the facial images which were used to generate it.
  • the invention further provides a FIM which contains both a 3D facial biometric template and a 2D biometric template for a subject.
  • the invention using the FIM, further provides a process for determining the identity of a subject or for verifying the subject against a known FIM.
  • the process compares the matrix (thereby obtained through the image process) with a stored FIM which contains 3D facial biometric data and a 2D biometric template derived there from, to determine whether the two matrixes match.
  • the identification matrix is small enough to be written to portable identification cards, added to a database along with numerous similar matrixes or transmitted electronically.
  • the stored matrix may be included on a portable data carrier such as an identity card, a 2D Barcode printout, smart card or any portable media capable of storing data.
  • the verification process can be used to verify that the data carrier belongs to the individual presenting it.
  • the accuracy of the verification process can be improved by repeatedly generating identification matrixes from the subject and comparing them with the stored matrix until a predetermined correlation level is achieved.
  • the stored FIM can be found from a database of similar matrixes using the index key value of the FIM.
  • the identity of an unknown individual can be ascertained by first finding all the matching index values which equal the newly generated FIM and then comparing each of the stored matrixes until a best match is found.
  • the invention allows for duplicate index key values.
  • FIG. 0 - This is a high level overview of the main modules forming the core of the invention.
  • FIG. 1 A high level diagrammatic representation of the process for producing a FIM in accordance with the invention
  • Figure 2 A high level diagrammatic representation of the process for verifying the identity of a subject using the matrix.
  • Figure 3 This is a high level diagrammatic representation of the process for identifying a subject using the FIM recalled from storage.
  • a standard PC 750 MHz processor, 128 MB of RAM, 10 gigabyte hard drive, 15 inch CRT or TFT LCD screen, keyboard, mouse, video Grabber Card
  • Microsoft Windows 2000 Professional Operating System System requires a secure O/S
  • an approved video camera for the low-end systems a USB Video Web Camera can be utilised
  • lighting units as specified. (Basically the lights illuminate the subject with light which is outside of the spectrum visible to a human being.)
  • an optional control device such as a touch screen may be utilised. If Process Step 7 (as described below) is implemented then a device, such as a smart card reader or a printer capable of outputting 2D barcode would be required.
  • the enrolment process involves the following steps:
  • Image capture This is the act of obtaining a digital image of the subject as a direct result of an input into the system from either a digital video camera or a Web camera. Depending on which methodology is used to obtain the third coordinate (required for the 3D biometric engine) either a single camera is used or a stereoscopic camera system. For details of the two different methodologies which are available see the entry below on pseudo 3D image calculation and stereoscopic 3D image. Whilst the image capture is being processed the system will normally display to the subject a copy of the image capture, however, in certain circumstances this will not be the case. Both the input from the Web Camera and the video camera provide a number of frames which are available for capture and utilisation by the system.
  • the process monitors the quality of the image and assigns a value which relates to the acceptableness of the image quality provided. In the event that the image quality is adversely affected by strong peripheral lighting the subject would be advised of the problem.
  • a series of "live" images are preferably obtained from the subject from whom successive pairs of biometric identifiers are created as described. The newly formed identifiers are compared to previously stored identifiers until such time as a perfect match is achieved or the verification level reaches a preset figure. The image can be taken at various angles covering the front of the subject.
  • Step 2D and 3D images are created separately in order that they may be passed to the appropriate biometric engine.
  • the 3D image will be calculated from a single camera system or a 2D image will be calculated from the images provided by the stereoscopic system.
  • Process Step 3 The 2D data is passed to the 2D biometric engine which performs a number of tasks and results in the 2D biometric template.
  • Process Step 4 The 3D data is passed to the 3D biometric engine which performs a number of tasks and results in the creation of a 3D biometric template.
  • Process Step 5 Indexing, at this point data from the 3D biometric engine is used to create an index value which has a specific correlation to the features of the subject involved. [The system allows for a duplicated index value, for example, identical twins, who are truly identical, may well produce a calculated value which in itself is identical.] At this point the two sets of data (2D and 3D templates) linking the index values are encapsulated together, this data set is known as a FIM.
  • Process Step 6 The FIM produced is appended to a database.
  • Process Step 7. (Optional) If required the FIM calculated can be output directly to a portable data storage product.
  • the two best biometric identifiers are then converted into a binary format which is suitable for storage. These biometric identifiers are then stored as a FIM, along with the "best match" image of the subject and any other desired data. Since the size of the matrix is remarkably small it can be recorded on a plastic or paper identity card, e.g. airline boarding pass or a company ID card, a smart card or any suitable carrier.
  • the matrix could, for example, be incorporated into a barcode in PDF417 format (capacity about 1200 bytes), a USB token, a memory stick or similar binary recording means. In the case of a barcode the information can be printed using a laser printer for paper/cardboard or a commercial 150 dpi thermal printer for plastic cards. Other biometric, image or reference data may be included on the card as desired.
  • the FIM may be transmitted rapidly by any communication means, for example Blue tooth fax or email, to enable the FIM to be used at a remote location.
  • the verification process involves the following steps:
  • Process Step 1 The FIM or unique ID of the person to be verified is entered into the system, the method and type of input device by which data is entered into the system can vary considerably for example; PDF417 (2D barcode), smart cards, smart keys, proximity card, keyboard entry or even a floppy disk could be the means of triggering the system to verify the subject.
  • the FIM does not actually have to be entered into the system, the input to the system only needs to identify the location of the FIM, it is required that the location of the FIM is available to the system. If a touch screen facility is provided the image may be displayed on the touch screen so that the required facial image can be selected from a group of people by touching the appropriate face.
  • Process Step 2 The 2D and 3D images are acquired as per step 1 of the enrolment. Depending on the application either a single or multi camera system would be utilised.
  • Step 3 Image separation.
  • the 2D and 3D images are created separately in order that they may be passed to the appropriate biometric engine.
  • the 3D data will be calculated from the single image (provided from a one camera system) or 2D data will be calculated from the images provided by the stereoscopic system.
  • the 2D data is passed to the 2D biometric engine which performs a number of tasks resulting in the creation of a 2D biometric template.
  • the 3D data is passed to the 3D biometric engine which performs a number of tasks and results in the creation of a 3D biometric template.
  • Process Step 4 The 2D and 3D biometric engines are used to compare the data held in the FIM against the newly acquired image.
  • the user as part of the system setup, can set a predefined threshold with regard to the accuracy of the system; this value is used in determining whether there is an acceptable match between the two sets of data.
  • Process Step 5 Dependent on the results of Process Step 4, an action or series of actions could be implemented, for example a trigger signal to open a door release.
  • Process Step 6 In most circumstances a message will be output indicating the results of the verification. This might be through the means of a message displayed on a computer screen, an audible response or the illumination of a sign.
  • the system can record the details of the verification process which has just been enacted in the form of an Audit Trail. This ability enables the verification process to form a major part of an automated Time and Attendance system.
  • the system can also run a training module which updates the FIM with the latest acquired image of the subject. In the case of a failed verification or where the input device failed a particular test (valid date) the system may withhold the return of the input device or overwrite the device effectively canceling its operational use.
  • Process Step 1 The 2D and 3D images are acquired as per step 1 of the enrolment. Depending on the application either a single or multi camera system would be utilised.
  • Process Step 2 Image separation, at this point the 2D and 3D images are created separately in order that they may be passed to the appropriate biometric engine.
  • the 3D data will be calculated from the single image (provided from a one camera system) or 2D data will be calculated from the images provided by the stereoscopic system.
  • the 2D data is passed to the 2D biometric engine which performs a number of tasks and results in the 2D biometric template.
  • the 3D data is passed to the 3D biometric engine which performs a number of tasks and results in the creation of a 3D biometric template.
  • Process Step 4 Using the key value generated at Process Step 3 the system recovers all FIM's with that specific key value.
  • the 2D and 3D biometric engines are used to compare the data held in the returned FIM's against the FIM of the newly acquired image.
  • the user as part of the system setup, can set a predefined threshold with regard to the accuracy of the system; this value is used in determining whether there is an acceptable match between the two sets of data.
  • Process Step 5 Dependent on the results of Process Step 4, an action or series of actions could be implemented, for example a trigger signal needed to open the door release. In most circumstances a message will be output indicating the results of the verification. This might be through the means of a message displayed on a computer screen, an audible response or the illumination of a sign.
  • the system can record the details of the identification process which has just been enacted to form an Audit Trail.
  • the system can also run a training module which updates the FIM with the latest acquired image of the subject thus helping to maintain an accurate collection of images of the subject.
  • the system may withhold the return of the input device or overwrite the device effectively canceling its operational use.
  • the enrolment and verification processes are not restricted to live CCTVfor image capture.
  • Digital images can be acquired by means of digital still cameras, laptop computers with camera attachments, i.e. a Web camera, a PDA hand held barcode scanner with an image capture head, scanned from printed documents or photographs, camera-enabled WAP phones, etc. Images when acquired by these means would normally result in the 2D biometric process only, being applied.
  • Uses of the identification method include travel documents, banking, healthcare, social security, passports and immigration, education, the prison service, ATM machines, retail, secure access, document security, internet services and identification of football fans.
  • the figure 4a shows a single camera system where a projector is used to illuminate the subject with a calibrated grid; the frequency of illumination is outside that of a person's ability to register.
  • the subject is further illuminated by a balanced light source (i.e. light falling of each side of the subjects face is balance to reflect an equal level of light) providing light of a wavelength outside the visible spectrum.
  • the camera is capable of switching, via software control, between colour and the wavelength(s) of the emitters used in the projector and lamps.
  • Figure 4b uses two cameras the angle between each camera is known and fixed; the lighting characteristics are the same as for figure 4a. The images from both cameras are fed simultaneously to the computer system.
  • At figure 4c multiple cameras are used, at least one pair of cameras will be used as described at figure 4b, the additional camera or cameras will normally be used to take colour images of the subject in addition to providing extra calibration information for the stereoscopic imaging process.
  • a single camera is used, meeting the range of wavelength sensitivity for the lighting and camera conditions as at figure 4a, the subject is partially surrounded by tracking which will enable the camera to track at high speed from one side of the subject to the other taking a number of images at predetermined points along the location of the tracking.
  • the lighting fix does not have to be applied in every situation, if a constant level of light is available then the lighting fix may not need to be applied.
  • Image capture is the process of obtaining a digital image of the subject normally as a direct result of an input into the system from either a digital video camera or a Web camera. Either a single camera or a stereoscopic/multi camera system is used (applying the appropriate methodology) to obtain the third coordinate required for the 3D biometric engine.
  • a framing box is positioned around the face, in order to achieve the maximum clarity of the image either a telescopic lens is used to zoom in on the subject or the image is digital enlarged to fill the framing box.
  • Process Step 3 The main process, see steps 3a and 3b below, like a software 'Read ahead' operation, sets an initial value which the main process can use as a comparator.
  • Process Step 3a The captured image is passed for processing by the 2D biometric engine; the first template created is a master.
  • Process Step 3b The template is passed back to the application including an assessment score and size of the template.
  • Process Step 3 (Repeated). Further n images of the subject are taken (the actual number selected is determined in the program setup, for example 10 images may be taken of which only 5 are used in the process) The images are passed to the 2D biometric engine in order to determine that they are (a) valid template material (b) the score obtained is of an acceptable value (c) and the size of template is acceptable.
  • Process Step 4 The templates are sorted in order, based on their quality and their size; a predetermined number of templates are selected to be encapsulated into the FIM.
  • Process Step 5 The selected templates are used to create a 2D person template which is then written to the application database.
  • Process Step 1 Projection using non-visible light of a calibration grid
  • the target subject is illuminated using a wavelength of the spectrum, which is not visible to the naked eye. Included within the projected light is a calibration grid (horizontal and vertical lines at a known distance), which through the use of filters and an appropriate camera device which can detect the grid, are captured along with the image.
  • the camera used must have both a high-resolution lens and image capture element (Charge Coupled Device (CCD) or Low Noise CMOS) element and must also be able to see the wavelength of light projected.
  • CCD Charge Coupled Device
  • CMOS Low Noise CMOS
  • Process Step 2 The captured image contains the target subject overlaid with the lines of the grid, because the contours of the surface where the horizontal and vertical lines of the calibration grid are overlaid over the subject's image the surface contour is deformed (appears to bend).
  • Process Step 3 Preparation of image data.
  • two images are required; a standard image incorporating colour information and a calibrated image which incorporates the standard image overlaid with the calibration grid.
  • Process Step 4 Linear scan of image data.
  • the standard image overlaid with the information from the generated calibration grid is then placed in 3D space (R3), this gives the effect of a plane in 3D space (with all z co-ordinates set to zero), note that the calibration grid has the same resolution as the projected calibration grid.
  • a linear scan of the image can now be performed; the process involves scanning along each of the horizontal calibration grid lines, from left to right, top to bottom. For each pass of the scan, the generated calibration grid is adjusted, at each intersecting point on the grid the corresponding picture element is checked in the image which incorporates both the image and calibration grid. If the pixel information reveals a colour, which is part of the calibrated grid itself the next scan is processed, otherwise the point referenced needs to be adjusted until all checks have been processed.
  • Process Step 5 Applying image texture data.
  • a specific area of interest (nose, mouth etc.) in the standard image can be used to map across the surface area of the patch, normally however the area of the face (forehead down to the bottom of the chin) would be used.
  • Values for u and v now range from 0.0 through to 1.0 these values can now be scaled to the device co-ordinates, however most 3D texture mapping hardware uses unit values ranging from 0.0 to 1.0.
  • Figure 7 shows the layout of a pair of camera, for use in stereoscopic image capture.
  • the principles on how this process functions are well known (see Mori, K., Kidode, M & Asada, H. (1973). An iterative prediction and correction method for automatic stereo comparison. Comp. Graph. Image Process).
  • the figure indicates how two cameras can be used to determine depth information of a subject. By taking an image point from camera 1 (the point x1,y1) and searching for the same point in the image capture by camera 2 (the point x2,y2) and using image block cross-correlation it is possible to calculate the depth of this point in the real world.
  • the stereo search will have two dimensions (x2,y2). Whilst there are different methods for calculating depth information, the use of the block cross-correlation method is preferred (even though costly in processor time) as it provides a more reliable way of obtaining the z1 values.
  • the information obtained is used to build up a data set of the x,y,z coordinates for use by other modules within the application.
  • Process Step 1 Obtain Working Resolution.
  • the size of the patch array (resolution) is obtained directly from the resolution provided by the image capture device; effectively this establishes the working parameters of the patch.
  • the unit scale of the patch is 2 by 2 (represented as a true mathematical model).
  • a scaling factor can be applied creating a true (real world) scale.
  • the minimum working resolution will be 90 units thus the scale is 90/2, which is 45 units.
  • Process Step 2 Initialise Patch Variables. Two variables are required in order to specify the points in the patch these are commonly known as u and v. The variables u and v are the working variables required for the generation of the 3D co-ordinates. At initialisation both the u and v variables are initialised to the value minus one, this representing the top left co-ordinate of the patch. The top left coordinate serves as the starting point for the linear scan process. Further variables are defined in order for the process to be completed.
  • Process Step 3 Generation of a Patch Constraint Points.
  • This process generates 3D point components, where the x and y components are calculated and the third component (z) is set to the value zero.
  • the patch at this stage represents a flat plane in 3D space, standing upright spanning across the Y axis, with a width that spans across the X axis.
  • This process also serves as an initialisation process for the patch array.
  • Process Step 4. Import of constraint data. Constraint information can be imported into the process from the data output by the 3D image capture engine in the form of 3D co-ordinates. The 3D coordinates now act as constraint points to be used in the generation of the curves; this provides the ability to automate the process of patch curve generation.
  • Process Step 5 Generation of Patch Curves.
  • the generation of patch curves is accomplished by taking a collection of constraint points and calculating each part of the curve.
  • the smoothness of the curve is variable depending on the level smoothness required. This process is applied across the u array of constraint points with the results feeding back into the patch array on completion of the u array being filled the process is repeated for the v array of constraint points.
  • the final result is a grid of curves that breaks the patch down into quadrants with all curves being inter-connecting.
  • Process Step 6 Storage of 3D template information via a database.
  • the final process is the creation of a 3D template, which is in the form of a FIM 3D Class; this is then stored in a database. Information held in the database can be recalled to be rendered on the display device.
  • the application module has the ability to rotate the model about a central axis +/- 15 degrees.
  • the graph-matching module has been developed at the University of York (UK); it is an advanced system for finding solutions to constraint-based problems.
  • the solution prevents the exponential complexity of the problem by using methods based on that originally proposed by Hancock and Kittler.
  • This method was extended and developed as neural relaxation by Austin and Turner, further development introduced correlation matrix memories as a calculation engine for the system.
  • This technique and its extensions are embodied in the Advanced Uncertain Reasoning Architecture (AURA) graph matching engine (from the University of York) versions 1.0 onwards.
  • AURA Advanced Uncertain Reasoning Architecture
  • the engine is use to delimit a face from the surrounding background and for defining the key features of the face, for example eyes, mouth, nose etc.
  • This element of the invention uses a face / facial characteristics detection algorithm, which is able to handle a wide variety of variations in colour images; the engine also makes of pixel depth data from information supplied by the Patch Engine.
  • the uses of colour information can simplify the task of finding faces in different environmental lighting conditions.
  • the process for locating facial images/facial characteristics is shown below:
  • Modelling skin tone colour involves isolating the skin colour space and using the cluster associated with the skin colour in this space.
  • ROI Region of Interest
  • a 3D Depth filter to aid the segmentation process. It follows that the reverse is also true and that the skin tone information can be mapped to the Patch Engine output as a fourth component.
  • the invention also applies the method of the colour difference between adjacent pixels, proportional to the original image colour difference.
  • ROI detection relies on skin tones, which allow (eye, nose etc) areas of interest to be processed. This process does not isolate a face in a scene but isolates areas of interest.
  • TRUE positives with FALSE positives and various other percentage results, a probability face graph can be constructed and a decision threshold implemented.
  • a neural net is now trained with results taken from the ROI process; both FALSE and TRUE results are used, resulting in face detection in a scene.
  • the index module works with information provided from several modules, the major ones being the delimitation of facial features and the 3D depth calculations. Common facial features, for example eyes, eyes to ear edge, nose to closed-jaw, depth of eyebrows etc can be used as elements of a compound index. The element values would be a normalized set of data based on five or more images of the subject.
  • the delimitation of facial features module has to determine whether a true element value is used or if there is a concern regarding accuracy of the element then whether an average value should be substituted. The more elements used the greater the ability to separate out individuals and the lower the risk of using average values for some of the elements in the key. Each element has a one bit Boolean flag to indicate whether a real or substituted value has been supplied to the element.
  • L.Eye (205); R.Eye (204); Nose to Closed-Jaw (650); Eye to Eye (200) the key would like 205b204b650B200b, 650 is a substituted value.
  • Using 30 or more elements creates a highly selective and searchable index; identification can be against any part(s) of the key or the whole key.
  • the invention will identify a subject, relative to a specific camera which is mounted at a known position in the environment, with the subjects face being at various angles.
  • the subjects head position is first aligned to 0 degrees, which is perpendicular to the camera from which the subject will be tracked from.
  • the first enrolment image is taken, the subject then rotates their head by a number of degrees and a second enrolment image is taken. This procedure is then repeated until the subject's head is at right angles to the camera, that is the subject's head is at 90 degrees to the camera.
  • the enrolment procedure is completed, there will be a collection of enrolled images spanning the head 90 degrees either side of a front of head view.
  • a new template is created holding a number of images representing the subject covering an angular range of 180 degrees.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Collating Specific Patterns (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)

Abstract

A facial identification matrix (FIM) is acquired by obtaining digital image information from a subjects face, including depth information, using conventional stereoscopic imaging techniques. A single camera technique is described in which an invisible grid is projected onto the face. Biometric engines are used to create 2D and 3D biometric templates from which facial indexing data is extracted and stored with the two templates. The FIM may be stored on a portable data carrier for verifying the identity or a subject or written to a database for efficient identification of a subject. To detect fraud a facial patch engine is included which compares regions of interest in at least two facial images of the same subject and detects selective movement in a proportion of the image, e.g. caused by blinking.

Description

IDENTITY SYSTEMS
TECHNICAL FIELD OF THE INVENTION
This invention relates to a Facial Identification Matrix (FIM). A FIM is a means of holding facial biometric data; specifically for application across the whole spectrum of security and identification requirements.
BACKGROUND
Biometrics is the science of automatically identifying individuals based on their unique physiological and/or behavioural characteristics. Biometric information based on the unique characteristics of a person's face, iris, voice, fingerprints, signature, palm prints, hand prints, or DNA can all be used to authenticate a person's identity or establish an identity from a database.
The use of facial biometric information has a number of advantages. Existing methods have a good level of reliability and the information can be obtained quickly using non-intrusive techniques. Furthermore, the data is not sensitive to superficial facial features such as sunglasses and beards; primarily because the system can undertake training of the data whilst processing the image. One of the major methods of the first generation of facial biometric software, based on "neural net" creates a template from a two-dimensional (2D) image of the subjects face; although this can be very accurate, both the angle at which the image is captured and the nature of the lighting are critical. Problems also arise with individuals having low skin contrast levels (e.g. people of Afro-Caribbean origin).
A second generation of facial biometric software is now available which has overcome some of the issues with regard to image angle, lighting and skin contrast.
SUMMARY OF THE INVENTION
In simple terms the invention allows for images captured to be rendered into 3D models, and depending on the number of cameras used, the image can be accurately rotated through 360 degrees. Our invention makes use of a 3D Biometric Engine [The 3D Biometric Engine includes the following modules; 3D Patch Engine, 3D Graph Engine, 3D Feature Selector Engine, 3D Indexing Engine and a control module which controls the modules and the input and output functions of the system.]. This takes data from a 3D camera system which is passed to both a 2D and separate 3D Biometric Engine; the 3D engine enables a correlated index key to be created. The present invention provides a process for producing a FIM [A FIM contains both 2D and 3D templates, indexing data and optionally a copy of the original captured.] which is more accurate than any currently existing biometric identifier.
The present invention proposes a methodology for producing a FIM from a subject which includes:
- The linking of various known, and new, techniques to produce a facial biometric product with new and enhanced capabilities. - obtaining a 3D facial image of the subject with improved and balanced contrasted levels through the use of modified lighting and filtering methods;
- producing both 2D and 3D biometric data from the captured facial image;
- providing the capability of using the 3D data to generate a 2D biometric template of the subjects face and manipulate images captured from various angles;
- providing the capability of using the 2D data to generate a 3D biometric template of the subjects face and manipulate images captured from various angles;
- through the use of the Patch Engine the invention provides the ability to detect motion at minute levels, down to the level of 3 to 4 pixels, for example the partial blink of an eye-lid.
- the ability to provide a 3D depth filter, to work with Skin Segmentation, in defining Region of Interest (ROI) to a much greater degree of accuracy than currently available.
- combining the two sets of biometric data to form a unique FIM for the subject, thus providing the means of creating a correlated indexing system.
The invention is more accommodating for poor lighting conditions and weak skin contrast levels, the range of angles through which a successful image can be captured is increased through the ability to correct angular displacement. The use of the two different kinds of biometric data significantly increases the overall accuracy of identification when measured against using either of the biometric methods individually.
The accuracy of the matrix-generating process can be improved by obtaining a plurality (verification set) of facial images from the subject, generating the two kinds of biometric information for each successive image, and comparing each new set of information with the information previously generated until a predetermined correlation level is achieved. The FIM may be stored together with the facial images which were used to generate it.
It will be appreciated that the invention further provides a FIM which contains both a 3D facial biometric template and a 2D biometric template for a subject.
The invention, using the FIM, further provides a process for determining the identity of a subject or for verifying the subject against a known FIM. The process compares the matrix (thereby obtained through the image process) with a stored FIM which contains 3D facial biometric data and a 2D biometric template derived there from, to determine whether the two matrixes match.
The identification matrix is small enough to be written to portable identification cards, added to a database along with numerous similar matrixes or transmitted electronically. In one use of the process the stored matrix may be included on a portable data carrier such as an identity card, a 2D Barcode printout, smart card or any portable media capable of storing data. Thus, the verification process can be used to verify that the data carrier belongs to the individual presenting it.
The process results in a considerable increase in the accuracy with which the identity of an individual can be verified, to a level previously unobtainable using non-intrusive biometric methodology. There is also a significant reduction in the number of false acceptances.
The accuracy of the verification process can be improved by repeatedly generating identification matrixes from the subject and comparing them with the stored matrix until a predetermined correlation level is achieved.
In another use of the process the stored FIM can be found from a database of similar matrixes using the index key value of the FIM. Thus, for example, the identity of an unknown individual can be ascertained by first finding all the matching index values which equal the newly generated FIM and then comparing each of the stored matrixes until a best match is found. The invention allows for duplicate index key values.
BRIEF DESCRIPTION OF THE DRAWINGS
The following description and the accompanying drawings referred to therein are included by way of non-limiting example in order to illustrate how the invention may be put into practice. In the drawings:
Figure 0 - This is a high level overview of the main modules forming the core of the invention.
Figure 1 - A high level diagrammatic representation of the process for producing a FIM in accordance with the invention;
Figure 2 - A high level diagrammatic representation of the process for verifying the identity of a subject using the matrix.
Figure 3 - This is a high level diagrammatic representation of the process for identifying a subject using the FIM recalled from storage.
Figures 4a, b, c & d - Concerns image capture and lighting issues.
Figure 5 - Basic 2D & 3D Enrolment.
Figure 6 - Capture a 3D Image using a Single Camera Source
Figure 7 - Capture a 3D Image using a Pair of Cameras
Figure 8 - Creation of a 3D Image of the face.
DETAILED DESCRIPTION OF THE DRAWINGS
Overview: [See Figure 0.]
This is a high level overview of the main modules involved in the invention.
It explains which modules are linked together and in which order.
Equipment
Suggested equipment required for use by the various processes described. A standard PC (750 MHz processor, 128 MB of RAM, 10 gigabyte hard drive, 15 inch CRT or TFT LCD screen, keyboard, mouse, video Grabber Card) Microsoft Windows 2000 Professional Operating System (System requires a secure O/S), an approved video camera, (for the low-end systems a USB Video Web Camera can be utilised) lighting units as specified. (Basically the lights illuminate the subject with light which is outside of the spectrum visible to a human being.) In certain circumstances an optional control device such as a touch screen may be utilised. If Process Step 7 (as described below) is implemented then a device, such as a smart card reader or a printer capable of outputting 2D barcode would be required.
Enrolment: [See Figure 1.]
The enrolment process involves the following steps:
Process Step 1. Image capture. This is the act of obtaining a digital image of the subject as a direct result of an input into the system from either a digital video camera or a Web camera. Depending on which methodology is used to obtain the third coordinate (required for the 3D biometric engine) either a single camera is used or a stereoscopic camera system. For details of the two different methodologies which are available see the entry below on pseudo 3D image calculation and stereoscopic 3D image. Whilst the image capture is being processed the system will normally display to the subject a copy of the image capture, however, in certain circumstances this will not be the case. Both the input from the Web Camera and the video camera provide a number of frames which are available for capture and utilisation by the system. The process monitors the quality of the image and assigns a value which relates to the acceptableness of the image quality provided. In the event that the image quality is adversely affected by strong peripheral lighting the subject would be advised of the problem. A series of "live" images are preferably obtained from the subject from whom successive pairs of biometric identifiers are created as described. The newly formed identifiers are compared to previously stored identifiers until such time as a perfect match is achieved or the verification level reaches a preset figure. The image can be taken at various angles covering the front of the subject.
Process Step 2. Image separation. At this point the 2D and 3D images are created separately in order that they may be passed to the appropriate biometric engine. Depending on the method used to obtain the third coordinate either the 3D image will be calculated from a single camera system or a 2D image will be calculated from the images provided by the stereoscopic system.
Process Step 3. The 2D data is passed to the 2D biometric engine which performs a number of tasks and results in the 2D biometric template.
Process Step 4. The 3D data is passed to the 3D biometric engine which performs a number of tasks and results in the creation of a 3D biometric template.
Process Step 5. Indexing, at this point data from the 3D biometric engine is used to create an index value which has a specific correlation to the features of the subject involved. [The system allows for a duplicated index value, for example, identical twins, who are truly identical, may well produce a calculated value which in itself is identical.] At this point the two sets of data (2D and 3D templates) linking the index values are encapsulated together, this data set is known as a FIM.
Process Step 6. The FIM produced is appended to a database. Process Step 7. (Optional) If required the FIM calculated can be output directly to a portable data storage product.
The two best biometric identifiers are then converted into a binary format which is suitable for storage. These biometric identifiers are then stored as a FIM, along with the "best match" image of the subject and any other desired data. Since the size of the matrix is remarkably small it can be recorded on a plastic or paper identity card, e.g. airline boarding pass or a company ID card, a smart card or any suitable carrier. The matrix could, for example, be incorporated into a barcode in PDF417 format (capacity about 1200 bytes), a USB token, a memory stick or similar binary recording means. In the case of a barcode the information can be printed using a laser printer for paper/cardboard or a commercial 150 dpi thermal printer for plastic cards. Other biometric, image or reference data may be included on the card as desired.
The FIM may be transmitted rapidly by any communication means, for example Blue tooth fax or email, to enable the FIM to be used at a remote location.
Verification: [See Figure 2.]
The verification process involves the following steps:
Process Step 1. The FIM or unique ID of the person to be verified is entered into the system, the method and type of input device by which data is entered into the system can vary considerably for example; PDF417 (2D barcode), smart cards, smart keys, proximity card, keyboard entry or even a floppy disk could be the means of triggering the system to verify the subject. The FIM does not actually have to be entered into the system, the input to the system only needs to identify the location of the FIM, it is required that the location of the FIM is available to the system. If a touch screen facility is provided the image may be displayed on the touch screen so that the required facial image can be selected from a group of people by touching the appropriate face.
Process Step 2. The 2D and 3D images are acquired as per step 1 of the enrolment. Depending on the application either a single or multi camera system would be utilised.
Process Step 3. Image separation. At this point the 2D and 3D images are created separately in order that they may be passed to the appropriate biometric engine. Depending on the method used to obtain the third coordinate either the 3D data will be calculated from the single image (provided from a one camera system) or 2D data will be calculated from the images provided by the stereoscopic system. The 2D data is passed to the 2D biometric engine which performs a number of tasks resulting in the creation of a 2D biometric template. The 3D data is passed to the 3D biometric engine which performs a number of tasks and results in the creation of a 3D biometric template.
Process Step 4. The 2D and 3D biometric engines are used to compare the data held in the FIM against the newly acquired image. The user, as part of the system setup, can set a predefined threshold with regard to the accuracy of the system; this value is used in determining whether there is an acceptable match between the two sets of data. Process Step 5. Dependent on the results of Process Step 4, an action or series of actions could be implemented, for example a trigger signal to open a door release.
Process Step 6. In most circumstances a message will be output indicating the results of the verification. This might be through the means of a message displayed on a computer screen, an audible response or the illumination of a sign.
Process Step 7. The system can record the details of the verification process which has just been enacted in the form of an Audit Trail. This ability enables the verification process to form a major part of an automated Time and Attendance system. The system can also run a training module which updates the FIM with the latest acquired image of the subject. In the case of a failed verification or where the input device failed a particular test (valid date) the system may withhold the return of the input device or overwrite the device effectively canceling its operational use.
Identification: [See Figure 3.]
Process Step 1. The 2D and 3D images are acquired as per step 1 of the enrolment. Depending on the application either a single or multi camera system would be utilised.
Process Step 2. Image separation, at this point the 2D and 3D images are created separately in order that they may be passed to the appropriate biometric engine. Depending on the method used to obtain the third coordinate either the 3D data will be calculated from the single image (provided from a one camera system) or 2D data will be calculated from the images provided by the stereoscopic system. The 2D data is passed to the 2D biometric engine which performs a number of tasks and results in the 2D biometric template. The 3D data is passed to the 3D biometric engine which performs a number of tasks and results in the creation of a 3D biometric template.
Process Step 3. Using the 3D biometric data a new index key is generated.
Process Step 4. Using the key value generated at Process Step 3 the system recovers all FIM's with that specific key value. The 2D and 3D biometric engines are used to compare the data held in the returned FIM's against the FIM of the newly acquired image. The user, as part of the system setup, can set a predefined threshold with regard to the accuracy of the system; this value is used in determining whether there is an acceptable match between the two sets of data.
Process Step 5. Dependent on the results of Process Step 4, an action or series of actions could be implemented, for example a trigger signal needed to open the door release. In most circumstances a message will be output indicating the results of the verification. This might be through the means of a message displayed on a computer screen, an audible response or the illumination of a sign.
Process Step 6. The system can record the details of the identification process which has just been enacted to form an Audit Trail. The system can also run a training module which updates the FIM with the latest acquired image of the subject thus helping to maintain an accurate collection of images of the subject. In the case of a failed verification or where the input device failed a particular test (valid date) the system may withhold the return of the input device or overwrite the device effectively canceling its operational use.
The enrolment and verification processes are not restricted to live CCTVfor image capture. Digital images can be acquired by means of digital still cameras, laptop computers with camera attachments, i.e. a Web camera, a PDA hand held barcode scanner with an image capture head, scanned from printed documents or photographs, camera-enabled WAP phones, etc. Images when acquired by these means would normally result in the 2D biometric process only, being applied.
Uses of the identification method include travel documents, banking, healthcare, social security, passports and immigration, education, the prison service, ATM machines, retail, secure access, document security, internet services and identification of football fans.
Image Capture and Lighting: [See Figures 4a - 4dJ The system makes use of cameras which have been tested in order to ensure their compliance for use with the invention, the cameras whether video cameras or Web cameras require high-grade optics and CCDs, the resolution and scanning levels of the system are also important, poor quality products will not perform adequately.
The figure 4a shows a single camera system where a projector is used to illuminate the subject with a calibrated grid; the frequency of illumination is outside that of a person's ability to register. The subject is further illuminated by a balanced light source (i.e. light falling of each side of the subjects face is balance to reflect an equal level of light) providing light of a wavelength outside the visible spectrum. The camera is capable of switching, via software control, between colour and the wavelength(s) of the emitters used in the projector and lamps.
Figure 4b uses two cameras the angle between each camera is known and fixed; the lighting characteristics are the same as for figure 4a. The images from both cameras are fed simultaneously to the computer system.
At figure 4c multiple cameras are used, at least one pair of cameras will be used as described at figure 4b, the additional camera or cameras will normally be used to take colour images of the subject in addition to providing extra calibration information for the stereoscopic imaging process.
At figure 4d a single camera is used, meeting the range of wavelength sensitivity for the lighting and camera conditions as at figure 4a, the subject is partially surrounded by tracking which will enable the camera to track at high speed from one side of the subject to the other taking a number of images at predetermined points along the location of the tracking.
Note: The lighting fix does not have to be applied in every situation, if a constant level of light is available then the lighting fix may not need to be applied.
Basic 2D & 3D Enrolment: [See Figure 5]
Process Step 1. Image capture, is the process of obtaining a digital image of the subject normally as a direct result of an input into the system from either a digital video camera or a Web camera. Either a single camera or a stereoscopic/multi camera system is used (applying the appropriate methodology) to obtain the third coordinate required for the 3D biometric engine. A framing box is positioned around the face, in order to achieve the maximum clarity of the image either a telescopic lens is used to zoom in on the subject or the image is digital enlarged to fill the framing box.
Process Step 2. Once the application has determined that the subject is correctly positioned within the framing box, the user is required to confirm that his face is correctly positioned; the process of enrolment can now commence.
Process Step 3. The main process, see steps 3a and 3b below, like a software 'Read ahead' operation, sets an initial value which the main process can use as a comparator.
Process Step 3a. The captured image is passed for processing by the 2D biometric engine; the first template created is a master.
Process Step 3b. The template is passed back to the application including an assessment score and size of the template.
Process Step 3 (Repeated). Further n images of the subject are taken (the actual number selected is determined in the program setup, for example 10 images may be taken of which only 5 are used in the process) The images are passed to the 2D biometric engine in order to determine that they are (a) valid template material (b) the score obtained is of an acceptable value (c) and the size of template is acceptable.
Process Step 4. The templates are sorted in order, based on their quality and their size; a predetermined number of templates are selected to be encapsulated into the FIM.
Process Step 5. The selected templates are used to create a 2D person template which is then written to the application database.
Process Step 6 (not shown in Figure). The 3D enrolment follows a similar process but uses the 3D biometric engine.
Capture a 3D Image Using a Single Camera as Source: [See Figure 6]
Process Step 1. Projection using non-visible light of a calibration grid
The target subject is illuminated using a wavelength of the spectrum, which is not visible to the naked eye. Included within the projected light is a calibration grid (horizontal and vertical lines at a known distance), which through the use of filters and an appropriate camera device which can detect the grid, are captured along with the image. The camera used must have both a high-resolution lens and image capture element (Charge Coupled Device (CCD) or Low Noise CMOS) element and must also be able to see the wavelength of light projected. Modern CCDs can now work on resolutions up to 16.8M pixels.
Process Step 2. The captured image contains the target subject overlaid with the lines of the grid, because the contours of the surface where the horizontal and vertical lines of the calibration grid are overlaid over the subject's image the surface contour is deformed (appears to bend).
Process Step 3. Preparation of image data. In order to build a 3D model from 2D information using a calibration grid two images are required; a standard image incorporating colour information and a calibrated image which incorporates the standard image overlaid with the calibration grid.
Process Step 4. Linear scan of image data. The standard image overlaid with the information from the generated calibration grid is then placed in 3D space (R3), this gives the effect of a plane in 3D space (with all z co-ordinates set to zero), note that the calibration grid has the same resolution as the projected calibration grid. A linear scan of the image can now be performed; the process involves scanning along each of the horizontal calibration grid lines, from left to right, top to bottom. For each pass of the scan, the generated calibration grid is adjusted, at each intersecting point on the grid the corresponding picture element is checked in the image which incorporates both the image and calibration grid. If the pixel information reveals a colour, which is part of the calibrated grid itself the next scan is processed, otherwise the point referenced needs to be adjusted until all checks have been processed.
Process Step 5. Applying image texture data. A specific area of interest (nose, mouth etc.) in the standard image can be used to map across the surface area of the patch, normally however the area of the face (forehead down to the bottom of the chin) would be used. Texture mapping co-ordinates (s, t) can be calculated from the u and v co-ordinates that map the actual surface of the patch, taking into account patch dimensions (top/left = -1,-1 and bottom/right = 1,1).
Calculating actual texture mapping co-ordinates: s = (u+1.0)/2 t = (v+1.0)/2 ;
Values for u and v now range from 0.0 through to 1.0 these values can now be scaled to the device co-ordinates, however most 3D texture mapping hardware uses unit values ranging from 0.0 to 1.0.
Capture a 3D Image Using a Stereoscopic / Multiple Camera System as a Source: [See Figure 7.]
Figure 7 shows the layout of a pair of camera, for use in stereoscopic image capture. The principles on how this process functions are well known (see Mori, K., Kidode, M & Asada, H. (1973). An iterative prediction and correction method for automatic stereo comparison. Comp. Graph. Image Process). The figure indicates how two cameras can be used to determine depth information of a subject. By taking an image point from camera 1 (the point x1,y1) and searching for the same point in the image capture by camera 2 (the point x2,y2) and using image block cross-correlation it is possible to calculate the depth of this point in the real world. As camera 1 will have the coordinate system (x1,y1,z1), which will be different to that of the second camera 2 (x2,y2,z2), the stereo search will have two dimensions (x2,y2). Whilst there are different methods for calculating depth information, the use of the block cross-correlation method is preferred (even though costly in processor time) as it provides a more reliable way of obtaining the z1 values. The information obtained is used to build up a data set of the x,y,z coordinates for use by other modules within the application.
Creation of a 3D Image of the face using a 3D Patch Engine [See Figure 8.]
Process Step 1. Obtain Working Resolution. The size of the patch array (resolution) is obtained directly from the resolution provided by the image capture device; effectively this establishes the working parameters of the patch. The unit scale of the patch is 2 by 2 (represented as a true mathematical model). A scaling factor can be applied creating a true (real world) scale. The minimum working resolution will be 90 units thus the scale is 90/2, which is 45 units.
Process Step 2. Initialise Patch Variables. Two variables are required in order to specify the points in the patch these are commonly known as u and v. The variables u and v are the working variables required for the generation of the 3D co-ordinates. At initialisation both the u and v variables are initialised to the value minus one, this representing the top left co-ordinate of the patch. The top left coordinate serves as the starting point for the linear scan process. Further variables are defined in order for the process to be completed.
Process Step 3. Generation of a Patch Constraint Points. This process generates 3D point components, where the x and y components are calculated and the third component (z) is set to the value zero. Note that the patch at this stage represents a flat plane in 3D space, standing upright spanning across the Y axis, with a width that spans across the X axis. This process also serves as an initialisation process for the patch array. Process Step 4. Import of constraint data. Constraint information can be imported into the process from the data output by the 3D image capture engine in the form of 3D co-ordinates. The 3D coordinates now act as constraint points to be used in the generation of the curves; this provides the ability to automate the process of patch curve generation.
Process Step 5. Generation of Patch Curves. The generation of patch curves is accomplished by taking a collection of constraint points and calculating each part of the curve. The smoothness of the curve is variable depending on the level smoothness required. This process is applied across the u array of constraint points with the results feeding back into the patch array on completion of the u array being filled the process is repeated for the v array of constraint points. The final result is a grid of curves that breaks the patch down into quadrants with all curves being inter-connecting.
Process Step 6. Storage of 3D template information via a database. The final process is the creation of a 3D template, which is in the form of a FIM 3D Class; this is then stored in a database. Information held in the database can be recalled to be rendered on the display device. The application module has the ability to rotate the model about a central axis +/- 15 degrees.
3D Graph Pattern Matching
The graph-matching module has been developed at the University of York (UK); it is an advanced system for finding solutions to constraint-based problems. The solution prevents the exponential complexity of the problem by using methods based on that originally proposed by Hancock and Kittler. This method was extended and developed as neural relaxation by Austin and Turner, further development introduced correlation matrix memories as a calculation engine for the system. This technique and its extensions are embodied in the Advanced Uncertain Reasoning Architecture (AURA) graph matching engine (from the University of York) versions 1.0 onwards. The underlying AURA library is the Copyright University of York and Cybula Ltd.
3D Feature Selector Engine
The engine is use to delimit a face from the surrounding background and for defining the key features of the face, for example eyes, mouth, nose etc.
Skin Chrominance Process
This element of the invention (3D Feature Selector Engine) uses a face / facial characteristics detection algorithm, which is able to handle a wide variety of variations in colour images; the engine also makes of pixel depth data from information supplied by the Patch Engine. The uses of colour information can simplify the task of finding faces in different environmental lighting conditions. The process for locating facial images/facial characteristics is shown below:
1. Acquire the image in colour, i 2. Provide for a lighting compensation.
3. Collect information on skin tone.
4. Create skin segmentation data.
5. Consolidate information on skin tones in order to create an overall skin image.
Lighting Compensation and Skin Colour Detection
It is known that the appearance of the skin colour will change under different lighting conditions. Modelling skin tone colour involves isolating the skin colour space and using the cluster associated with the skin colour in this space.
Skin Segmentation
A problem arises when two faces (or a face and skin coloured objects) in a scene, when one face is actually touching or overlapping the other, causing them to be segmented from the scene as a single object and hence forming a single Region of Interest (ROI). The solution to this problem in the past has been to use an edge enhancement filter before skin segmentation; care must be exercised in how strongly the filters are applied, as enhancement filters tend to degrade the overall image. As part of this invention we now apply a 3D Depth filter to aid the segmentation process. It follows that the reverse is also true and that the skin tone information can be mapped to the Patch Engine output as a fourth component. The invention also applies the method of the colour difference between adjacent pixels, proportional to the original image colour difference.
Region of Interest Detection
The essence of ROI detection relies on skin tones, which allow (eye, nose etc) areas of interest to be processed. This process does not isolate a face in a scene but isolates areas of interest. Using a combination of TRUE positives with FALSE positives and various other percentage results, a probability face graph can be constructed and a decision threshold implemented. A neural net is now trained with results taken from the ROI process; both FALSE and TRUE results are used, resulting in face detection in a scene.
Indexing, using 3D Elements. The index module works with information provided from several modules, the major ones being the delimitation of facial features and the 3D depth calculations. Common facial features, for example eyes, eyes to ear edge, nose to closed-jaw, depth of eyebrows etc can be used as elements of a compound index. The element values would be a normalized set of data based on five or more images of the subject. The delimitation of facial features module has to determine whether a true element value is used or if there is a concern regarding accuracy of the element then whether an average value should be substituted. The more elements used the greater the ability to separate out individuals and the lower the risk of using average values for some of the elements in the key. Each element has a one bit Boolean flag to indicate whether a real or substituted value has been supplied to the element.
The following is an illustration of the concept using only four elements: L.Eye (205); R.Eye (204); Nose to Closed-Jaw (650); Eye to Eye (200) the key would like 205b204b650B200b, 650 is a substituted value. Using 30 or more elements creates a highly selective and searchable index; identification can be against any part(s) of the key or the whole key.
Capturing images at various angles. (Optional Feature)
Description of method.
The invention will identify a subject, relative to a specific camera which is mounted at a known position in the environment, with the subjects face being at various angles.
Enrolment Process The subjects head position is first aligned to 0 degrees, which is perpendicular to the camera from which the subject will be tracked from. The first enrolment image is taken, the subject then rotates their head by a number of degrees and a second enrolment image is taken. This procedure is then repeated until the subject's head is at right angles to the camera, that is the subject's head is at 90 degrees to the camera. Once a number of enrolment images have been taken, with the head rotating from centre to right, the procedure is repeated for the opposite side of the face. Note, there is no need to enrol the image of the subject's first head position at centre left, as this would duplicate with the first enrolment image made. When the enrolment procedure is completed, there will be a collection of enrolled images spanning the head 90 degrees either side of a front of head view. A new template is created holding a number of images representing the subject covering an angular range of 180 degrees.
Verification Process
The normal verification process is now applied with the new template information, which can now manage multi-angled images previously outside of the scope of the system.
It will be appreciated that the features disclosed herein may be present in any feasible combination. Whilst the above description lays emphasis on those areas which, in combination, are believed to be new, protection is claimed for any inventive combination of the features disclosed herein.

Claims

1. A facial identification matrix which incorporates both a 2D and a 3D biometric facial template derived from the face of a subject together with facial indexing data for identifying the subject based on facial reference features.
2. A facial identification matrix according to Claim 1, which additionally incorporates a digital image of the subjects face.
3. A digital storage medium holding a facial identification matrix according to Claim 1 or 2.
4. A database holding a plurality of facial identification matrixes according to Claim 1 or 2 which are derived from different subjects.
5. A method of acquiring a facial identification matrix from a subject which includes the steps of:
- using a subjects face to obtain digital image information therefrom, including depth information;
- creating a 2D biometric template from the digital image information;
- creating a 3D biometric template from the digital image information;
- extracting facial indexing data from the biometric template data; and
- linking the two biometric templates and the facial indexing data to create a facial identification matrix for the subject.
6. A method according to Claim 5, in which the facial identification matrix additionally incorporates a digital image of the subjects face.
7. A method according to Claim 5, in which the depth information is obtained by projecting a pattern onto the subjects face, recording a pattern in a digital image and calculating the depth information from the distortion of the pattern caused by the contours of the face.
8. A method according to Claim 7, in which the depth information is calculated using an image of the subjects face which is contemporaneously recorded in the visible spectrum.
9. A method according to Claim 7, in which the pattern is projected and recorded at a non-visible wavelength.
10. A method according to Claim 5, in which each biometric template is generated by creating biometric reference templates from successive sets of digital image information and selecting the best template according to predetermined selection criteria.
11. A method according to Claim 5, in which corresponding regions of interest in at least two successive digital images of said face are compared to detect any movement therein, and a warning is signalled in the absence of selective movement within specific regions of interest.
12. A method according to Claim 5, in which the facial identification matrix is recorded on a digital storage medium.
13. A method according to Claim 5, in which the facial identification matrix is added to a database holding a plurality of facial identification matrixes which are derived from different subjects.
14. A method of verifying the identity of a subject which includes acquiring a facial identification matrix from a subject in accordance with Claim 5, comparing said facial identification matrix with a facial identification matrix recorded on a digital storage medium, and signalling whether or not the two matrixes correspond within a specified tolerance.
15. A method of identifying a subject which includes acquiring a facial identification matrix from a subject in accordance with Claim 5, comparing the facial indexing data thereof with the facial indexing data of matrixes stored in a database, and retrieving any facial identification matrix which matches within a specified tolerance.
16. Apparatus for acquiring a facial identification matrix from a subject which includes:
- image capture means for obtaining digital image information from subjects face, including depth information;
- a 2D engine for creating a 2D biometric template from the digital image information;
- a 3D engine for creating a 3D biometric template from the digital image information;
- indexing means arranged to extract facial indexing data from the biometric template data; and
- output means for providing a facial identification matrix for the subject which includes the two biometric templates and the facial indexing data.
17. Apparatus according to Claim 16, in which the facial identification matrix additionally incorporates a digital image of the subjects face.
18. Apparatus according to Claim 16, in which the depth information is obtained by projecting a pattern onto the subjects face, recording a pattern in a digital image and calculating the depth information from the distortion of the pattern caused by the contours of the face.
19. Apparatus according to Claim 18, in which the depth information is calculated using an image of the subjects face which is contemporaneously recorded in the visible spectrum.
20. Apparatus according to Claim 18, in which the pattern is projected and recorded at a non-visible wavelength.
21. Apparatus according to Claim 16, in which each biometric template is generated by creating biometric reference templates from successive sets of digital image information and selecting the best template according to predetermined selection criteria.
22. Apparatus according to Claim 16, in which corresponding regions of interest in at least two successive digital images of said face are compared to detect any movement therein, and a warning is signalled in the absence of selective movement within specific regions of interest.
23. Apparatus according to Claim 16, in which the facial identification matrix is recorded on a digital storage medium.
24. Apparatus according to Claim 16, in which the facial identification matrix is added to a database holding a plurality of facial identification matrixes which are derived from different subjects.
25. Apparatus for verifying the identity of a subject which includes means for acquiring a facial identification matrix from a subject in accordance with Claim 16, means for comparing said facial identification matrix a facial identification matrix recorded on a digital storage medium, and means for signalling whether or not the two matrixes correspond within a specified tolerance.
26. Apparatus for identifying a subject which includes means for acquiring a facial identification matrix from a subject in accordance with Claim 16, means for comparing the facial indexing data thereof with the facial indexing data of matrixes stored in a database, and means for retrieving any facial identification matrix which matches within a specified tolerance.
27. A method of detecting fraud in a facial identification system, which includes comparing corresponding regions of interest in at least two successive digital images acquired from a subjects face to detect movement therein, and signalling a warning in the absence of selective movement within specific regions of interest.
28. A facial patch engine which compares regions of interest in at least two facial images of the same subject and detects selective movement in a proportion of the image.
PCT/GB2001/003322 2000-07-25 2001-07-24 Identity systems WO2002009024A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU2001272667A AU2001272667A1 (en) 2000-07-25 2001-07-24 Identity systems

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB0018161.0 2000-07-25
GBGB0018161.0A GB0018161D0 (en) 2000-07-25 2000-07-25 Identity systems

Publications (1)

Publication Number Publication Date
WO2002009024A1 true WO2002009024A1 (en) 2002-01-31

Family

ID=9896258

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/GB2001/003322 WO2002009024A1 (en) 2000-07-25 2001-07-24 Identity systems

Country Status (3)

Country Link
AU (1) AU2001272667A1 (en)
GB (2) GB0018161D0 (en)
WO (1) WO2002009024A1 (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2396002A (en) * 2002-10-09 2004-06-09 Canon Kk Gaze tracking system
EP1471455A2 (en) * 2003-04-15 2004-10-27 Nikon Corporation Digital camera
WO2005008570A2 (en) * 2003-05-27 2005-01-27 Honeywell International Inc. Face identification verification using frontal and side views
WO2005064525A1 (en) * 2003-12-30 2005-07-14 Kield Martin Kieldsen A method and apparatus for providing information relating to a body part of a person, such as for identifying the person
WO2005096768A2 (en) * 2004-04-06 2005-10-20 Rf Intelligent Systems, Inc. Combined 2d/3d facial authentication
WO2005098744A2 (en) * 2004-04-06 2005-10-20 Rf Intelligent Systems, Inc. Handheld biometric computer for 2d/3d image capture
FR2880158A1 (en) * 2004-12-23 2006-06-30 Sagem Individual identifying and fraud detecting method, involves comparing two dimensional characteristics of image detected on individual with two dimensional characteristics of stored image, in case of coherence
US7158097B2 (en) 2002-10-09 2007-01-02 Canon Kabushiki Kaisha Gaze tracking system
US7599531B2 (en) 2004-08-24 2009-10-06 Tbs Holding Ag Method and arrangement for optical recording of biometric finger data
US7606395B2 (en) 2004-08-25 2009-10-20 Tbs Holding Ag Method and arrangement for optical recording of data
US8675926B2 (en) 2010-06-08 2014-03-18 Microsoft Corporation Distinguishing live faces from flat surfaces
US8823864B2 (en) 2007-06-28 2014-09-02 Sony Corporation Image capturing apparatus and associated methodology for auto-focus and facial detection
WO2015086964A1 (en) * 2013-12-12 2015-06-18 Orange Imaging system for rendering with an impression of relief
WO2017060329A1 (en) 2015-10-06 2017-04-13 Janssen Vaccines & Prevention B.V. Methods for preventing plastic-induced degradation of biologicals
CN113228037A (en) * 2018-12-18 2021-08-06 创新先进技术有限公司 Creating iris signatures to reduce search space for biometric recognition systems

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0353385A (en) * 1989-07-21 1991-03-07 Nippon Denki Sekiyuritei Syst Kk Feature extracting device
US5801763A (en) * 1995-07-06 1998-09-01 Mitsubishi Denki Kabushiki Kaisha Face image taking device
DE19712844A1 (en) * 1997-03-26 1998-10-08 Siemens Ag Method for three-dimensional identification of objects

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0353385A (en) * 1989-07-21 1991-03-07 Nippon Denki Sekiyuritei Syst Kk Feature extracting device
US5801763A (en) * 1995-07-06 1998-09-01 Mitsubishi Denki Kabushiki Kaisha Face image taking device
DE19712844A1 (en) * 1997-03-26 1998-10-08 Siemens Ag Method for three-dimensional identification of objects

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
LOPEZ R ET AL: "3D HEAD POSE COMPUTATION FROM 2D IMAGES: TEMPLATES VERSUS FEATURES", PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING. (ICIP). WASHINGTON, OCT. 23 - 26, 1995, LOS ALAMITOS, IEEE COMP. SOC. PRESS, US, vol. 2, 23 October 1995 (1995-10-23), pages 599 - 602, XP000624040, ISBN: 0-7803-3122-2 *
PATENT ABSTRACTS OF JAPAN vol. 015, no. 205 (P - 1206) 27 May 1991 (1991-05-27) *
TORU ABE ET AL: "AUTOMATIC IDENTIFICATION OF HUMAN FACES BY 3-D SHAPE OF SURFACES-USING VERTICES OF B-SPLINE SURFACE", SYSTEMS & COMPUTERS IN JAPAN, SCRIPTA TECHNICA JOURNALS. NEW YORK, US, vol. 22, no. 7, 1991, pages 96 - 104, XP000259345, ISSN: 0882-1666 *
YACOOB Y ET AL: "LABELING OF HUMAN FACE COMPONENTS FROM RANGE DATA", CVGIP IMAGE UNDERSTANDING, ACADEMIC PRESS, DULUTH, MA, US, vol. 60, no. 2, 1 September 1994 (1994-09-01), pages 168 - 178, XP000484201, ISSN: 1049-9660 *

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7158097B2 (en) 2002-10-09 2007-01-02 Canon Kabushiki Kaisha Gaze tracking system
GB2396002A (en) * 2002-10-09 2004-06-09 Canon Kk Gaze tracking system
GB2396002B (en) * 2002-10-09 2005-10-26 Canon Kk Gaze tracking system
EP1471455A2 (en) * 2003-04-15 2004-10-27 Nikon Corporation Digital camera
EP1471455A3 (en) * 2003-04-15 2005-05-25 Nikon Corporation Digital camera
US9147106B2 (en) 2003-04-15 2015-09-29 Nikon Corporation Digital camera system
WO2005008570A2 (en) * 2003-05-27 2005-01-27 Honeywell International Inc. Face identification verification using frontal and side views
WO2005008570A3 (en) * 2003-05-27 2005-05-06 Honeywell Int Inc Face identification verification using frontal and side views
US7421097B2 (en) 2003-05-27 2008-09-02 Honeywell International Inc. Face identification verification using 3 dimensional modeling
WO2005064525A1 (en) * 2003-12-30 2005-07-14 Kield Martin Kieldsen A method and apparatus for providing information relating to a body part of a person, such as for identifying the person
WO2005096768A2 (en) * 2004-04-06 2005-10-20 Rf Intelligent Systems, Inc. Combined 2d/3d facial authentication
WO2005098744A2 (en) * 2004-04-06 2005-10-20 Rf Intelligent Systems, Inc. Handheld biometric computer for 2d/3d image capture
WO2005096768A3 (en) * 2004-04-06 2007-01-04 Rf Intelligent Systems Inc Combined 2d/3d facial authentication
WO2005098744A3 (en) * 2004-04-06 2007-03-01 Rf Intelligent Systems Inc Handheld biometric computer for 2d/3d image capture
US7599531B2 (en) 2004-08-24 2009-10-06 Tbs Holding Ag Method and arrangement for optical recording of biometric finger data
US7606395B2 (en) 2004-08-25 2009-10-20 Tbs Holding Ag Method and arrangement for optical recording of data
FR2880158A1 (en) * 2004-12-23 2006-06-30 Sagem Individual identifying and fraud detecting method, involves comparing two dimensional characteristics of image detected on individual with two dimensional characteristics of stored image, in case of coherence
WO2006070099A3 (en) * 2004-12-23 2006-09-21 Sagem Defense Securite Method for identifying a person from the person's features, with fraud detection
WO2006070099A2 (en) * 2004-12-23 2006-07-06 Sagem Defense Securite Method for identifying a person from the person's features, with fraud detection
US8823864B2 (en) 2007-06-28 2014-09-02 Sony Corporation Image capturing apparatus and associated methodology for auto-focus and facial detection
US8675926B2 (en) 2010-06-08 2014-03-18 Microsoft Corporation Distinguishing live faces from flat surfaces
WO2015086964A1 (en) * 2013-12-12 2015-06-18 Orange Imaging system for rendering with an impression of relief
FR3015161A1 (en) * 2013-12-12 2015-06-19 Orange RECOVERY INSTALLATION FOR RESTITUTION WITH RELIEF PRINTING
WO2017060329A1 (en) 2015-10-06 2017-04-13 Janssen Vaccines & Prevention B.V. Methods for preventing plastic-induced degradation of biologicals
CN113228037A (en) * 2018-12-18 2021-08-06 创新先进技术有限公司 Creating iris signatures to reduce search space for biometric recognition systems
CN113228037B (en) * 2018-12-18 2024-03-15 创新先进技术有限公司 Creating iris identifications to reduce search space for biometric identification systems

Also Published As

Publication number Publication date
AU2001272667A1 (en) 2002-02-05
GB0019555D0 (en) 2000-09-27
GB0018161D0 (en) 2000-09-13

Similar Documents

Publication Publication Date Title
JP3271750B2 (en) Iris identification code extraction method and device, iris recognition method and device, data encryption device
Tirunagari et al. Detection of face spoofing using visual dynamics
US7715596B2 (en) Method for controlling photographs of people
US7983451B2 (en) Recognition method using hand biometrics with anti-counterfeiting
Kose et al. Countermeasure for the protection of face recognition systems against mask attacks
Miao et al. A hierarchical multiscale and multiangle system for human face detection in a complex background using gravity-center template
EP3807792B1 (en) Authenticating an identity of a person
EP1629415B1 (en) Face identification verification using frontal and side views
Ferrara et al. Face image conformance to iso/icao standards in machine readable travel documents
US20090161925A1 (en) Method for acquiring the shape of the iris of an eye
US20150347833A1 (en) Noncontact Biometrics with Small Footprint
Galbally et al. A review of iris anti-spoofing
EP0731426A2 (en) Process for encrypting a fingerprint onto an I.D. card
Pala et al. Iris liveness detection by relative distance comparisons
WO2002009024A1 (en) Identity systems
Benalcazar et al. Synthetic ID card image generation for improving presentation attack detection
Benlamoudi Multi-modal and anti-spoofing person identification
Singh et al. FDSNet: Finger dorsal image spoof detection network using light field camera
CN111429156A (en) Artificial intelligence recognition system for mobile phone and application thereof
Rani et al. Pre filtering techniques for face recognition based on edge detection algorithm
JP4670619B2 (en) Biological information verification system
CN111160137B (en) Intelligent business processing equipment based on biological 3D information
Grabovskyi et al. Facial recognition with using of the microsoft face API Service
CN111291586A (en) Living body detection method, living body detection device, electronic apparatus, and computer-readable storage medium
Chiesa Revisiting face processing with light field images

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: COMMUNICATION UNDER RULE 69 EPC (EPO FORM 1205A OF 04.06.2003)

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: JP