Technical Specifications and Usage Recommendations
General recommendations for facial recognition:
- Face recognition accuracy of the VeriLook algorithm heavily depends on the quality of a face image. Image quality during enrollment is important, as it influences the quality of the face template.
- 32 pixels is the recommended minimal distance between eyes for a face on image or video stream to perform face template extraction reliably. 64 pixels or more recommended for better face recognition results. Note that this distance should be native, not achieved by resizing an image.
- Several images during enrollment are recommended for better facial template quality which results in user experience improvement during recognition.
- Additional enrollments may be needed when facial hair style changes, especially when beard or mustache is grown or shaved off.
- Persons wearing face masks or respirators can be recognized without separate enrollment. Face quality check should be disabled for this scenario.
The face recognition engine has certain tolerance to face posture:
head roll (tilt) – ±180 degrees (configurable);
- ±15 degrees default value is the fastest setting which is usually sufficient for most near-frontal face images.
head pitch (nod) – ±15 degrees from frontal position.
- The head pitch tolerance can be increased up to ±25 degrees if several views of the same face that covered different pitch angles were used during enrollment.
head yaw (bobble) – ±90 degrees from frontal position (default value).
- Smaller yaw tolerance values are not recommended to be used except if the target system does not meet the system requirements.
- Several views of the same face can be enrolled to the database to cover the whole ±90 degrees yaw range from frontal position.
- head roll (tilt) – ±180 degrees (configurable);
Face liveness check:
- The face liveness check algorithm was tested by iBeta and proven to be compliant with ISO 30107-3 Biometric Presentation Attack Detection Standards.
- A stream of consecutive images (usually a video stream from a camera) or a single image (in some modes) are required for the face liveness detection.
- When the liveness check is enabled, it is performed by the face engine before feature extraction. If the face in the stream or frame fails to qualify as "live", the features are not extracted.
- Only one face should be visible in these frames.
- 80 pixels is the recommended minimal distance between eyes (IOD) for a face to perform liveness check reliably. 100 pixels or more recommended for smoother performance.
- During passive liveness checks the face should be still and the user has to look directly at the camera with ±15 degrees tolerances for roll, pitch and yaw to experience the best performance.
- Optionally, ICAO compliance check can be used to strengthen the liveness check.
Users can enable these liveness check modes:
- Active – the engine requests the user to perform certain actions like blinking or moving one's head. All requested actions should be performed to pass the liveness check. This mode can work with both colored and grayscale images. 5 frames per second or better frame rate required.
- Passive – the engine analyzes certain facial features while the user stays still in front of the camera for a short period of time. Colored images are required for this mode. 10 frames per second or better frame rate is required. Better score is achieved when users do not move at all.
- Passive + Blink – the engine analyzes certain facial features while the user stays still in front of the camera for a short period of time, when the engine requests the user to blink. Colored images are required for this mode. 10 frames per second or higher frame rate required.
- Passive then active – the engine first tries the passive liveness check, and if it fails, tries the active check. This mode requires colored images.
- Simple – the engine requires user to turn head from side to side while looking at camera. This mode can work with both colored and grayscale images. 5 frames per second or better frame rate recommended.
- Single frame passive – the engine uses a neural network to estimate if a face image is not inserted in front of a camera using a paper photo or smartphone screen. This mode does not need any interaction from the user.
VeriLook 13.0 face detection algorithm can run in maximal speed or maximal accuracy modes. The face detection times in the table below are provided for 640 x 480 pixels images as ranges, where the smallest time corresponds to the maximal speed mode, and the largest time – to the maximal accuracy. The head pitch tolerance in the table below is always ±15°. The head yaw tolerance in the table below is always ±90°. The face detection times are specified for a PC with Intel Core i7-8700K processor.
|VeriLook 13.0 face detection algorithm performance for all faces in a frame on PC-based systems|
|Roll tolerance||Time (milliseconds)|
VeriLook biometric template extraction and matching algorithm is designed to run on multi-core processors allowing to reach maximum possible performance on the used hardware. Face template extraction is performed after all faces are detected in a frame. The template extraction time does not depend on image size.
|VeriLook 13.0 face engine specifications|
|Embedded / mobile (1)
|Template extraction components||Mobile
|Template extraction time (seconds)||1.34||1.20||1.34||0.60|
|Template matching components||Mobile Face Matcher||Face Matcher|
|Template matching speed
(faces per second)
|Single face record size in a template (bytes)||322|
(1) Requires to be run on iOS devices or Android devices based on at least Snapdragon S4 system-on-chip with Krait 300 processor (4 cores, 1.51 GHz).
(2) Requires to be run on PC or laptop with at least Intel Core i7-8700K processor.