Biometric acceptability

FACE ENGINE SPECIFICATIONS AND RECOMMENDATIONS

General recommendations for facial recognition:

  • Face recognition accuracy of the MegaMatcher algorithm heavily depends on the quality of a face image. Image quality during enrollment is important, as it influences the quality of the face template.
  • 32 pixels is the recommended minimal distance between eyes for a face on image or video stream to perform face template extraction reliably. 64 pixels or more recommended for better face recognition results. Note that this distance should be native, not achieved by resizing an image.
  • Several images during enrollment are recommended for better facial template quality which results in improvement of recognition quality and reliability.
  • Additional enrollments may be needed when facial hair style changes, especially when beard or mustache is grown or shaved off.
  • Persons wearing face masks or respirators can be recognized without separate enrollment. Face quality check should be disabled for this scenario.

The face recognition engine has certain tolerance to face posture:

  • head roll (tilt) ? ±180 degrees (configurable);

                    ±15 degrees default value is the fastest setting which is usually sufficient for most near-frontal face images.

  • head pitch (nod) ? ±15 degrees from frontal position.

                    The head pitch tolerance can be increased up to ±25 degrees if several views of the same face that covered different pitch angles were used during enrollment.

  • head yaw (bobble) ? ±90 degrees from frontal position (default value).

                Smaller yaw tolerance values are not recommended to be used except if the target system does not meet the system requirements.

                Several views of the same face can be enrolled to the database to cover the whole ±90 degrees yaw range from frontal position.

Face liveness check:

  • A stream of consecutive images (usually a video stream from a camera) or a single image (in some modes) are required for the live face detection.
  • When the liveness check is enabled, it is performed by the face engine before feature extraction. If the face in the stream fails to qualify as "live", the features are not extracted.
  • Only one face should be visible in these frames.
  • 80 pixels is the recommended minimal distance between eyes (IOD) for a face to perform liveness check reliably. 100 pixels or more recommended for smoother performance.
  • During passive liveness checks the face should be still and the user has to look directly at the camera with ±15 degrees tolerances for roll, pitch and yaw to experience the best performance.
  • Optionally, ICAO compliance check can be used to strengthen the liveness check.
  • Users can enable these liveness check modes:

                Active ? the engine requests the user to perform certain actions like blinking or moving one's head. All requested actions should be performed to pass the liveness check. This mode can work with both colored and grayscale images. 5 frames per second or better frame rate required.

                Passive ? the engine analyzes certain facial features while the user stays still in front of the camera for a short period of time. Colored images are required for this mode. 10 frames per second or better frame rate is required. Better score is achieved when users do not move at all.

                Passive + Blink ? the engine analyzes certain facial features while the user stays still in front of the camera for a short period of time, when the engine requests the user to blink. Colored images are required for this mode. 10 frames per second or higher frame rate required.

                Passive then active ? the engine first tries the passive liveness check, and if it fails, tries the active check. This mode requires colored images.

                Simple ? the engine requires user to turn head from side to side while looking at camera. This mode can work with both colored and grayscale images. 5 frames per second or better frame rate recommended.

                Single frame passive ? the engine uses a neural network to estimate if a face image is not inserted in front of a camera using a paper photo or smartphone screen. This mode does not need any interaction from the user.

Reference: https://www.neurotechnology.com/megamatcher-technical-specifications.html#face-engine?

No luck finding what you're looking for?

Let us know your question or query, click “E-mail Us” and we’ll get back to you as soon as possible.