Video Stream Processing
Overview
To process video streams, Face SDK uses VideoWorker
interface object, responsible for thread control and synchronization routine. You only need to provide decoded video frames and register a few callback functions.
VideoWorker
object can be used to:
- track faces on video streams
- create templates
- estimate age, gender, and emotions
- estimate liveness
- match detected faces with the database
- match faces detected within a specified time period
See an example of using VideoWorker
in video_recognition_demo.
Related tutorials:
- Face Tracking in a Video Stream
- Face Identification in a Video Stream
- Detection and Recognition of Masked Faces
Create VideoWorker Object
To create VideoWorker
, use FacerecService.createVideoWorker
method.
Examples
- C++
- C#
- Java
- Python
pbio::FacerecService::Config video_worker_config("video_worker_lbf.xml");
video_worker_config.overrideParameter("search_k", 3);
pbio::VideoWorker::Ptr video_worker = service->createVideoWorker(
pbio::VideoWorker::Params()
.video_worker_config(video_worker_config)
.recognizer_ini_file(recognizer_config)
.streams_count(streams_count)
.processing_threads_count(processing_threads_count)
.matching_threads_count(matching_threads_count)
.age_gender_estimation_threads_count(age_gender_estimation_threads_count)
.emotions_estimation_threads_count(emotions_estimation_threads_count)
.short_time_identification_enabled(enable_sti)
.short_time_identification_distance_threshold(sti_recognition_threshold)
.short_time_identification_outdate_time_seconds(sti_outdate_time)
);
FacerecService.Config video_worker_config = new FacerecService.Config("video_worker_lbf.xml");
video_worker_config.overrideParameter("search_k", 3);
VideoWorker video_worker = service.createVideoWorker(
new VideoWorker.Params()
.video_worker_config(video_worker_config)
.recognizer_ini_file(recognizer_config)
.streams_count(streams_count)
.processing_threads_count(processing_threads_count)
.matching_threads_count(matching_threads_count)
.age_gender_estimation_threads_count(age_gender_estimation_threads_count)
.emotions_estimation_threads_count(emotions_estimation_threads_count)
.short_time_identification_enabled(enable_sti)
.short_time_identification_distance_threshold(sti_recognition_threshold)
.short_time_identification_outdate_time_seconds(sti_outdate_time)
);
FacerecService.Config video_worker_config = service.new Config("video_worker_lbf.xml");
video_worker_config.overrideParameter("search_k", 3);
VideoWorker video_worker = service.createVideoWorker(
new VideoWorker.Params()
.video_worker_config(video_worker_config)
.recognizer_ini_file(recognizer_config)
.streams_count(streams_count)
.processing_threads_count(processing_threads_count)
.matching_threads_count(matching_threads_count)
.age_gender_estimation_threads_count(age_gender_estimation_threads_count)
.emotions_estimation_threads_count(emotions_estimation_threads_count)
.short_time_identification_enabled(enable_sti)
.short_time_identification_distance_threshold(sti_recognition_threshold)
.short_time_identification_outdate_time_seconds(sti_outdate_time)
);
video_worker_config = Config("video_worker_lbf.xml")
video_worker_config.override_parameter("search_k", 3)
video_worker_params = video_worker.Params()
video_worker_params.video_worker_config = video_worker_config
video_worker_params.recognizer_ini_file = recognizer_config
video_worker_params.streams_count = streams_count
video_worker_params.processing_threads_count = processing_threads_count
video_worker_params.matching_threads_count = matching_threads_count
video_worker_params.age_gender_estimation_threads_count = age_gender_estimation_threads_count
video_worker_params.emotions_estimation_threads_count = emotions_estimation_threads_count
video_worker = service.create_video_worker(video_worker_params)
Where:
video_worker_config
: path to the configuration file forVideoWorker
orFacerecService.Config
object.video_worker_params
: parameters ofVideoWorker
constructor.recognizer_config
: the configuration file of the method used (see Face Identification).streams_count
: the number of video streams.processing_threads_count
: the number of threads for template extraction. These threads are common for all video streams and distribute resources evenly across all video streams regardless of their workload (except for the video streams without faces in the frame).matching_threads_count
: the number of threads for comparing templates with the database. Like processing threads, these threads distribute the workload evenly across all video streams.age_gender_estimation_threads_count
: the number of threads for age and gender estimation. Like processing threads, these threads distribute the workload evenly across all video streams.emotions_estimation_threads_count
: the number of threads for emotion estimation. Like processing threads, these threads distribute the workload evenly across all video streams.enable_sti
: the flag enables short time identification.sti_recognition_threshold
: the distance threshold for short time identification.sti_outdate_time
: time period in seconds for short time identification.
All available configuration files are stored in the conf folder of the Face SDK distribution.
Provide Video Frames
To provide video frames, call VideoWorker.addVideoFrame
. This method is thread-safe, so you can provide frames from different threads without additional synchronization.
The method arguments are frame
(frame color image), stream_id
(video stream identifier) and timestamp_microsec
(frame timestamp in microseconds). The method returns frame_id
(integer frame identifier) which will be used in callbacks to identify this frame.
Callbacks
To return frame processing results, VideoWorker
uses a set of callbacks. Processing data received from different threads is organized into a data
structure, passed as a callback argument. Attributes of data
structure are different for each type of callback.
The following types of callbacks are available:
- TrackingCallbackU (Face tracking, estimation of gender, age and emotions, liveness estimation)
- TrackingLostCallbackU (Face tracking)
- TemplateCreatedCallbackU (Creating templates)
- MatchFoundCallbackU (Face identification)
- StiPersonOutdatedCallbackU (Short-time identification)
To add a callback, use the VideoWorker.add(callback name)
method. The method returns callback_id
(integer identifier of the callback). To remove a callback, use the VideoWorker.remove(callback name)
method with callback_id
as an argument.
Face Tracking
If VideoWorker
is used only for face tracking, specify matching_thread=0
and processing_thread=0
, and apply the standard Face Detector license. To create VideoWorker
for one stream, set the streams_count=1
parameter.
You can use two callbacks for face tracking:
VideoWorker.TrackingCallbackU (Tracking callback)
While a person is in the camera's field of view, the system forms a person's track (a sequence of video stream frames, which belongs to the same person). Each track is assigned its own integer identifier (track_id
).
Video stream frames are passed to the tracking conveyor - a structure responsible for synchronizing and distributing frames among processing threads. After processing each frame, the Tracking callback is called, which returns the tracking results. Tracking results are passed in TrackingCallbackData
structure (C++, Java, C#, Python).
Tracking callbacks with the same stream_id
are called in ascending frame_id
order. Therefore, if a callback with stream_id = 1
and frame_id = 102
was received immediately after a callback with stream_id = 1
and frame_id = 100
, it means the frame with frame_id = 101
was skipped for the video stream 1.
VideoWorker.TrackingLostCallbackU (TrackingLost callback)
After track is lost (for example, when a person leaves the frame), this callback returns the best sample and face template. The best sample can be empty if the weak_tracks_in_tracking_callback
configuration parameter is enabled. Processing results are passed in TrackingLostCallbackData
structure (C++, Java, C#, Python).
This is the last callback called for the pair of <stream_id, track_id>
. That is, after this callback no Tracking, MatchFound or TrackingLost callback for this stream_id
can contain a sample with the same track_id
.
For each pair of <stream_id, track_id>
, mentioned in the Tracking callback, there is exactly only one TrackingLost callback, except for the tracks removed during the reset of video stream with VideoWorker.resetStream
method.
The method returns track_id_threshold
, an integer that means that all tracks removed during the resetStream had track_id < track_id_threshold
, and all new tracks will have track_id >= track_id_threshold
. After resetStream returns control, no callbacks related to the previous tracks will be called, including TrackingLost callback.
Exceptions in the callbacks will be caught and rethrown in the VideoWorker.checkExceptions
member function. Therefore, you need to call the VideoWorker.checkExceptions
method from time to time to check for errors.
Do not call the methods that change the state of VideoWorker
inside the callbacks to avoid a deadlock. That is, only the VideoWorker.getMethodName
and VideoWorker.getStreamsCount
member functions are safe for calling in callbacks.
Creating Templates
To track faces and create biometric templates, you need to specify matching_thread=0
and processing_thread>0
, and apply the Video Engine Standard license. To create VideoWorker
for one stream, set the parameters streams_count=1
, processing_threads_count=1
and matching_threads_count=0
.
VideoWorker.TemplateCreatedCallbackU (TemplateCreated callback)
VideoWorker.TemplateCreatedCallbackU
returns template generation results in TemplateCreatedCallbackData
structure (C++, Java, C#, Python). This callback is called, whenever a template is created within the VideoWorker
.
Template creation is enabled for all video streams by default. You can enable/disable template creation for a specific video stream using the VideoWorker.disableProcessingOnStream
and VideoWorker.enableProcessingOnStream
methods.
Estimation of Age, Gender, and Emotions
To estimate age and gender, specify the parameter age_gender_estimation_threads_count > 0
. To estimate emotions, set emotions_estimation_threads_count > 0
.
The data about age, gender, and emotions is returned in VideoWorker.TrackingCallbackU
. The data about emotions is constantly updated, and the data about age and gender is updated only if there is a sample of better quality.
By default estimation of age, gender, and emotions is enabled for all video streams.
To disable estimation of age, gender, and emotions on a specified video stream, use the following methods:
VideoWorker.disableAgeGenderEstimationOnStream
(age and gender)VideoWorker.disableEmotionsEstimationOnStream
(emotions)
To enable estimation of age, gender, and emotions on a specified video stream again, use the following methods:
VideoWorker.enableAgeGenderEstimationOnStream
(age and gender)VideoWorker.enableEmotionsEstimationOnStream
(emotions)
Liveness Estimation
Active Liveness
This type of liveness estimation presupposes that a user needs to perform certain actions. For example, "turn the head", "blink", etc.
To enable active liveness, set the enable_active_liveness
parameter in VideoWorker
configuration file to 1
. All faces for identification will then take several checks:
- SMILE: smile. To use the
SMILE
check, specify the number of threads in theemotions_estimation_threads_count
parameter of theVideoWorker
object. - BLINK: blink
- TURN_UP: turn your head up
- TURN_DOWN: turn your head down
- TURN_RIGHT: turn your head to the right
- TURN_LEFT: turn your head to the left
- PERSPECTIVE: face position check (move your face closer to the camera)
To enable/disable checks, open the configuration file active_liveness_estimator.xml
and specify the values 0
(check disabled) or 1
(check enabled) for one or more checks: check_smile
, check_perspective
, check_blink
, check_turn_up
, check_turn_down
, check_turn_right
, check_turn_left
.
Also, in the active_liveness_estimator.xml
configuration file, you can configure the following parameters:
check_count
: number of enabled checks. If the specified number is less than the number of enabled checks, the system will randomly select several checks to take, and if it is greater than the number of enabled checks, some checks will be repeated several times.max_frames_wait
: check waiting time (in frames). If the check has not started during this time, it is considered failed, and the statusCHECK_FAIL
is returned.rotation_yaw_threshold
: threshold for passing the check. To pass the check, you need to turn your head at the specified angle.rotation_pitch_threshold
: threshold for passing the check. To pass the check, you need to turn your head at the specified angle.blinks_threshold
: threshold for passing the check. 0.3 is default value, 0 is eye closed, 1 is eye open. Blinking is counted if the set value is equal to or less than 0.3.blinks_number
: number of blinks required to pass the check.emotion_threshold_smile
: threshold for checkingsmile
. The smaller the specified value, the weaker you need to smile to pass the test.face_align_angle
: before starting the check and between the checks, the face must be in a neutral position (frontal to the camera). The parameter value is the maximum allowable deviation of the head rotation angles to start the check.
The order of checks can be random (this mode is selected by default), or set during initialization (by creating a list of non-repeating checks).
See an example of setting a checklist in the video_recognition_demo
sample (C++/C#/Java/Python).
The check status is returned in the active_liveness_result
attribute of Tracking callback. This attribute contains the following fields:
verdict
: status of the current check (theActiveLiveness.Verdict
object)type
: type of check (theActiveLiveness.LivenessChecks
object)progress_level
: check passing progress - a number in the range of [0,1]
The ActiveLiveness.Verdict
object contains the following fields:
ALL_CHECKS_PASSED
: all checks passedCURRENT_CHECK_PASSED
: current check passedCHECK_FAIL
: check failedWAITING_FACE_ALIGN
: waiting for neutral face positionNOT_COMPUTED
: liveness is not estimatedIN_PROGRESS
: check is in progress
Face Identification
For face tracking, template creation and matching with the database, specify matching_thread>0
and processing_thread>0
in VideoWorker
constructor parameters, and apply the Video Engine Extended license. To create VideoWorker
for one stream, set the parameters streams_count=1
, processing_threads_count=1
and matching_threads_count=1
.
To set or change the database, use the VideoWorker.setDatabase
method. This method can be called any time. Pass elements
(a vector of base elements) and acceleration
(type of search acceleration) as arguments.
Identification works the following way: a template extracted from the sample of the tracked face, is compared with templates from the database. If the distance to the closest database element is less than distance_threshold
specified for this element, then a match is found.
VideoWorker.MatchFoundCallbackU
callback is called after N
consecutive matches with the elements that belong to the same person, and returns the matching result. The result is passed in the structure MatchFoundCallbackData
(C++, Java, C#, Python). The N
number can be set in the configuration file in the <consecutive_match_count_for_match_found_callback>
parameter.
You can set the <not_found_match_found_callback>
parameter to 1
to enable this callback after N
consecutive mismatches (mismatch happens when the closest element is beyond its distance_threshold
). In this case match_result
of the first element in VideoWorker.MatchFoundCallbackData.search_results
will be at zero distance, and the person_id
and element_id
identifiers will be equal to VideoWorker.MATCH_NOT_FOUND_ID
.
This callback will be called after at least one Tracking callback and before a TrackingLost callback with the same stream_id
and track_id
.
The maximum number of elements returned in the VideoWorker.MatchFoundCallbackData.search_results
is set in the configuration file in the search_k
parameter. This number can be changed by the FacerecService.Config.overrideParameter
method. Click here for the examples.
Short-Time Identification
Short time identification does not require a separate license. To use this function, specify at least one thread for template creation (processing_thread>0
).
Short time identification (STI) is used to recognize a person who was in the frame some time ago, even if this person is not in the database, or identification is disabled. With STI enabled a person who was tracked, lost, and then tracked again within, for example, one minute, will be recognized as the same person.
To enable STI, use the enable_sti
flag in the parameters of the VideoWorker constructor.
When a person enters the frame, the system creates a sample, used to extract a biometric template. Then, this template is compared with the face templates of people who left the frame no more than sti_outdate_time
seconds ago (sti_outdate_time
is set in the parameters of the VideoWorker constructor).
Matched tracks are grouped as sti_person
. ID of this group (sti_person_id
) is equal to the track_id
value of the first element that formed the group sti_person
.
When a specific group sti_person
exceeds the specified period sti_outdate_time
, VideoWorker.StiPersonOutdatedCallbackU
is called. The result is passed in the structure StiPersonOutdatedCallbackData
(C++, Java, C#, Python).