gaze-estimation-adas-0002#
Use Case and High-Level Description#
This is a custom VGG-like convolutional neural network for gaze direction estimation.
Example and Gaze Vector Definition#
The network takes three inputs: square crop of left eye image, square crop of right eye image, and three head pose angles - (yaw, pitch, and roll) (see figure). The network outputs 3-D vector corresponding to the direction of a person’s gaze in a Cartesian coordinate system in which z-axis is directed from person’s eyes (mid-point between left and right eyes’ centers) to the camera center, y-axis is vertical, and x-axis is orthogonal to both z,y axes so that (x,y,z) constitute a right-handed coordinate system.
Specification#
Metric |
Value |
---|---|
GFlops |
0.139 |
MParams |
1.882 |
Source framework |
PyTorch* |
Validation Dataset#
Two random held out individuals from an internal dataset containing images of 60 people with different gaze directions.
Validation Results#
The accuracy of gaze direction prediction is evaluated through the use of MAE of angle (in degrees) between the ground truth and predicted gaze direction.
Dataset |
MAE, degrees |
Standard deviation of AE, degrees |
---|---|---|
Internal dataset |
6.95 |
3.58 |
Inputs#
Blob, name:
left_eye_image
, shape:1, 3, 60, 60
in the formatB, C, H, W
, where:B
- batch sizeC
- number of channelsH
- image heightW
- image width
Blob, name:
right_eye_image
, shape:1, 3, 60, 60
in the formatB, C, H, W
, where:B
- batch sizeC
- number of channelsH
- image heightW
- image width
Blob, name:
head_pose_angles
, shape:1, 3
in the formatB, C
, where:B
- batch sizeC
- number of channels
Outputs#
The net output is a blob with name gaze_vector
and the shape: 1, 3
, containing Cartesian coordinates of gaze direction vector. Please note that the output vector is not normalizes and has non-unit length.
Demo usage#
The model can be used in the following demos provided by the Open Model Zoo to show its capabilities:
Legal Information#
[*] Other names and brands may be claimed as the property of others.