Next Article in Journal
A Review of Recent Advances in Human-Motion Energy Harvesting Nanogenerators, Self-Powering Smart Sensors and Self-Charging Electronics
Previous Article in Journal
Calibration of a Low-Cost Methane Sensor Using Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Model-Based 3D Gaze Estimation Using a TOF Camera

1
School of Chemical Process Automation, Shenyang University of Technology, Liaoyang 111003, China
2
School of Control Science and Engineering, Dalian University of Technology, Dalian 116024, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(4), 1070; https://doi.org/10.3390/s24041070
Submission received: 18 December 2023 / Revised: 27 January 2024 / Accepted: 2 February 2024 / Published: 6 February 2024

Abstract

:
Among the numerous gaze-estimation methods currently available, appearance-based methods predominantly use RGB images as input and employ convolutional neural networks (CNNs) to detect facial images to regressively obtain gaze angles or gaze points. Model-based methods require high-resolution images to obtain a clear eyeball geometric model. These methods face significant challenges in outdoor environments and practical application scenarios. This paper proposes a model-based gaze-estimation algorithm using a low-resolution 3D TOF camera. This study uses infrared images instead of RGB images as input to overcome the impact of varying illumination intensity in the environment on gaze estimation. We utilized a trained YOLOv8 neural network model to detect eye landmarks in captured facial images. Combined with the depth map from a time-of-flight (TOF) camera, we calculated the 3D coordinates of the canthus points of a single eye of the subject. Based on this, we fitted a 3D geometric model of the eyeball to determine the subject’s gaze angle. Experimental validation showed that our method achieved a root mean square error of 6.03° and 4.83° in the horizontal and vertical directions, respectively, for the detection of the subject’s gaze angle. We also tested the proposed method in a real car driving environment, achieving stable driver gaze detection at various locations inside the car, such as the dashboard, driver mirror, and the in-vehicle screen.

1. Introduction

Gaze-estimation technology has been receiving increasing attention due to its significant application value. This technology has been applied in numerous fields such as computer communication and text input [1], human–computer interaction [2], virtual reality (VR) [3], psychological research [4], advertising and entertainment gaming [5], animation creation [6], and driver distraction monitoring [7]. Eyeball gazing is a form of non-verbal communication. By observing the gaze points of people in images or scenes, key information hidden in the images can be revealed. Before the application of convolutional neural networks (CNNs) in gaze estimation, the primary research methods for gaze estimation were traditional machine learning and methods based on eyeball geometric models. Mora et al. [8] used gaze appearance images as ground truth images, from which they built appearance models and used regression and KNN machine-learning methods to determine the gaze angle of the subjects. Wood [9] used 3D head scans and realistic rendering methods to create a synthetic eye image dataset and trained a gaze-estimation model on this dataset using machine-learning methods. Chen et al. [10] and Li et al. [11] created realistic 3D eyeball models through facial landmarks detection, allowing for 3D gaze estimation with just a single calibration from the subject. Their 3D eye model does not employ the corneal reflection points of external light sources mentioned by Guestrin et al. [12], a method which is capable of estimating the point of gaze only when the head is held completely still. Furthermore, 3D eye-modeling techniques that utilize corneal reflection points are often considered to be intrusive and unfriendly to the subject, as they constitute an invasive approach to gaze estimation.
Following the rapid advancements in the research and application of convolutional neural networks (CNNs) in the fields of computer vision and deep learning, numerous scientists have begun to incorporate CNNs into the domain of gaze estimation. Zhang et al. [13] proposed the first appearance-based deep gaze-estimation method, and MPIIGaze [14] has become a commonly used dataset for gaze estimation among researchers. Subsequently, Zhang et al. introduced a gaze-estimation method based on the full-face images of subjects [15]. Unlike Ref. [13], which only used the eye region images of the subjects as the model input, Ref. [15] used the entire face image as input, achieving more accurate gaze estimation on the MPIIGaze dataset than [13]. Park et al. [16] proposed a novel 3D gaze image representation method. They used a deep convolutional neural network architecture to handle the projections of the eyeball and iris in 3D space onto a two-dimensional map, and regressed on the map to obtain the gaze angle. Bâce et al. [17] combined appearance-based gaze estimation with optical flow in the eye region to jointly analyze single-channel eye movement dynamics.
In the current landscape of appearance-based gaze estimation, the most advanced model on the MPIIGaze dataset is FAZE [18], while the L2CS [19] model achieved the second-highest global performance ranking on both the Gaze360 and MPIIGaze datasets. These models perform exceptionally well in scenarios using visible light cameras. However, in situations such as nighttime driving tasks, models trained on RGB images prove to be ineffective. As Murthy et al. [20] noted, existing appearance-based gaze-estimation models lack generalization in driving tasks and have not been evaluated for deployment in real-world applications. Vasli et al. [21] proposed a geometry-based gaze-estimation method for drivers, utilizing the physical constraints of the automotive cockpit environment. However, their model, like that of Vora et al. [22], focuses on determining the gaze region of the driver rather than providing a specific and precise gaze angle. Shah et al. [23] believe that using visual area analysis to determine whether a driver is distracted is a fundamental concept. Their work achieved high precision in detecting the gaze angle of drivers in both vertical and horizontal directions. However, it is regrettable that they still only used visible light cameras. Vicente et al. [24] conducted commendable early research. They used a near-infrared camera and a visible light camera to track the driver’s line of sight and evaluated the effects during both day and night. They particularly considered obstructions from sunglasses and extreme head pose. Their work was meticulous and reliable.
In contrast to the aforementioned studies, our proposed method for 3D gaze estimation, suitable for automobile drivers, exclusively utilizes infrared images. This approach is instrumental in overcoming the challenges posed by variations in lighting intensity in practical application scenarios, particularly enhancing its utility during nighttime. Considering economic factors in real-world applications, unlike [24], we do not employ multiple cameras, rather, we utilize a single time-of-flight (TOF) camera equipped with an infrared sensor. This study aims to explore a method for implementing 3D gaze estimation on low-resolution TOF cameras (see Figure 1) and to evaluate its accuracy and feasibility in real automotive driving situations. We trained a model using YOLOv8 to detect eye regions and landmarks to obtain the positions of eye landmarks including the pupils and left and right eye corners of the subjects. Then, we performed image distortion correction, followed by the calculation of the 3D coordinates of these eye landmarks using the depth image captured by the TOF camera. We then fit a 3D eyeball model in both horizontal and vertical directions, and on this basis, calculated the final gaze angle. Our primary contributions include:
  • We propose a novel method for fitting a 3D eye model. This approach calculates the 3D center of the eyeball using just two eye corners of a single eye. By employing fewer constraint conditions, it reduces computational burden;
  • We propose an infrared image dataset for gaze estimation, named “Infrared Gaze Dataset (IRGD)”, created using a TOF camera with a low resolution of 300,000 (640 × 480) pixels and a precision of 1% within a working range of 1 m. This dataset features a larger gaze angle range than MPIIGaze, with participants freely moving their heads during the recording process. Our work aims to achieve gaze estimation under natural conditions;
  • Using mean absolute error (MAE) and root mean square error (RMSE) as performance evaluation metrics, we assessed the gaze-estimation method presented in this paper on the benchmark dataset IRGD. We verified that the maximum absolute deviation in gaze angle detection of the proposed model is less than 9 degrees. Furthermore, the model’s reliability in detecting the driver’s line of sight angles was validated in real driving tasks with Toyota commercial vehicles.
Figure 1. The overall process of the proposed model-based 3D gaze estimation using a TOF camera method. The green arrow represents the subject’s gaze direction.
Figure 1. The overall process of the proposed model-based 3D gaze estimation using a TOF camera method. The green arrow represents the subject’s gaze direction.
Sensors 24 01070 g001
The remainder of this paper is organized as follows: Section 2 reviews existing gaze-estimation techniques. Section 3 provides a detailed introduction to our proposed IRGD dataset, comparing it with other publicly available gaze-estimation datasets. Section 4 elaborates on our gaze-estimation method using a TOF camera, including the computation for image distortion correction and detailed modeling of the eyeball model when subjects gaze in horizontal and vertical directions. Section 5 presents the accuracy and evaluation results of the gaze-estimation method proposed in this paper. We also ran five existing state-of-the-art gaze-estimation models on the benchmark dataset IRGD and compared their results with our work. Finally, we collected images of driver gaze during real driving tasks in a Toyota commercial SUV, and evaluated the detection results of the proposed model on these images. Section 6 concludes our work and provides prospects for further research.

2. Related Works

In this chapter, we briefly reviewed existing gaze-estimation methods and works related to the method or content of this paper, especially the facial landmarks detection and eyeball center localization methods.

2.1. Appearance-Based Gaze Estimation

Different classifications of existing gaze-estimation methods will yield different results when viewed from different angles. From the principle of gaze-estimation models, existing methods can be divided into model-based methods and regression-based methods [25], and regression-based methods can be further subdivided into feature-based methods and appearance-based methods. Feature-based methods learn the mapping function from eye features to gaze direction, with the offset vector between the inner eye corner and the pupil center being the most commonly used feature mapping. However, researchers have discovered a fact: when the tester looks in different directions, the inner eye corner will move [26]. Therefore, the inner eye corner cannot serve as a stable anchor point. Some researchers use the nostril position as the feature vector for regression, but thus far no researchers have confirmed which set of facial feature points is the most ideal [9]. Appearance-based methods learn the mapping relationship from the appearance of the eye to the gaze point or gaze angle, avoiding feature tracking. Park et al. [27] believe that there is a strong relationship between what the user is looking at and the appearance of the user’s eyes. At present, the latest appearance-based methods are almost all implemented based on CNNs. Another study by Park et al. used a CNN architecture to train a stacked hourglass network on the synthetic image dataset UnityEyes to estimate eye area feature points, and on this basis, the gaze direction was obtained through regression [28]. However, as they stated, the most effective appearance-based methods are black-box solutions and understanding why and when they work may be a challenge. Furthermore, although the human eye indeed has three degrees of freedom of motion, due to the physiological relationship between the human eye and the head, people always turn their heads before moving their eyes when they gaze in a certain direction. Therefore, appearance-based gaze estimation is easily affected by head pose.

2.2. Model-Based Gaze Estimation

Compared with appearance-based methods, model-based methods can obtain more accurate gaze estimates, but they rely on high-resolution images to fit and track local features used to estimate geometric parameters [29]. A common practice is to use a high-resolution infrared camera to capture the corneal reflection points (Purkinje images) of the subject because the pupil of the eyeball can reflect infrared light well and thus the pupil position can be easily identified on infrared images [30]. By fitting the eyeball with the pupil and corneal reflection points, the direction of the subject’s gaze can be detected with high accuracy on high-resolution infrared images.
Scientists introduced the concepts of visual axis and optical axis to analyze the gaze direction of the human eye. The angle between the visual axis and the optical axis is called the kappa angle, which is approximately 5°. Traditional model-based gaze-estimation methods need to use this physical relationship between the optical axis and the visual axis to solve for the visual axis. Chen et al. [10] proposed a model-based 3D gaze-estimation eyeball modeling method. They used a generic 3D face model composed of six rigid facial points to obtain the 3D coordinates of the subject’s facial landmarks and then calculated the 3D coordinates of the canthus midpoint, eyeball center point, pupil point, and corneal center point of the subject’s single eye in order to calculate the optical axis, and then calculate the 3D visual axis and gaze point position. LI et al. [11] used a research method similar to [10]. Although this traditional model-based gaze-estimation method avoids the intrusion on the subjects as in [12], their solution process is very complex and computationally intensive. Taking the work of [10] as an example, in the calibration process of the subject, for N calibration points, there are 7N equations, with 6N+5 unknowns. Compared with these previous works, we only use the two canthus points of the subject’s single eye to approximate the position of the center of the 3D eyeball. This simple solution method is feasible in our research.
In subsequent model-based gaze-estimation methods, the gaze vector is defined as a line passing through the center of the eyeball to the iris [31]. This definition ignores the kappa angle between the optical axis and the visual axis, omits the calculation of deriving the visual axis from the optical axis, and is equally effective in detecting the gaze direction of the subject. The advantage of the model-based method is that it can consider changes in head pose more effectively than appearance-based methods [30]. This paper, like [8], uses a camera with depth information for gaze estimation under free head movement. The difference is that the depth camera we use also has infrared images.

2.3. Facial Landmarks Detection

In the current research, a commonality between model-based methods and appearance-based methods is the need for facial landmarks detection on the subject before gaze estimation. The purpose of detecting facial landmarks in appearance-based gaze-estimation methods is to calculate the head pose of the subject [30] and to standardize the image in conjunction with the head pose before inputting the subject’s gaze image into the neural network [13,15,32,33], to reduce the negative impact of head pose on gaze estimation. Ref. [15] used Alexnet to perform appearance-based full-face gaze estimation on 37,667 images with facial landmarks, Ref. [30] used the open-source model of OpenFace [34] to detect the facial landmarks of the subjects, and ref. [28] used the Tensorflow framework to train a stacked hourglass network for detecting eye region feature points on the annotated synthetic dataset UnityEyes.
The purpose of facial landmarks detection in model-based gaze-estimation methods is to fit and solve the positions of each eye component. Ref. [24] uses the SDM tracker to detect the positions of seven eye landmarks, including the pupil point, on a single-eye image of the subject and calculates the 3D coordinates of the pupil based on this. In existing research methods for estimating the gaze of car drivers, Ref. [23] uses an improved YOLOv4 face detector to detect the driver’s face. The YOLO series of neural networks are one-stage object-detection models, which are widely used in the field of artificial intelligence due to their fast detection speed, high accuracy, and ease of deployment. Unlike [23], this paper uses the open-source YOLOv8-pose [35] module to train a model for detecting the eye area and landmarks of the subject in infrared images on the IRGD dataset labeled with eye corner points, pupil points, etc.
YOLOv8 is an open-source convolutional neural network encompassing functionalities for object detection, instance segmentation, and landmark detection. By default, it utilizes the SiLU (sigmoid linear unit) activation function. The architecture of YOLOv8 comprises two primary components: the backbone and the head. The backbone, a modified version of CSPDarknet53, primarily serves as a feature extractor, responsible for deriving features from the input images. The head, consisting of multiple convolutional layers followed by a series of fully connected layers, is tasked with predicting the bounding boxes, object detection scores, and class probabilities for detected objects within the image. YOLOv8 employs CIoU (complete intersection over union) and DFL (distance-IoU front-end loss) for bounding box location loss, alongside BCE (binary cross-entropy) for classification loss. The network is standardly configured with 53 layers. In this study, the YOLOv8s neural network was utilized to train a subject’s eye landmark detection model. The images input into the network were of a default size of 640 × 480 pixels. Upon invoking the trained landmark detection model for inference on single-subject gaze images within the test set, the model outputs a tensor of dimensions [2,27], where 2 corresponds to the detected left and right eyes of the subject, while 27 represents the detected attributes of a single eye. These attributes include the bounding box’s position (x, y, w, h), the confidence of the bounding box, the probabilities associated with the categories, and the 2D coordinates and confidences of seven landmark points.

Data Augmentation

Data augmentation is an effective method to prevent overfitting in eye landmark detection models. As illustrated in Figure 2, we augment the images from the IRGD dataset using computer vision methods such as Gaussian filtering, mean filtering, median filtering, brightness enhancement, contrast enhancement, adding salt and pepper noise, horizontal flipping, rotation, and scaling. Through this approach, the amount of data used to train the eye landmark detection model in this study was increased by a factor of six from the original 75,000 images in the IRGD dataset, resulting in a total of 525,000 images. Furthermore, this paper divides the dataset into a training set and a testing set in a 4:1 ratio. These data augmentation techniques serve as a regularization mechanism for training neural network models in facial landmark detection. During the training process, data augmentation introduces randomness into the data fed into the network. This randomness is then mitigated in the test set to enhance the network’s generalization ability. Implementing this practice is expected to improve the robustness of the facial landmark detection model in complex scenarios with variable lighting conditions.
The eye landmarks detection model we trained, based on YOLOv8, outputs the coordinates of seven eye landmarks for a single eye image of the subject, including the left and right eye corners, the pupil, and the upper and lower eyelid points (see Figure 3). To improve work efficiency, we annotate the eye landmarks of participants on the IRGD dataset in a semi-automatic manner to quickly obtain training set data.

2.4. 3D Eyeball Center Positioning Method

The center of the eyeball is an eye component hidden outside the image, which cannot be directly observed in the gaze image like the eyelids and pupils. Existing model-based gaze-estimation literature varies in their computation and fitting methods for the 3D center of the eyeball. Refs. [10,11] detect the head pose angle and sequentially calculate the 3D pupil point coordinates and 3D corneal center point coordinates in the facial landmarks, then obtain the 3D coordinates of the eyeball through individual calibration of the subject. Their calculation method is quite complex. Vicente et al. [24] assumed that the eyeball is spherical and is at a rigid point relative to the head model, thereby approximating the 3D coordinate position of the eyeball. However, this is merely an approximate estimation method and unfortunately, they have not proven the effectiveness of their method.

2.4.1. Localization Method Based on Eye Corners and Eyeball Radius

Modern medical data on human eyeball anatomy indicate that the average radius of the human eye is 12 mm. In model-based gaze estimation, a novel method for locating the eyeball center involves obtaining the 3D coordinates of the left and right eye corners to compute the 3D midpoint of the eye corners. The eyeball center is then defined as the position 12 mm away from the eye corner midpoint in the Z-axis direction of the camera coordinate system, thereby transforming the model-based method into a calibration-free gaze-estimation method. Ref. [30] used the open-source OpenFace model to detect the head pose of the subject, combined with the 3D coordinates of the left and right eye corners and the constant value of the 12 mm eyeball radius to calculate the 3D coordinates of the subject’s eyeball center. The subject’s gaze angle is then computed based on this.
We replicated the eyeball center localization method mentioned in [30]. To eliminate the interference of head pose on the eyeball center localization method mentioned in [30], we calculated the horizontal gaze angle (yaw) on the gaze images of the subject’s lizard movement collected by the TOF camera (see Figure 4). The feasibility of this 3D eyeball center localization method was judged by the absolute deviation size between the calculated gaze angle and the ground-truth angle. We manually annotated the left and right eye corner points and pupil points of these six images and strictly calculated the 3D eyeball center coordinates according to the eyeball center localization method of [30]. We found that within ±30° of small gaze angle calculation, the eyeball center localization method mentioned in [30] can obtain very accurate gaze estimation, with an absolute deviation angle value within 5°. However, this eyeball center localization method fails in gaze estimation at large angles greater than 30°.

2.4.2. 3D Eyeball Fitting Based on Nonlinear Least Square Method

In the many methods of fitting the 3D eyeballs, we noticed a type of nonlinear least squares method. Like most studies [21,24], we assume that the eyeball is spherical. We replicated the method of fitting the eyeball in [36], solving for the coordinate values of the 3D eyeball center through eight marked points on a single-eye image (see Figure 5), with the number of marked points satisfying the condition of array size ≥ 5 required for the solving process. We collected images at approximate distances of 200 mm, 300 mm, and 400 mm from the TOF camera to the subject’s head, and the radius values of the fitted 3D eyeball models are shown in Table 2.
The nonlinear least squares method has demonstrated highly accurate fitting results in fitting the 3D eyeball model. As shown in Table 2 the fitted eyeball radius is very close to the average value of 12 mm of the human eye. However, it relies on high-resolution eye images and a large number of marked points. The more marked points involved in fitting the eyeball and the more dispersed the distribution of marked points on the sclera, the more accurate the solved eyeball model. However, obtaining as many marked points as possible on low-resolution images is a challenge. We collected detailed images of the eyes of subjects wearing glasses and not wearing glasses at different distances between 200 mm and 500 mm from a 300,000-pixel TOF camera (see Figure 6). When the distance between the subject’s eyes and the TOF camera exceeds 300 mm, the contrast between the sclera and cornea severely decreases, and the only easily observable details in the subject’s eye area are the corners of the eyes and the pupils. Therefore, the eyeball center locating method based on nonlinear least squares does not apply to the low-resolution infrared images used in this study. We propose a method to fit the 3D eyeball center using only the left and right corner points of a single eye (see Section 4.3). The subject only needs to undergo one calibration to complete the solution calculation.

3. Dataset

To achieve stable 3D gaze estimation under various complex backgrounds and lighting conditions, we used a TOF camera equipped with an infrared sensor to record the benchmark dataset IRGD for gaze estimation. The use of an infrared camera can overcome the negative impact of uneven changes in lighting intensity on gaze estimation. The gaze-estimation model based on infrared images can operate in two different scenarios, day and night, which is extremely beneficial for the gaze estimation of car drivers and has been reflected in the work of [24].

3.1. Existing Dataset

Currently, many public datasets for gaze estimation have been generated in previous studies. Due to the different purposes and focuses of these studies, the characteristics of their public datasets also vary. Siegfried et al. [37] researched gaze estimation in conversation and operation scenarios and recorded the Gaze_VFOA dataset, which is a video dataset including three sub-datasets such as ManiGaze [38]. The ManiGaze dataset is especially used for assessing the gaze estimation of subjects during human–computer interaction. Other video datasets include EYEDIAP [39] and EVE [27], of which EVE was recorded on a computer monitor using a network port camera, aiming to estimate the gaze of subjects watching videos. Kellnhofer et al. recorded the large-scale gaze tracking dataset Gaze360 [40] which includes indoor and outdoor environments. They used the AprilTag image that was also used in the ManiGaze dataset during the recording process, which is convenient for marking target locations in 3D space. Similar to Gaze360, the MPIIGaze dataset is also used for gaze estimation in outdoor environments. This dataset is collected from subjects gazing at laptop screens in daily life. Due to the limitations of the recording scenario, the gaze angle range of the MPIIGaze dataset is relatively small. Due to the privacy protection set by the authors, researchers can only access the images of the eye area in the MPIIGaze dataset. If researchers want to explore the role of information from other parts of the face in gaze estimation, they can access the MPIIFaceGaze dataset, which comes from the same source as MPIIGaze. The latter is based on the former and provides complete facial area images and facial feature point annotations. For research on gaze estimation in extreme head poses, the ETH-XGaze dataset is often used. This dataset provides high-resolution images under extreme head pose and gaze directions with over a million samples [32]. The ETH-XGaze dataset contains images with a resolution of 24 MP (6000 × 4000 pixels). Another dataset with high-resolution images is Columbia [41], with an image resolution of 18 MP (5184 × 3456). However, the Columbia dataset only has 5880 images, and compared to the MPIIGaze dataset, Columbia has a smaller gaze angle range. Krafka et al. proposed a large-scale dataset, GazeCapture [42], for eye-tracking on mobile phones and tablets. The gaze-estimation model, iTracker, they developed based on the GazeCapture dataset was used in a commercial mobile phone software. Because this work is based on gaze estimation on small planes such as mobile phones or tablets, GazeCapture, like MPIIGaze, has the disadvantage of a smaller gaze angle range.
Synthetic datasets are another type of dataset commonly used in gaze-estimation work. Unlike real human eye gaze images, synthetic datasets are based on the scanning and reconstruction of real human eyes using 3D rendering models. However, this does not mean that synthetic datasets are not practical. In fact, synthetic datasets avoid the angle errors in the recording process of real eye datasets. For example, in our research experience, due to physiological factors, the subjects’ eyes inevitably become tired during the long data-recording process, leading to errors in gaze angle. The recording process of real human eye datasets requires the subjective cooperation of the subjects. Sugano et al. [33] proposed a synthetic dataset, UT Multi-view, which contains multi-view gaze images reconstructed from real human eyes. Other synthetic datasets include SynthesEyes and UnityEyes. SynthesEyes contains 11,382 synthesized close-up eye images and 10 dynamic eye area models; these synthesized images have achieved photo-realistic quality [43]. The UnityEyes dataset is generated by rendering 3D eye area models, capturing a large number of eye appearance variations, and can estimate challenging eye gaze images [44]. Wood has created open-source software to generate the UnityEyes dataset, and researchers can use this software to generate synthetic eye images with custom resolutions.
Most existing popular gaze datasets are RGB images with high resolution. However, image blur is a factor that leads to inaccuracies in the analysis of human social interactions involving gaze [45]. Therefore, it is particularly necessary to explore gaze estimation on low-resolution images. The low-resolution IR image dataset proposed in this paper addresses the challenge of RGB images in existing public datasets being easily affected by lighting conditions. A summary and comparison of the popular open-source datasets and the IRGD dataset proposed in this study can be seen in Table 3.

3.2. Data Collection Procedure

The TOF camera can simultaneously capture infrared images and depth images. The infrared images provide 2D information about the human eye during the gaze process, while the depth images provide 3D information about the 2D eye landmarks in the camera coordinate system. TOF cameras have been widely used in the automotive intelligent cockpit industry for driver behavior detection such as smoking and phone calling. To save economic costs for car manufacturers, we perform model-based 3D gaze estimation based on the automotive-grade TOF camera with a resolution of 300,000 pixels and an accuracy of 1% (1 m), which is widely used in cars.
Before officially recording the IRGD dataset, we created and calibrated a standard plane for gaze estimation using a level, based on a standard wall (see Figure 7a). This standard plane consists of a grid of squares with a side length of 200 mm, arranged in 17 columns and 11 rows. We ordered the grid, and the intersections of the grid on the standard plane are the gaze points, which are located in the coordinate system of the standard plane. To increase the diversity of lighting conditions during data recording, we chose the location of the standard plane at the indoor doorway to approach the lighting changes in the external environment and organized participants to record data at three different times: morning, noon, and evening. The TOF camera is fixed on the standard plane and does not change its position during data recording. The size of the standard plane is 3.4 m × 2.2 m, which is sufficient to meet the recording of gaze data at extreme angles. We created a standard plane coordinate system X w Y w Z w and camera coordinate system X c Y c Z c (see Figure 7b). Since the fixed position of the camera on the standard plane is known, the 3D coordinate values of each gaze point in the camera coordinate system can be obtained through the transformation relationship between the two coordinate systems. The specific transformation calculation is as follows:
P x = ( s × ( n u m ( c o l × n u m c o l ) 1 ) c a m e r a x )
                                              P y = s × n u m c o l c a m e r a y
P z = c a m e r a z
where ( P x ,   P y , P z ) represents the 3D coordinates of the gaze point on the standard plane in the coordinate system of the TOF camera. ( c a m e r a x , c a m e r a y , c a m e r a z ) denotes the fixed position of the TOF camera in the coordinate system of the standard plane. s represents the size of the regular quadrilateral grid on the standard plane, which is set to 200 mm in this study. n u m represents the gaze point number, and c o l represents the number of regular quadrilateral grids per row on the standard plane.
To establish a tacit understanding with the participants during the data-recording process, we developed a camera acquisition software and ran it on a laptop. Unlike Columbia and UT Multi-view, we did not use a head fixation device, allowing participants in the data recording to perform natural eye movements as described in the EVE dataset, without any constraints. Participants freely gaze at the gaze points on the standard plane and inform the camera acquisition software of the current gaze point number via voice commands. The camera acquisition software captures the participant’s gaze image 0.5 s later. Figure 8 shows some examples from the IRGD dataset we recorded.

3.3. Dataset Characteristics

The process of recording the dataset was quite laborious [28], and the creation of the IRGD dataset proposed in this paper took a considerable amount of time. We ultimately organized 15 participants to join this work, including 7 individuals wearing nearsighted glasses, 9 males, and 6 females, all of whom were Asian. Each participant recorded data at five different distances from the TOF camera: 200 mm, 300 mm, 400 mm, 500 mm, and 600 mm (see Figure 8). The TOF camera had a sampling frame rate of 30 fps and a sampling interval of 4, capturing 1000 IR images and 1000 depth images for each participant at each distance, resulting in a total of 75K IR images and their corresponding 75K depth images in the IRGD dataset. These ample data laid a solid foundation for training the eye landmarks detection model using YOLOv8. To provide a comprehensive description of the characteristics of the IRGD dataset, we discuss it in detail from the following aspects.

3.3.1. Image Resolution

The IRGD dataset proposed in this paper includes both IR and depth images. These two types of images were captured by an automotive-grade TOF camera with 300,000 pixels, both of which have a low resolution of 640 × 480.

3.3.2. Gaze Direction

The IRGD dataset is composed of gaze images from 15 participants, each of whom sequentially gazed at 35 target gaze points on a standard plane. These gaze points form a gaze area of 1.4 m × 0.8 m on the standard plane, with the participants’ gaze angles in the horizontal and vertical directions within this gaze area being ±50°.

3.3.3. Light Condition

In this study, a lux meter was used to record the ambient illumination levels during data acquisition. To enhance the adaptability of the gaze-estimation model to variations in illumination intensity, the benchmark dataset IRGD was recorded under a wide range of lighting conditions, from complete darkness (0 lux) to illumination levels nearing outdoor sunlight exposure (170,000 lux). The automotive-grade TOF camera utilized in this work demonstrated stable performance across this spectrum of illumination intensities.

3.3.4. Head Pose

Existing work on head pose handling can be divided into two categories: one is to normalize the images using the detected head pose angles before training the network to reduce the impact of head pose on gaze estimation [15], and the other is to use head pose to compensate for the estimated gaze angle [24]. The head pose handling approach Mora et al. mentioned is more ingenious than that of [15]; they use an RGBD camera and rigid ICP algorithm to track the subject’s head pose frame by frame to generate the subject’s frontal gaze image. This method has the advantage of decoupling the gaze angle from the head pose, significantly reducing the complexity of gaze-estimation problems caused by the head pose [8]. However, for cases where the subject’s eyes are self-occluded at larger gaze angles, it is not possible to restore the subject’s frontal gaze image from the gaze image with the head pose. The gaze-estimation method proposed in this paper calculates the subject’s eyes in 3D space, so it is not disturbed by head pose like traditional 2D-image-based gaze-estimation methods. Our work does not directly regress gaze angles from gaze images, so there is no need to normalize the images like [33]. Since head pose detection is not the goal of this study, we only use existing head pose detection methods to statistically analyze the head pose angles of the IRGD dataset. We use the 6-point facial method [46] and the 12-point method [47] to detect the head poses of participants in the manually annotated IRGD dataset, but the detected results have significant deviations in both direction and angle. Subsequently, we used the state-of-the-art head pose estimation model 6DRepNet [48] on the BIWI dataset to detect and calculate the absolute values of the average head pose angles of participants gazing at 35 gaze points in the IRGD dataset (see Figure 9).

4. Method

In this chapter, we will focus on the distortion correction calculations on the IR gaze images taken by the TOF camera and the methods of fitting the 3D eyeball center in the horizontal and vertical directions. We will discuss the drawbacks of using the TOF camera directly for gaze estimation and the methods to address this issue.

4.1. Image Distortion Correction

Distortion refers to the degree of distortion of the image formed by an optical system relative to the object itself. It is an inherent characteristic of optical lenses, directly caused by the inconsistent magnification between the edge and the center of the lens. Radial distortion is caused by the shape of the lens, while tangential distortion is caused by the tilt of the lens and the imaging plane during the camera assembly process [49]. Thanks to the improvement of camera assembly and production technology, the distortion values of current industrial cameras and network cameras are relatively small, with the former generally having a distortion value of only 1–2 pixels, and the latter only a few pixels. The average distortion value of the TOF camera used in this study is 7 pixels, with different distortion sizes at different positions on the image, and the smallest distortion value in the central area. For gaze-estimation methods on 2D images, image distortion may not affect the accuracy of gaze estimation like other factors. However, our work requires extracting the depth value at the pixel coordinates of the target point on the depth image, and the depth values of different target points may not be the same. Before correcting the distortion of the eye landmarks on the subject’s gaze image, we use OpenCV’s camera calibration tool [50] to obtain the intrinsic matrix and distortion coefficients of the TOF camera:
I n t r i n s i c   M a t r i x = [ f x 0 c x 0 f y c y 0 0 1 ]
D i s t o r t i o n   C o e f f i c i e n t = ( k 1 , k 2 , p 1 , p 2 , k 3 )
where ( k 1 , k 2 , k 3 ) represents the radial distortion coefficient, and ( p 1 , p 2 ) denotes the tangential distortion coefficient.
In Section 2.3, we presented the output results of the eye landmarks detection model trained using YOLOv8. For the 7 landmarks obtained on the single-eye image of the subject shown in Figure 3b, the calculation methods for radial and tangential distortion correction are as follows:
r i 2 = u i 2 + v i 2
u i ´ = u i ( 1 + k 1 r i 2 + k 2 r i 4 + k 3 r i 6 ) + 2 p 1 u i v i + p 2 ( r i 2 + 2 u i 2 ) + u i
v i ´ = v i ( 1 + k 1 r i 2 + k 2 r i 4 + k 3 r i 6 ) + 2 p 1 ( r i 2 + 2 v i 2 ) + 2 p 2 u i v i + v i
where ( u i , v i ) represents the 2D pixel coordinates of the i -th target point among the 7 landmarks detected on the original gaze image, and (   u i ´ , v i ´ ) denotes the pixel coordinates of the i -th landmarks point after distortion correction. We only correct the distortion of the eye landmarks in the subject’s gaze image, rather than every pixel of the image. This approach is beneficial for improving the computational efficiency of gaze estimation.

4.2. Integration of Depth Information

After eye landmarks detection and distortion correction on the IR gaze images captured by the TOF camera, we utilize the depth information from the TOF camera to convert the coordinates of the landmarks in the 2D image coordinate system into 3D coordinates in the camera coordinate system. The detailed calculation method is as follows:
X i = Z i ( u i ´ w 2 ) f x
Y i = Z i ( v i ´ h 2 ) f y
Z i = g i 255 ( f n ) + n
where ( X i , Y i , Z i ) is the 3D coordinate value corresponding to the 2D pixel coordinate ( u i ´ , v i ´ ) of the i -th eye landmark, f x and f y are the intrinsic parameters of the TOF camera. w and h are the width and height of the image, which are 640 and 480, respectively, in this study. g i is the grayscale value at point ( u i ´ , v i ´ ) on the depth image, f is the maximum distance value of the TOF camera’s range measurement, and n is the minimum distance value of the TOF camera’s range.

4.3. Fitting of Eyeball

In this study, we fit the center of the eyeball using the left and right canthus points on the image of a single eye of the subject. We draw inspiration from Abdelrahman et al. [19] who used two fully connected layers to independently regress each gaze angle (yaw and pitch) to improve the prediction accuracy of gaze estimation. We model the eyeball in the horizontal (yaw) and vertical (pitch) directions when the subject is gazing (see Figure 10). We use the coordinates of the subject’s eyeball center fitted in the horizontal and vertical directions to calculate the gaze angles in the two directions, respectively. This study assumes that the left and right canthus points of the subject are on the spherical surface with the eyeball center as the center and the distance to the surface of the eyeball is d i f . For the center of the eyeball in the horizontal direction, its Y-coordinate value is set as the average of the left and right canthus points. Assuming that the radius of the eyeball model in the horizontal direction R 1 and d i f are known, the detailed calculation of the center of the eyeball is as follows:
y e y e = Y 1 + Y 2 2
( x e y e X 1 ) 2 + ( y e y e Y 1 ) 2 + ( z e y e Z 1 ) 2 = ( R 1 + d i f ) 2
( x e y e X 4 ) 2 + ( y e y e Y 4 ) 2 + ( z e y e Z 4 ) 2 = ( R 1 + d i f ) 2
where ( X 1 , Y 1 , Z 1 ) and ( X 4 , Y 4 , Z 4 ) are the 3D coordinates of the left and right canthus points of the subject after distortion correction, respectively. ( x e y e , y e y e , z e y e ) is the calculated center coordinate of the subject’s eyeball in the horizontal gaze direction in the TOF camera coordinate system, and z e y e is taken as the maximum value solution in the calculation.
For the vertical direction of the eyeball model, the X-coordinate value of the eyeball center is set as the average of the left and right eye corner points. The following method is used to solve for the position of the eyeball center:
x e y e ´ = X 1 + X 4 2
( x e y e ´ X 1 ) 2 + ( y e y e ´ Y 1 ) 2 + ( z e y e ´ Z 1 ) 2 = ( R 2 + d i f ) 2
( x e y e ´ X 4 ) 2 + ( y e y e ´ Y 4 ) 2 + ( z e y e ´ Z 4 ) 2 = ( R 2 + d i f ) 2
where R 2 represents the radius value of the participant’s eyeball when gazing in the vertical direction. Assuming R 2 is known, the vertical coordinate ( x e y e ´ , y e y e ´ , z e y e ´ ) of the eyeball center can be calculated. In fact, in the specific calculation process, the value of y e y e ´ depends on the participant’s vertical eyeball pitch state. If the coordinate axis direction of the TOF camera coordinate system X c Y c Z c in Figure 7b is used to define the positive and negative values of the gaze angles in horizontal and vertical directions, then   y e y e ´ takes the minimum value solution when the participant is looking down (pitch in the positive direction), and y e y e ´ takes the maximum value solution when the participant is looking up. This study uses the aspect ratio of the participant’s eye appearance to determine the pitch state of the eyeball. We randomly selected participants for experiments to explore the trend of changes between the vertical gaze angle p i t c h of the eyeball and the aspect ratio p of the participant’s eye appearance. Figure 11 presents our experimental results.
From the research results of Baltrušaitis [51], we find that for gaze-estimation tasks, the accuracy of gaze angle detection in the vertical direction of the participants is lower than that in the horizontal direction, which is a typical phenomenon in gaze estimation. The reason for this phenomenon is that there are fewer pixels on the iris used for Y-axis estimation, and the pupil being covered by the upper and lower eyelids is also a significant factor. Therefore, it is highly meaningful to study the gaze angles of the participants in the horizontal and vertical directions separately. The work of predecessors has validated the value of our research.

4.4. Compute the Aspect Ratio of Eye Appearance

The purpose of calculating and statistically analyzing the aspect ratio of the eye appearance of the participant’s eyes before gaze estimation is to determine the pitch state of the participant’s eyeball in the vertical direction, to obtain the correct solution for the participant’s eyeball center in the vertical direction. This work is primarily accomplished in the calibration phase of Section 4.6. Section 2.3 presents the results of the participant’s eye landmark detection. This paper uses the 2 upper eyelid points and 2 lower eyelid points output from the participant’s single-eye gaze image to calculate the aspect ratio of the participant’s eye appearance. It is worth noting that when manually annotating the training set images for landmark detection, the first upper eyelid point is annotated at 1/3 of the participant’s upper eyelid contour, and the second upper eyelid point is annotated at 2/3 of the upper eyelid contour. Similarly, the first lower eyelid point is annotated at 1/3 of the lower eyelid contour, and the second lower eyelid point is annotated at 2/3 of the lower eyelid contour (see Figure 3b). The eye landmark detection model trained under these annotation position constraints can output 4 eyelid point coordinates that also satisfy the position relationship. This constraint calculates the aspect ratio of the participant’s eye appearance more accurately. The calculation of the aspect ratio of the participant’s eye appearance is as follows:
u a = u 2 ´ + u 3 ´ 2 , v a = v 2 ´ + v 3 ´ 2 , Z a = g a 255 f n + n u b = u 5 ´ + u 6 ´ 2 , v b = v 5 ´ + v 6 ´ 2 ,   Z b = g b 255 f n + n X a = Z a u a w 2 f x , Y a = Z a v a h 2 f y X b = Z b u b w 2 f x , Y b = Z b v b h 2 f y p = X a X b 2 + Y a Y b 2 + Z a Z b 2 X 1 X 4 2 + Y 1 Y 4 2 + Z 1 Z 4 2
where ( u 2 , ´ v 2 ´ ) represents the pixel coordinates of the first upper eyelid point on the distortion-corrected gaze image of the participant, and ( u 3 ´ , v 3 ´ ) represents the pixel coordinates of the second upper eyelid point after distortion correction. ( u a , v a ) is the pixel coordinates of the midpoint of the two upper eyelid points. Similarly, ( u 5 , ´ v 5 ´ ) and ( u 6 ´ , v 6 ´ ) are the pixel coordinates of the first lower eyelid point and the second lower eyelid point after distortion correction, respectively, and ( u b , v b ) is their midpoint. g a and g b are the grayscale values of ( u a , v a ) and ( u b , v b ) on the depth image of the TOF camera, respectively. ( X a , Y a , Z a ) and ( X b , Y b , Z b ) are the 3D coordinates in the TOF camera coordinate system corresponding to ( u a , v a ) and ( u b , v b ) , respectively. p is the calculated aspect ratio of the participant’s eye appearance.

4.5. Drawbacks of TOF Cameras in Operation

The infrared light emitted by the TOF camera has strong penetration and can directly illuminate the pupil area. In addition, the pupil reflects infrared light more than the cornea and lens; hence, the pupil presents higher contrast and brightness in the infrared image, making it easy to detect on the infrared image. The TOF camera relies on the infrared pulse light reflected by the object under test to measure the distance from the object to the camera, but at certain special gaze angles, the subject’s pupil will absorb the infrared pulse light from the TOF camera. This results in the pupil being easy to detect on the 2D infrared image but unable to extract depth information from on the depth image (see Figure 12).
This study does not use the method of obtaining distance information of landmarks on the subject’s face using a 3D face model [10,13,15,32], but calculates the depth value under the TOF camera coordinate system for each eye landmark using the grayscale value on the depth image. We found that the depth values of the left and right eye corners of the subject can always be stably obtained. However, in certain special head poses or gaze angles, due to the pupil absorbing the infrared light from the TOF camera, a ‘black hole’ appears at the corresponding position on its depth image. This results in a grayscale value of 0 for the pupil point on the depth image, thereby making it impossible to solve for the 3D coordinate value of the pupil point.
In response to the operational drawback of the TOF camera being unable to extract the depth information of the pupil point in some subjects’ gaze images, we use the constraint relationship that the distance between the subject’s pupil point and the center of the eyeball is a certain value, combined with the 2D pixel coordinates of the pupil point detected on the IR gaze image, to solve for the 3D coordinates of the pupil. Since we model the subject’s horizontal and vertical gaze directions separately, the distance between the pupil point and the center of the eyeball in these two directions is two different values for the subject. Assuming that the distance in 3D space between the pupil point and the center of the eyeball in the horizontal gaze direction of the subject is d 1 , then:
{ X 7 = Z 7 ( u 7 ´ w 2 ) f x Y 7 = Z 7 ( v 7 ´ h 2 ) f y ( x e y e X 7 ) 2 + ( y e y e Y 7 ) 2 + ( z e y e Z 7 ) 2 = d 1 2
where ( u 7 ´ , v 7 ´ ) represents the pixel coordinate values of the pupil outputted from the eye landmark detection model after distortion correction on the subject’s IR gaze image. ( X 7 , Y 7 , Z 7 ) signifies the calculated 3D coordinate values of the pupil point in the TOF camera coordinate system when the subject gazes in the horizontal direction. The value of Z 7 is obtained by taking the minimum solution.
Assuming that the distance between the pupil point and the center of the eyeball in the vertical gaze direction of the subject is d 2 , we then have:
{ X 7 ´ = Z 7 ´ ( u 7 ´ w 2 ) f x Y 7 ´ = Z 7 ´ ( v 7 ´ h 2 ) f y ( x e y e ´ X 7 ´ ) 2 + ( y e y e ´ Y 7 ´ ) 2 + ( z e y e ´ Z 7 ´ ) 2 = d 2 2
where ( X 7 ´ , Y 7 ´ , Z 7 ´ ) represents the calculated 3D coordinate values of the pupil point when the subject gazes in the vertical direction, and Z 7 ´ is obtained by taking the minimum solution.
After addressing the operational drawbacks of the TOF camera and obtaining the 3D coordinate values of the pupil points under the camera coordinate system, the coordinates of the eyeball center from Section 4.3 can be used to determine the gaze angles y a w and p i t c h   of the subject in the horizontal and vertical directions, respectively:
y a w = 180 π tan 1 ( X 7 x e y e z e y e Z 7 )
p i t c h = 180 π tan 1 ( Y 7 ´ y e y e ´ z e y e ´ Z 7 ´ )

4.6. Person-Dependent Calibration

The average value of the human eyeball radius is 12 mm, and the distance from the center of the eyeball to the pupil is 13.1 mm [11]. Through our measurements, the average value of d i f among different subjects is 5 mm. Using these constants to assign values to the above variables and calculate the final gaze angles in the horizontal and vertical directions, we found that precise calculation results can be obtained at the gaze points of individual subjects, but these constants do not apply to most subjects, and the deviation of the calculated gaze angles far exceeds 10°. The calculation of the subject’s eyeball center and pupil in this paper is based on the known conditions of R 1 , R 2 , d 1 , d 2 , and d i f . These subject-specific parameters need to be obtained through calibration. The research results of Zhang et al. [13] show that among the various challenges faced by unconstrained gaze, the performance gap in gaze estimation caused by individual differences is 40%, which is higher than the 25% of gaze range and 35% of lighting conditions. Therefore, in model-based gaze estimation, individual calibration is also an important task, which can significantly improve the accuracy of gaze estimation.
Chen et al. [10] proposed a 9-point method for calibrating individual-specific eyeball parameters of subjects, while Shah et al. [23] suggested that users could be instructed to look at specific points on the screen to adjust the intrinsic parameters of the eye model to avoid time-consuming calibration procedures. In the experimental results of [25], the more calibration samples there are, the smaller the average deviation of the gaze angles calculated by the algorithm, and as the number of samples increases, the calibrated eyeball parameters approach the ground-truth parameters. Our work combines their calibration approaches:
  • Assuming there is a calibration screen in the subject’s gaze space and as many gaze points as possible are spread across the calibration screen (see Figure 13). The number of gaze points is denoted as M, and the position relationship of each gaze point with the TOF camera is recorded. The ground-truth gaze angle of the eyeball when the subject gazes at the gaze point in the r -th row and c -th column on the calibration screen in the horizontal direction is denoted as α r c , and the ground-truth gaze angle of the eyeball when gazing at the gaze point in the r -th row and c -th column in the vertical direction is denoted as β r c ;
  • Constraining the values of R 1 , R 2 , d 1 , and d 2 within a certain range and traversing these parameters with a step size of 0.1 within this range, we calculate the absolute deviation between the gaze angle values obtained by each set of parameters at each gaze point and the ground-truth angle values, using the aforementioned gaze-estimation calculation method;
  • Computing the average absolute gaze angle deviation corresponding to each set of parameters at M gaze points in N calibrations. The set of R 1 , R 2 , d 1 , and d 2 parameters that yield the smallest average absolute deviation is outputted and denoted as the ground-truth parameters of the subject’s eyeball.
In the specific experiment, since the influence of the d i f value on the gaze angle is consistent with R 1 and R 2 , it can be considered that the calibration effect of d i f is included in the calibration effect of R 1 and R 2 . Therefore, we set d i f as a fixed value of 5 mm and only calibrate R 1 and R 2 . From the perspective of human eye anatomy, the parameters of the human eye must have reasonable values [25]. After fully considering the prior values of the structural parameters of the human eye in medicine, we set the values of R 1 , R 2 , d 1 , and d 2 to [5,40].
In the calibration process, the calculation of the absolute deviation between the gaze angle of the subject’s eyeball obtained by the gaze-estimation method in this paper and the ground-truth angle corresponding to the gaze point on the calibration screen is as follows:
e = | α r c y a w r c ( R 1 d 1 ) | , R 1 [ 5 ,   40 ] , d 1 [ 5 ,   40 ]
e ´ = | β r c p i t c h r c ( R 2 d 2 ) | , R 2 [ 5 , 40 ] , d 2 [ 5 ,   40 ]
where y a w r c ( R 1 d 1 ) represents, within the range of [5,40] when the horizontal radius of the subject’s eyeball is R 1 and the distance from the center of the eyeball to the pupil is d 1 , the horizontal gaze angle calculated by the gaze-estimation method proposed in this paper when the subject gazes at the gaze point at the r -th row and c -th column on the calibration screen. Similarly, p i t c h r c ( R 2 d 2 ) represents, within the range of [5,40] when the vertical radius of the subject’s eyeball is R 2 and the distance from the center of the eyeball to the pupil is d 2 , the vertical gaze angle calculated by the gaze-estimation method in this paper when the subject gazes at the gaze point at the r -th row and c -th column. e   represents the absolute deviation between the horizontal gaze angle calculated under the current set of ( R 1 , d 1 ) parameters and the ground-truth angle, while e ´ represents the absolute deviation between the vertical gaze angle of the subject calculated under the current set of ( R 2 , d 2 ) parameters and the ground-truth angle.
The calculation method for the average gaze-estimation angle deviation of M gaze points on the calibration screen in N calibrations is as follows:
e ( y a w ) ¯ = i = 1 N j = 1 M e i j N M
e ( p i t c h ) ¯ = i = 1 N j = 1 M e i j ´ N M
where e ( y a w ) ¯ represents the average angle deviation in the horizontal direction, e ( p i t c h ) ¯ represents the average angle deviation in the vertical direction, e i j represents the absolute deviation between the gaze angle value corresponding to the j -th gaze point on the calibration screen and the ground-truth angle during the i -th calibration by the subject in the horizontal direction, and e i j ´ represents the absolute deviation between the gaze angle value corresponding to the j -th gaze point on the calibration screen and the ground-truth angle during the i -th calibration by the subject in the vertical direction.
In this study, the calibration screen used for calibrating the subject-specific eye structure parameters is not a fixed scene. The subject should choose an appropriate calibration screen based on the actual application of gaze estimation. For example, when implementing gaze estimation for drivers in a car, the car’s central control screen can be used as the calibration screen. When performing eye tracking for gamers on a computer, the computer monitor can be used as the calibration screen. The selection of the calibration screen should be combined with the subject’s work environment, and the subject should perform complete and as many gaze point calibrations as possible. Figure 14 shows the calibration results of the eye structure parameters for three randomly selected subjects.

5. Experiment

For the model-based 3D gaze-estimation method presented in this paper, we demonstrate its accuracy and effectiveness in detecting the gaze of subjects through several experiments. Following existing experimental approaches, we evaluate the method using a leave-one-out strategy on one subject from the IRGD dataset, while the data from the remaining 13 subjects are used for training the eye landmark detection model based on YOLOv8. We apply the trained model to detect facial images of the evaluation subject at different distances and calculate the detection deviation of the seven landmarks on the subject’s single-eye image by comparing the Euclidean distance between the detected landmarks’ pixel coordinates and the ground-truth landmarks’ pixel coordinates. The final calculated mean deviation is 7.42 pixels. Since four of the seven landmarks (eyelid points) are only used for determining the subject’s eyeball pitch state, when disregarding the eyelid points detection deviation, the model’s average detection deviation for the two eye corner points and the pupil point used to calculate the gaze angle is 1.86 pixels. Therefore, we believe that a YOLOv8 landmark detection model with a good detection accuracy has been trained on the IRGD dataset.

5.1. Pupil Depth Error

We conducted experiments on the IRGD dataset by dividing it into two different groups, male and female, to test the accuracy of the pupil depth calculation method proposed in this paper. In total, nine male subjects participated in 28 experimental sessions within a range of 500 mm–600 mm from the TOF camera, while six female subjects participated in 23 sessions within a range of 400 mm–500 mm. We intentionally aimed to verify the accuracy of pupil depth calculation at different distances. Figure 15 shows the results of the calculated pupil depth values and the ground-truth pupil depth values obtained during the experiments.
For the male subject group, the mean absolute error between the calculated pupil depth values and the ground-truth pupil depth in the horizontal gaze direction was 9.23 mm, and in the vertical gaze direction, it was 17.81 mm. For the female subject group, the mean absolute error in the horizontal gaze direction was 6.73 mm, and in the vertical gaze direction, it was 14.39 mm. Overall, the pupil depth calculation method proposed in this study has a greater accuracy in the horizontal direction than in the vertical direction.

5.2. Deviation of Gaze Angle

Figure 16 presents the results of the gaze-estimation method proposed in this paper for the subjects’ eye tracking. The results in Section 2 indicate that when the subjects are closer to the TOF camera, the clarity of the eye region is higher. To evaluate the gaze-estimation model’s detection results at greater distances, we maintained the nine male subjects at a distance of 500 mm–600 mm from the TOF camera and the six female subjects at a distance of 400 mm–500 mm. These test ranges closely approximate the typical distances between drivers and the common camera installation locations in real driving scenarios. Figure 17 shows the comparison between the estimated gaze angles and the ground-truth angles for male subjects targeting 8 uniformly distributed points within the gaze space and female subjects targeting 16 points in the test environment. To evaluate the accuracy of our method in detecting the subjects’ gaze angles, we used the mean absolute error (MAE) as the performance evaluation criterion to calculate the average deviation between the model’s detection results and the ground-truth gaze angles for both male and female subject groups. Ultimately, the average deviation for the male subjects was 4.51° in the horizontal direction and 4.43° in the vertical direction. For the female subjects, the average deviation was 5.71° in the horizontal direction and 5.07° in the vertical direction.

5.3. Comparison with Other State-of-the-Art Methods

In this subsection, we combined the test data from both male and female groups and set up a testing environment for state-of-the-art gaze-estimation models such as OpenFace2.2, ETH-XGaze, MPIIGaze (also known as GazeNet), MPIIFaceGaze (also referred to as Spatial Weights CNN), and L2CS on local computer devices. These state-of-the-art models are all trained on RGB images. Kendrick et al. [52] showed that RGB images perform best in facial landmark detection before gaze estimation, and the model converges the fastest. However, when predicting a wide range of landmarks, grayscale images are superior to RGB images. RGB images are sensitive to lighting and partial occlusion, which lack feature points [53]. In our experiments, we found that these state-of-the-art models have issues with infrared image data, as they are unable to detect the facial landmarks of the subjects and thus cannot output gaze angles. OpenFace2.2 performs better in detecting infrared gaze images of subjects than ETH-XGaze, MPIIFaceGaze, and MPIIGaze, as the latter three often fail to output gaze angles. L2CS significantly deviates from the ground-truth size in the angle values detected from some infrared gaze images. We calculated the accuracy of these five state-of-the-art models and the gaze-estimation model proposed in this study based on RMSE (see Figure 18). The gaze-estimation method proposed in this study has a root mean square error of 6.03° in the horizontal direction (yaw) and 4.83° in the vertical direction (pitch). The method for calculating the root mean square error (RMSE) between the gaze angle values detected by the model for the subjects and the ground-truth gaze angle values is as follows:
R M S E ( y a w ) = 1 Q i = 1 Q ( y a w i y a w t , i ) 2
R M S E ( p i t c h ) = 1 Q i = 1 Q ( p i t c h i p i t c h t , i ) 2
where R M S E ( y a w ) represents the root mean square error (RMSE) between the model-predicted gaze angle and the ground-truth gaze angle in the horizontal direction, while R M S E ( p i t c h ) denotes the RMSE for the vertical direction. Q signifies the number of experimental samples. y a w i refers to the model-predicted gaze angle of the subject in the horizontal direction during the i-th experiment, and y a w t , i is the ground-truth gaze angle of the subject in the horizontal direction. Similarly, p i t c h i represents the model-predicted gaze angle in the vertical direction for the i-th experiment, and p i t c h t , i indicates the ground-truth gaze angle of the subject in the vertical direction.

5.4. Experiments in Real Driving Scenarios for Automobile

To validate the performance of our gaze-estimation model in real-world application scenarios, we conducted driver gaze detection in a real driving environment within a full-scale car cabin. We organized 15 subjects to gaze at five target points during car driving, including the dashboard, back mirror, in-vehicle screen, driver mirror, and driver window (see Figure 19). Before recording these gaze data, we first calibrated the five gaze points and the TOF camera fixed points in the Toyota business SUV used, to measure the ground-truth gaze angles of the 15 drivers. The subjects performed free driving tasks in the experiment, with the TOF camera fixed in front of the car steering wheel and at a distance of 450 mm–600 mm from the subjects. Figure 20 shows the mean absolute error between the detection results of our gaze-estimation model for these target points and the ground-truth gaze angles. In normal car driving, the distance from the driver to the dashboard is about 600 mm. Moreover, adjusting the installation position of the TOF camera and maintaining a closer distance between the camera and the driver can appropriately improve the accuracy of driver gaze detection.

5.5. Discussion

Calibration of the subject’s eyeball parameters is a very important task for our model-based gaze-estimation method. In the experiment, setting individual subjects’ eyeball radius and the distance from the center of the eyeball to the pupil as medical constants of 12 mm and 13.1 mm, respectively, can also achieve a good gaze angle detection accuracy, but not all subjects’ eyeball parameters are close to these two constants. Therefore, it is imperative to calibrate individual-specific eyeball parameters. Since we approximate the position of the subject’s eyeball center separately in the horizontal and vertical directions, the eyeball structure parameters estimated in these two gaze directions are also different. Summarizing the overall computational process, the implementation steps of the gaze-estimation method proposed in this paper are as follows:
  • Utilizing the YOLOv8s neural network, a model for eye landmark detection is trained. This landmark detection model is then employed to perform inference on the gaze images of subjects, yielding 2D coordinates of seven landmark points on the image of a single eye of the subject, including points at the corners of the eye and the pupil;
  • Distortion correction is applied to the gaze images of the subjects, resulting in the acquisition of the 2D coordinates of the seven eye landmarks after distortion correction;
  • By integrating the depth images captured simultaneously by the TOF camera during the subject’s gaze, the 3D coordinates of the seven landmark points are computed within the coordinate system of the TOF camera;
  • An eyeball model in the horizontal and vertical directions is constructed using the left and right corner points of a single eye of the subject. The aspect ratio of the eye’s appearance during the subject’s gaze is calculated. Based on this aspect ratio value, the solution for the eyeball’s center point in the horizontal and vertical gaze directions of the subject is determined;
  • Based on the distance constraint between the center of the eyeball and the pupil point, combined with the 2D coordinates of the pupil on the gaze image of the subject, the 3D coordinates of the pupil in the horizontal and vertical directions within the TOF camera coordinate system are calculated;
  • Calibration of the subject-specific parameters of the eyeball structure is conducted to ascertain the optimal structural parameters of the eyeball and the range of the subject’s eye appearance aspect ratio variation with the vertical gaze angle. Based on these optimal parameters, the gaze angles of the subject in the horizontal and vertical directions, specifically yaw and pitch, are calculated.
In this study, we evaluate the performance of five state-of-the-art gaze-estimation models, namely ETH-XGaze, L2CS, MPIIFaceGaze (also referred to as Spatial Weights CNN), MPIIGaze (also known as GazeNet), and OpenFace 2.2, on our proposed IRGD benchmark dataset for subject gaze detection. These state-of-the-art models were assessed on a local PC, and their effectiveness in detecting the gaze direction of subjects is illustrated in Figure 21. Compared to these state-of-the-art models, the gaze-estimation method proposed in this paper demonstrates the highest accuracy in subject gaze detection. The superior performance is attributed to the proposed method’s robust detection capabilities on low-resolution gaze images and images where the subject is situated at a greater distance from the camera. In gaze images characterized by low resolution, large head pose angles, and a substantial distance between the subject and the camera (≥500 mm), existing state-of-the-art gaze-estimation methods often yield significant discrepancies between the estimated and ground-truth gaze angles due to the blurriness of the image and the limited availability of extractable information from the subject’s eye region. In some instances, the estimated gaze direction may even be the opposite of the subject’s ground-truth line of sight. The gaze-estimation method proposed in this paper overcomes these limitations by utilizing only two eye corner points of a single eye to model the eyeball and employing easily observable pupil points to calculate the gaze angle. Remarkably, even on low-resolution images of 300,000 pixels, our method accurately computes the subject’s gaze angle, maintaining its precision at a distance of 600 mm between the subject and the camera. Our proposed method computes the subject’s gaze angle within the 3D coordinate system of a TOF camera, which significantly mitigates the influence of the head pose on gaze angle estimation. In contrast, existing state-of-the-art gaze-estimation approaches predict the subject’s gaze angle based on 2D images and are highly susceptible to perturbations caused by the subject’s head pose.
In this study, subjects were divided into male and female groups to conduct cross-gender research, aiming to investigate the influence of gender factors on gaze estimation. We found that the aspect ratio of female subjects’ eyes was slightly larger than that of males, but the average gaze angle deviation calculated was close to that of males. The driver gaze detection experiment was conducted within the normal driving distance of 450 mm–600 mm from the dashboard, and the final results confirmed that our proposed gaze-estimation method has stable detection capabilities at different distances. In all experimental results, the maximum gaze angle deviation of the subjects did not exceed 9°, which means that within a driving distance of 600 mm, the maximum detection deviation of our gaze-estimation model in the horizontal and vertical directions is 95 mm. In our Toyota SUV, the distances between various gaze target points inside the car are much greater than this deviation value. Therefore, the detection deviation of the gaze-estimation model proposed in this study is sufficient to distinguish and recognize the driver’s gaze fixation points within the car.

6. Conclusions

This paper presents a model-based 3D gaze estimation method using a TOF camera. Unlike most gaze-estimation methods that use RGB images, our gaze-estimation work is completed on low-resolution infrared images, which is extremely advantageous for applications with varying lighting conditions and complex environments, particularly suitable for driver gaze detection in night-time scenarios.
Our main contributions include proposing an infrared gaze dataset (IRGD) for a wide range of free head and eyeball movements and training an eye landmark detection model for the subject’s eye area on the IRGD dataset using YOLOv8. In addition, this study proposes a method to fit a 3D eyeball model using only two corner points of the subject’s single-eye image, enabling our gaze-estimation method to work stably in images where the subject is far from the camera or in low resolution. To address the drawback of the TOF camera not being able to capture the depth information of the subject’s pupil at certain special gaze angles, we solve for the 3D coordinates of the pupil point using the constraint relationship between the subject’s pupil and the center of the eyeball. This paper analyzes the subject’s gaze angle in two directions, horizontal and vertical, and the subject only needs to be calibrated once to calculate the centers of the eyeball and pupil under the horizontal and vertical gaze directions.
Through experimental comparison with other state-of-the-art gaze-estimation models, we have confirmed the effectiveness and advancement of our proposed gaze-estimation method. By comparing with the ground truth gaze angle, the maximum detection deviation of the gaze-estimation method proposed in this paper is less than 9° for the subjects. Using RMSE as the performance evaluation standard, our gaze-estimation method has detection deviations of 6.03° and 4.83° in the horizontal and vertical directions, respectively. We employed the proposed gaze-estimation method to detect gaze angles in subject images with a resolution of 640 × 480. The model proposed in this paper achieved a detection speed of 7.8 frames per second (FPS) when implemented on a local PC equipped with an Intel Core i7-11800H @ 2.30GHz CPU. We also evaluated the accuracy of our gaze-estimation method for driver gaze detection in real car driving scenarios. In future research, we plan to improve the robustness of the gaze-estimation model to extreme head poses and self-occlusion. We will explore the possibility of dynamically selecting the subject’s left or right eye in conjunction with the subject’s head pose to enhance the accuracy of gaze estimation.

Author Contributions

Conceptualization, K.S. and Y.L.; methodology, K.S. and Y.L.; data curation and writing—original draft, Y.L.; funding acquisition, validtion and project administration, Z.G.; visualization and investigation, J.G.; software and validation, Y.W.; writing—review and editing, K.S. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by Liaoning Provincial Department of Education Project (LJKQZ2021082), China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

All the participating authors have signed and given the informed consent statements.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy protection for data recorders.

Acknowledgments

This work was guided by insightful suggestions on research directions from our friend Yan Mu, and greatly assisted by our friends Zhandong Ge, Jinghui Tao, Xiaojun Wang, Shouda Wang, and Bangxuan Tang in data collection. We express our profound gratitude to them.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kar, A.; Corcoran, P. A review and analysis of eye-gaze estimation systems, algorithms and performance evaluation methods in consumer platforms. IEEE Access 2017, 5, 16495–16519. [Google Scholar] [CrossRef]
  2. Chandra, S.; Sharma, G.; Malhotra, S.; Jha, D.; Mittal, A.P. Eye tracking based human computer interaction: Applications and their uses. In Proceedings of the 2015 International Conference on Man and Machine Interfacing (MAMI), Bhubaneswar, India, 17–19 December 2015; pp. 1–5. [Google Scholar] [CrossRef]
  3. Kapp, S.; Barz, M.; Mukhametov, S.; Sonntag, D.; Kuhn, J. ARETT: Augmented reality eye tracking toolkit for head mounted displays. Sensors 2021, 21, 2234. [Google Scholar] [CrossRef] [PubMed]
  4. Valtakari, N.V.; Hessels, R.S.; Niehorster, D.C.; Viktorsson, C.; Nyström, P.; Falck-Ytter, T.; Kemner, C.; Hooge, I.T.C. A field test of computer-vision-based gaze estimation in psychology. Behav. Res. Methods 2023, 1–16. [Google Scholar] [CrossRef] [PubMed]
  5. Nandini, M.; Singh, J. A review of various state of art eye gaze estimation techniques. Adv. Comput. Intell. Commun. Technol. Proc. CICT 2021, 2019, 501–510. [Google Scholar] [CrossRef]
  6. Wen, Q.; Bradley, D.; Beeler, T.; Park, S.; Hilliges, O.; Yong, J.; Xu, F. Accurate Real-time 3D Gaze Tracking Using a Lightweight Eyeball Calibration. Comput. Graph. Forum 2020, 39, 475–485. [Google Scholar] [CrossRef]
  7. Sharma, P.K.; Chakraborty, P. A Review of Driver Gaze Estimation and Application in Gaze Behavior Understanding. arXiv 2023, arXiv:2307.01470. [Google Scholar]
  8. Mora, K.A.F.; Odobez, J.-M. Gaze estimation from multimodal Kinect data. In Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPR Workshops), Providence, RI, USA, 16–21 June 2012; pp. 25–30. [Google Scholar]
  9. William, W.E. Gaze Estimation with Graphics. Doctoral Dissertation, University of Cambridge, Cambridge, UK, 2017. [Google Scholar] [CrossRef]
  10. Chen, J.; Ji, Q. 3D gaze estimation with a single camera without IR illumination. In Proceedings of the 2008 19th International Conference on Pattern Recognition (ICPR), Tampa, FL, USA, 8–11 December 2008; pp. 1–4. [Google Scholar] [CrossRef]
  11. Li, J.; Shigang, L. Eye-model-based gaze estimation by RGB-D camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  12. Guestrin, E.; Eizenman, M. General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Trans. Biomed. Eng. 2006, 53, 1124–1133. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, X.; Sugano, Y.; Fritz, M.; Bulling, A. Appearance-based gaze estimation in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  14. Zhang, X.; Sugano, Y.; Fritz, M.; Bulling, A. MPIIGaze: Real-world dataset and deep appearance-based gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 41, 162–175. [Google Scholar] [CrossRef] [PubMed]
  15. Zhang, X.; Sugano, Y.; Fritz, M.; Bulling, A. It’s written all over your face: Full-face appearance-based gaze estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017. arXiv:1611.08860v4. [Google Scholar]
  16. Seonwook, P.; Spurr, A.; Hilliges, O. Deep pictorial gaze estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. arXiv:1807.10002v1. [Google Scholar]
  17. Bace, M.; Becker, V.; Wang, C.; Bulling, A. Combining gaze estimation and optical flow for pursuits interaction. In Proceedings of the ETRA ‘20: 2020 Symposium on Eye Tracking Research and Applications, Stuttgart, Germany, 2–5 June 2020. [Google Scholar] [CrossRef]
  18. Park, S.; De Mello, S.; Molchanov, P.; Iqbal, U.; Hilliges, O.; Kautz, J. Few-shot adaptive gaze estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  19. Abdelrahman, A.A.; Hempel, T.; Khalifa, A.; Al-Hamadi, A.; Dinges, L. L2CS-Net: Fine-grained gaze estimation in unconstrained environments. In Proceedings of the 2023 8th International Conference on Frontiers of Signal Processing (ICFSP), Corfu, Greece, 23–25 October 2023; pp. 98–102. [Google Scholar]
  20. Murthy, L.R.D.; Mukhopadhyay, A.; Biswas, P. Distraction detection in automotive environment using appearance-based gaze estimation. In Proceedings of the IUI ‘22: 27th International Conference on Intelligent User Interfaces, Helsinki, Finland, 22–25 March 2022. [Google Scholar] [CrossRef]
  21. Vasli, B.; Martin, S.; Trivedi, M.M. On driver gaze estimation: Explorations and fusion of geometric and data driven approaches. In Proceedings of the 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; pp. 655–660. [Google Scholar] [CrossRef]
  22. Vora, S.; Rangesh, A.; Trivedi, M.M. On generalizing driver gaze zone estimation using convolutional neural networks. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 849–854. [Google Scholar] [CrossRef]
  23. Shah, S.M.; Sun, Z.; Zaman, K.; Hussain, A.; Shoaib, M.; Pei, L. A driver gaze estimation method based on deep learning. Sensors 2022, 22, 3959. [Google Scholar] [CrossRef] [PubMed]
  24. Vicente, F.; Huang, Z.; Xiong, X.; la Torre, D.F.; Zhang, W.; Levi, D. Driver gaze tracking and eyes off the road detection system. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2014–2027. [Google Scholar] [CrossRef]
  25. Wang, K.; Ji, Q. 3D gaze estimation without explicit personal calibration. Pattern Recognit. 2018, 79, 216–227. [Google Scholar] [CrossRef]
  26. Sesma, L.; Villanueva, A.; Cabeza, R. Evaluation of pupil center-eye corner vector for gaze estimation using a web cam. In Proceedings of the ETRA ‘12: Eye Tracking Research and Applications, Stuttgart, Germany, 28 June 2012. [Google Scholar] [CrossRef]
  27. Park, S.; Aksan, E.; Zhang, X.; Hilliges, O. Towards end-to-end video-based eye-tracking. Computer Vision. In Proceedings of the ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XII 16. Springer International Publishing: Berlin/Heidelberg, Germany, 2020. [Google Scholar] [CrossRef]
  28. Park, S.; Zhang, X.; Bulling, A.; Hilliges, O. Learning to find eye region landmarks for remote gaze estimation in unconstrained settings. In Proceedings of the ETRA ‘18: 2018 Symposium on Eye Tracking Research and Applications, Warsaw, Poland, 14–17 June 2018; p. 21. [Google Scholar] [CrossRef]
  29. Odobez, J.-M.; Mora, K.F. Geometric Generative Gaze Estimation Model from RGB-D Sensors. 2022. Available online: https://www.idiap.ch/webarchives/sites/www.idiap.ch/project/g3e/index.html (accessed on 10 September 2022).
  30. Tamura, K.; Choi, R.; Aoki, Y. Unconstrained and Calibration-free gaze estimation in a room-scale area using a monocular camera. IEEE Access 2017, 6, 10896–10908. [Google Scholar] [CrossRef]
  31. Cristina, S.; Camilleri, K.P. Model-based head pose-free gaze estimation for assistive communication. Comput. Vis. Image Underst. 2016, 149, 157–170. [Google Scholar] [CrossRef]
  32. Zhang, X.; Park, S.; Beeler, T.; Bradley, D.; Tang, S.; Hilliges, O. Eth-xgaze: A large scale dataset for gaze estimation under extreme head pose and gaze variation. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part V 16. Springer International Publishing: Berlin/Heidelberg, Germany, 2020. [Google Scholar] [CrossRef]
  33. Yusuke, S.; Matsushita, Y.; Sato, Y. Learning-by-synthesis for appearance-based 3d gaze estimation. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, San Juan, PR, USA, 17–19 June 2014. [Google Scholar]
  34. Baltrusaitis, T.; Robinson, P.; Morency, L.-P. OpenFace: An open source facial behavior analysis toolkit. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016. [Google Scholar] [CrossRef]
  35. Jocher, G. YOLOv8 Pose Estimation Model. 2023. Available online: https://docs.ultralytics.com/tasks/pose/ (accessed on 3 July 2023).
  36. Use Non-Linear Least Squares to Fit a Function f to Data. 2022. Available online: https://docs.scipy.org/doc/scipy/reference/generated/scipy.optimize.curve_fit.html (accessed on 11 October 2022).
  37. Siegfried, R.; Odobez, J.-M. Robust unsupervised gaze calibration using conversation and manipulation attention priors. ACM Trans. Multimedia Comput. Commun. Appl. 2022, 18, 1–27. [Google Scholar] [CrossRef]
  38. Siegfried, R.; Aminian, B.; Odobez, J.-M. ManiGaze: A Dataset for Evaluating Remote Gaze Estimator in Object Manipulation Situations. In Proceedings of the ETRA ‘20: 2020 Symposium on Eye Tracking Research and Applications, Stuttgart, Germany, 2–5 June 2020. [Google Scholar] [CrossRef]
  39. Mora, F.; Alberto, K.; Monay, F.; Odobez, J.-M. Eyediap: A database for the development and evaluation of gaze estimation algorithms from rgb and rgb-d cameras. In Proceedings of the Symposium on Eye Tracking Research and Applications, Safety Harbor, FL, USA, 26–28 March 2014. [Google Scholar] [CrossRef]
  40. Kellnhofer, P.; Recasens, A.; Stent, S.; Matusik, W.; Torralba, A. Gaze360: Physically unconstrained gaze estimation in the wild. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  41. Smith, B.A.; Yin, Q.; Feiner, S.K.; Nayar, S.K. Gaze locking: Passive eye contact detection for human-object interaction. In Proceedings of the UIST’13: The 26th Annual ACM Symposium on User Interface Software and Technology, St. Andrews Scotland, UK, 8–11 October 2013; pp. 271–280. [Google Scholar] [CrossRef]
  42. Krafka, K.; Khosla, A.; Kellnhofer, P.; Kannan, H.; Bhandarkar, S.; Matusik, W.; Torralba, A. Eye tracking for everyone. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  43. Wood, E.; Baltruaitis, T.; Zhang, X.; Sugano, Y.; Robinson, P.; Bulling, A. Rendering of eyes for eye-shape registration and gaze estimation. In Proceedings of the IEEE international Conference on Computer Vision, Santiago, Chile, 7–13 December 2015. [Google Scholar]
  44. Wood, E.; Baltrušaitis, T.; Morency, L.-P.; Robinson, P.; Bulling, A. Learning an appearance-based gaze estimator from one million synthesised images. In Proceedings of the ETRA ‘16: 2016 Symposium on Eye Tracking Research and Applications, Charleston, CA, USA, 14–17 March 2016; pp. 131–138. [Google Scholar] [CrossRef]
  45. Wang, X.; Zhang, J.; Zhang, H.; Zhao, S.; Liu, H. Vision-based gaze estimation: A review. IEEE Trans. Cogn. Dev. Syst. 2021, 14, 316–332. [Google Scholar] [CrossRef]
  46. Mallick, S. Head Pose Estimation using OpenCV and Dlib. 2016. Available online: https://learnopencv.com/head-pose-estimation-using-opencv-and-dlib/ (accessed on 15 October 2023).
  47. Case on OpenCV Camera Calibration and Human Head Pose Estimation, CSDN Blog. 2021. Available online: https://blog.csdn.net/zb1165048017/article/details/115553641 (accessed on 9 October 2023).
  48. Hempel, T.; Abdelrahman, A.A.; Al-Hamadi, A. 6d Rotation representation for unconstrained head pose estimation. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 2496–2500. [Google Scholar] [CrossRef]
  49. Sadekar, K. Understanding Lens Distortion. 2020. Available online: https://learnopencv.com/understanding-lens-distortion/ (accessed on 8 October 2023).
  50. Camera Calibration Using OpenCV-Python Toolkit. 2023. Available online: https://docs.opencv.org/4.x/dc/dbb/tutorial_py_calibration.html (accessed on 11 October 2023).
  51. Baltrusaitis, T. Issue 969 of OpenFace Github. 2021. Available online: https://github.com/TadasBaltrusaitis/OpenFace/issues/969 (accessed on 29 September 2023).
  52. Kendrick, C.; Tan, K.; Walker, K.; Yap, M.H. Towards real-time facial landmark detection in depth data using auxiliary information. Symmetry 2018, 10, 230. [Google Scholar] [CrossRef]
  53. Venturelli, M.; Borghi, G.; Vezzani, R.; Cucchiara, R. From Depth Data to Head Pose Estimation: A Siamese Approach. In Proceedings of the International Conference on Computer Vision Theory and Applications, Setúbal, Portugal, 27 February–1 March 2017; pp. 194–201. [Google Scholar]
Figure 2. The partial effectiveness of data augmentation.
Figure 2. The partial effectiveness of data augmentation.
Sensors 24 01070 g002
Figure 3. The eye region and landmark detection model trained on the IRGD dataset using YOLOv8 shows the detection effect on the subject’s gaze image (a). The landmark detection model outputs 7 target points for a single-eye image of the subject (b): 1—Left eye corner point; 2—First upper eyelid point; 3—Second upper eyelid point; 4—Right eye corner point; 5—First lower eyelid point; 6—Second lower eyelid point; 7—Pupil point.
Figure 3. The eye region and landmark detection model trained on the IRGD dataset using YOLOv8 shows the detection effect on the subject’s gaze image (a). The landmark detection model outputs 7 target points for a single-eye image of the subject (b): 1—Left eye corner point; 2—First upper eyelid point; 3—Second upper eyelid point; 4—Right eye corner point; 5—First lower eyelid point; 6—Second lower eyelid point; 7—Pupil point.
Sensors 24 01070 g003
Figure 4. The subject maintained a head pose angle of 0° in both the horizontal and vertical directions and performed a series of coherent lizard movements. The green arrow indicates the ground-truth gaze direction, while the red arrow represents the final gaze direction obtained using the eyeball center calculation method proposed in [30]. As the subject’s gaze angle gradually increased, the deviation between the gaze angle calculated by this eyeball center localization method and the ground-truth gaze angle began to increase. Table 1 shows the results of our calculations.
Figure 4. The subject maintained a head pose angle of 0° in both the horizontal and vertical directions and performed a series of coherent lizard movements. The green arrow indicates the ground-truth gaze direction, while the red arrow represents the final gaze direction obtained using the eyeball center calculation method proposed in [30]. As the subject’s gaze angle gradually increased, the deviation between the gaze angle calculated by this eyeball center localization method and the ground-truth gaze angle began to increase. Table 1 shows the results of our calculations.
Sensors 24 01070 g004
Figure 5. Eight marked points are manually annotated on the image of the subject’s single eye. These points are randomly distributed on the sclera of the eye, not the cornea. We use these eight 3D coordinate points to fit the eyeball model and solve for the 3D coordinates of the eyeball center and the radius of the eyeball.
Figure 5. Eight marked points are manually annotated on the image of the subject’s single eye. These points are randomly distributed on the sclera of the eye, not the cornea. We use these eight 3D coordinate points to fit the eyeball model and solve for the 3D coordinates of the eyeball center and the radius of the eyeball.
Sensors 24 01070 g005
Figure 6. Eye detail images taken by the TOF camera at a distance of 200 mm–500 mm from the subject. The experiment is divided into two scenarios: the subject not wearing myopia glasses (top) and wearing glasses (bottom). The occlusion of glasses reduces some of the clarity and contrast of the subject’s eyes, but it is much less than the impact of a longer distance. When the distance between the subject and the TOF camera exceeds 300 mm, the only observable details in the eye area are the corners of the eyes and the pupil points.
Figure 6. Eye detail images taken by the TOF camera at a distance of 200 mm–500 mm from the subject. The experiment is divided into two scenarios: the subject not wearing myopia glasses (top) and wearing glasses (bottom). The occlusion of glasses reduces some of the clarity and contrast of the subject’s eyes, but it is much less than the impact of a longer distance. When the distance between the subject and the TOF camera exceeds 300 mm, the only observable details in the eye area are the corners of the eyes and the pupil points.
Sensors 24 01070 g006
Figure 7. Creating a standard plane with multiple gaze points using a level’s laser line (a) and fixing a TOF camera on the plane (b).
Figure 7. Creating a standard plane with multiple gaze points using a level’s laser line (a) and fixing a TOF camera on the plane (b).
Sensors 24 01070 g007
Figure 8. Sample pictures of the IRGD dataset proposed in this paper. We recorded gaze data at five different distances from the participant to the TOF camera, ranging from 200 mm to 600 mm. The TOF camera simultaneously collected IR images and depth images of the participant gazing at the gaze points on the standard plane. All participants performed natural eye movements and coherent head movements.
Figure 8. Sample pictures of the IRGD dataset proposed in this paper. We recorded gaze data at five different distances from the participant to the TOF camera, ranging from 200 mm to 600 mm. The TOF camera simultaneously collected IR images and depth images of the participant gazing at the gaze points on the standard plane. All participants performed natural eye movements and coherent head movements.
Sensors 24 01070 g008
Figure 9. The absolute values of the average head pose angles of the participants at 35 gaze points in the IRGD dataset. The maximum absolute angle of the participants’ head pose in the horizontal direction (yaw) is approximately 50°, while in the vertical direction (pitch), it is approximately 30°.
Figure 9. The absolute values of the average head pose angles of the participants at 35 gaze points in the IRGD dataset. The maximum absolute angle of the participants’ head pose in the horizontal direction (yaw) is approximately 50°, while in the vertical direction (pitch), it is approximately 30°.
Sensors 24 01070 g009
Figure 10. Independent modeling and solution of eyeball center coordinates in horizontal (a) and vertical (b) gaze directions of subjects.
Figure 10. Independent modeling and solution of eyeball center coordinates in horizontal (a) and vertical (b) gaze directions of subjects.
Sensors 24 01070 g010
Figure 11. Variation trends of the aspect ratio of eye appearance with vertical gaze angle in male (a) and female (b) participants. In male participants, the aspect ratio of eye appearance is less than 0.3 when the eyeball is looking down, while in female participants, the aspect ratio of eye appearance is less than 0.4 when the eyeball is looking down.
Figure 11. Variation trends of the aspect ratio of eye appearance with vertical gaze angle in male (a) and female (b) participants. In male participants, the aspect ratio of eye appearance is less than 0.3 when the eyeball is looking down, while in female participants, the aspect ratio of eye appearance is less than 0.4 when the eyeball is looking down.
Sensors 24 01070 g011
Figure 12. The drawback of the inability to extract pupil point depth values from the depth image of the TOF camera. For gaze images at certain special angles, the pupil point can be observed on its IR image (a), but due to the absorption of infrared light by the pupil, a ‘black hole’ appears at the position of the pupil point on the corresponding depth image (b) of the IR image.
Figure 12. The drawback of the inability to extract pupil point depth values from the depth image of the TOF camera. For gaze images at certain special angles, the pupil point can be observed on its IR image (a), but due to the absorption of infrared light by the pupil, a ‘black hole’ appears at the position of the pupil point on the corresponding depth image (b) of the IR image.
Sensors 24 01070 g012
Figure 13. Schematic diagram of the calibration process for individual-specific eyeball parameters of the subject.
Figure 13. Schematic diagram of the calibration process for individual-specific eyeball parameters of the subject.
Sensors 24 01070 g013
Figure 14. Calibration results of eyeball parameters for three subjects. We obtained the optimal eyeball structure parameters ( R 1 , d 1 ) and ( R 2 , d 2 ) for 3 subjects through 10 calibrations, each involving gazing at 20 gaze points. At the same time, we calculated the mean absolute deviation between the gaze angles in the horizontal direction (blue) and vertical direction (orange) computed from this set of parameters and the ground-truth angles.
Figure 14. Calibration results of eyeball parameters for three subjects. We obtained the optimal eyeball structure parameters ( R 1 , d 1 ) and ( R 2 , d 2 ) for 3 subjects through 10 calibrations, each involving gazing at 20 gaze points. At the same time, we calculated the mean absolute deviation between the gaze angles in the horizontal direction (blue) and vertical direction (orange) computed from this set of parameters and the ground-truth angles.
Sensors 24 01070 g014
Figure 15. Experiment results on calculating the average pupil depth information and corresponding ground-truth values in horizontal and vertical gaze directions for the male group (a) and female group (b).
Figure 15. Experiment results on calculating the average pupil depth information and corresponding ground-truth values in horizontal and vertical gaze directions for the male group (a) and female group (b).
Sensors 24 01070 g015
Figure 16. Results of the subject’s gaze detection. Column (a) presents the original gaze images of the subject, column (b) shows the results of eye landmark detection based on YOLOv8, and column (c) visualizes the subject’s gaze direction. The green arrow indicates the gaze direction detected by our model.
Figure 16. Results of the subject’s gaze detection. Column (a) presents the original gaze images of the subject, column (b) shows the results of eye landmark detection based on YOLOv8, and column (c) visualizes the subject’s gaze direction. The green arrow indicates the gaze direction detected by our model.
Sensors 24 01070 g016
Figure 17. Gaze angle detection results of male and female subject groups using the gaze-estimation method proposed in this study. Specifically, (a) represents the horizontal gaze results of the male group, (b) shows the vertical gaze results of the male group. (c) illustrates the horizontal gaze results of the female group, and (d) presents the vertical gaze results of the female subjects.
Figure 17. Gaze angle detection results of male and female subject groups using the gaze-estimation method proposed in this study. Specifically, (a) represents the horizontal gaze results of the male group, (b) shows the vertical gaze results of the male group. (c) illustrates the horizontal gaze results of the female group, and (d) presents the vertical gaze results of the female subjects.
Sensors 24 01070 g017aSensors 24 01070 g017b
Figure 18. Comparative accuracy results of our proposed gaze-estimation model with other state-of-the-art models in infrared gaze test images.
Figure 18. Comparative accuracy results of our proposed gaze-estimation model with other state-of-the-art models in infrared gaze test images.
Sensors 24 01070 g018
Figure 19. Detection results of driver’s partial gaze points in the interior of a Toyota business SUV. Green arrows indicate the driver’s gaze direction detected by our gaze-estimation model.
Figure 19. Detection results of driver’s partial gaze points in the interior of a Toyota business SUV. Green arrows indicate the driver’s gaze direction detected by our gaze-estimation model.
Sensors 24 01070 g019
Figure 20. Mean absolute error between the detected driver’s gaze angles and ground-truth angles at various gaze points inside the car.
Figure 20. Mean absolute error between the detected driver’s gaze angles and ground-truth angles at various gaze points inside the car.
Sensors 24 01070 g020
Figure 21. Detection effect of existing state-of-the-art gaze-estimation methods on the IRGD dataset proposed in this study, with arrows and lines indicating the predicted gaze direction of the subject by each model.
Figure 21. Detection effect of existing state-of-the-art gaze-estimation methods on the IRGD dataset proposed in this study, with arrows and lines indicating the predicted gaze direction of the subject by each model.
Sensors 24 01070 g021
Table 1. The angle calculation results of gaze estimation based on the 3D eyeball center localization method mentioned in [30].
Table 1. The angle calculation results of gaze estimation based on the 3D eyeball center localization method mentioned in [30].
NumberGround-Truth Yaw Angle [Degree]Distance from Pupil to TOF Camera [mm]Calculated Yaw Angle Value [Degree]Absolute Deviation [Degree]
141.6495.2930.710.9
227.01502.3533.196.18
38.39495.299.831.44
4−12.3498.82−12.930.63
5−30.39498.82−24.136.26
6−43.54502.35−52.939.39
Table 2. The radius value of the 3D eyeball model fitted using the method from [36].
Table 2. The radius value of the 3D eyeball model fitted using the method from [36].
The Distance from the Subject’s Head to the TOF Camera [mm]The Depth Values of Each Marked Point on the Depth Image [mm]The Radius Value of the Eyeball [mm]
12345678
≈200184.3184.3184.6188.2183.7188.2184.3185.611.8
≈300274.5274.5273.2274.4278.4279.1278.4274.512.5
≈400360.8360.8361.2362.2356.9361.2356.9360.811.5
Table 3. Comparison and summary of current open-source datasets and our proposed IRGD dataset.
Table 3. Comparison and summary of current open-source datasets and our proposed IRGD dataset.
DatasetsWith FaceTotalGaze PitchGaze YawImage Types
MPIIGazeNo213 K images±20°±20°RGB
MPIIFaceGazeYes213 K images±20°±20°RGB
EYEDIAPYes94 Videos±30°±40°RGB and RGBD
UnityEyesNoUser definedUser definedUser definedRGB
ETH-XGazeYesOver 1 M images±70°±120°RGB
ManiGazeYes11 K images (from videos)±40°±50°RGB and RGBD
Gaze360Yes172 K images−50°±140°RGB
ColumbiaYes5 K±10°±15°RGB
EVEYesOver 12 M images (from videos)±50°±60°RGB
GazeCaptureYesOver 2 M images±20°±20°RGB
UT Multi-viewNo64 K±36°±50°RGB
SynthesEyesNo11 K±25°±35°RGB
IRGDYes150 K±50°±50°IR and Depth
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shen, K.; Li, Y.; Guo, Z.; Gao, J.; Wu, Y. Model-Based 3D Gaze Estimation Using a TOF Camera. Sensors 2024, 24, 1070. https://doi.org/10.3390/s24041070

AMA Style

Shen K, Li Y, Guo Z, Gao J, Wu Y. Model-Based 3D Gaze Estimation Using a TOF Camera. Sensors. 2024; 24(4):1070. https://doi.org/10.3390/s24041070

Chicago/Turabian Style

Shen, Kuanxin, Yingshun Li, Zhannan Guo, Jintao Gao, and Yingjian Wu. 2024. "Model-Based 3D Gaze Estimation Using a TOF Camera" Sensors 24, no. 4: 1070. https://doi.org/10.3390/s24041070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop