Skip to main content

REVIEW article

Front. Psychol., 20 March 2024
Sec. Perception Science
This article is part of the Research Topic Role of Eye Movements in Vision, Attention, Decision-making, and Disease View all 6 articles

A review on personal calibration issues for video-oculographic-based gaze tracking

\r\nJiahui Liu,,Jiahui Liu1,2,3Jiannan Chi,,
Jiannan Chi1,2,3*Zuoyun YangZuoyun Yang1
  • 1School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing, China
  • 2Beijing Engineering Research Center of Industrial Spectrum Imaging, University of Science and Technology Beijing, Beijing, China
  • 3Shunde Innovation School, University of Science and Technology Beijing, Foshan, China

Personal calibration is a process of obtaining personal gaze-related information by focusing on some calibration benchmarks when the user initially uses a gaze tracking system. It not only provides conditions for gaze estimation, but also improves gaze tracking performance. Existing eye-tracking products often require users to conduct explicit personal calibration first, thereby tracking and interacting based on their gaze. This calibration mode has certain limitations, and there is still a significant gap between theoretical personal calibration methods and their practicality. Therefore, this paper reviews the issues of personal calibration for video-oculographic-based gaze tracking. The personal calibration information in typical gaze tracking methods is first summarized, and then some main settings in existing personal calibration processes are analyzed. Several personal calibration modes are discussed and compared subsequently. The performance of typical personal calibration methods for 2D and 3D gaze tracking is quantitatively compared through simulation experiments, highlighting the characteristics of different personal calibration settings. On this basis, we discuss several key issues in designing personal calibration. To the best of our knowledge, this is the first review on personal calibration issues for video-oculographic-based gaze tracking. It aims to provide a comprehensive overview of the research status of personal calibration, explore its main directions for further studies, and provide guidance for seeking personal calibration modes that conform to natural human-computer interaction and promoting the widespread application of eye-movement interaction.

1 Introduction

With the rapid development of artificial intelligence, eye-movement interaction is increasingly favored by people due to its efficient and real-time. Eye-movement interaction utilizes gaze information for interaction and its key technology is gaze tracking, which uses visual information from face or eye images to analyze the user's gaze direction or point-of-regard (POR). At present, gaze tracking has been used in human-computer interaction, medical diagnosis, virtual reality, and intelligent transportation (Drakopoulos et al., 2021; Hu et al., 2022; Li et al., 2022).

Video-oculographic (VOG)-based gaze tracking systems generally conduct a personal calibration process to obtain some user-specific parameters or images, as shown in Figure 1. 2D mapping-based methods take the eye invariant features (e.g., eye corner point or glint) as the benchmark, and construct a mapping model between eye variation features (e.g., pupil center or iris center) and 2D POR for gaze estimation. To determine the mapping model for a specific user, it is necessary to calibrate some of his/her specific parameters through personal calibration. At present, most personal calibration methods require users to stare at multiple calibration points on the screen, so as to obtain sufficient calibration data for mapping model construction (Cheng et al., 2017; Mestre et al., 2018; Hu et al., 2019; Uhm et al., 2020). Hu et al. (2019) asked users to focus on nine explicit calibration points while keeping their heads fixed during personal calibration, to estimate the POR using the eye-movement vector between iris center and corner points. Uhm et al. (2020) used four calibration points to calibrate the projection transformation matrix between eye image and screen.

Figure 1
www.frontiersin.org

Figure 1. Personal calibration processes in different methods.

3D geometry-based methods utilize visual features (such as pupil, iris, and glints) to calibrate eye invariant parameters (such as corneal radius and kappa angle) based on eyeball structure and geometric imaging model. Then, eye-variation parameters (such as eyeball center, corneal center, pupil center, and iris center) are estimated to reconstruct the optical axis (OA) of the eyeball, thereby determining the visual axis (VA) using the OA and the kappa angle. Due to a fixed deviation between the OA and the VA, called the kappa angle, at least a single-point calibration is generally required to calculate it (Lai et al., 2015). To estimate the VA in a simple system, some parameters such as corneal radius, distance between corneal center and pupil center also need to be calibrated, in addition to the kappa angle (Cristina and Camilleri, 2016; Zhou et al., 2017). Cristina and Camilleri (2016) detected a frontal eye and head pose during personal calibration, and then estimated the 3D gaze from a single camera based on a cylindrical head and spherical eyeball model.

Appearance-based methods take face or eye images as input, and learn the mapping of face or eye images to gaze information by using a large number of training samples with ground-truth labels, thereby predicting the gaze information for new images using the trained model. Most methods are calibration-free, but some methods use a few calibration samples to optimize the model and reduce the impact of individual differences (Krafka et al., 2016; Gu et al., 2021; Liu G. et al., 2021; Wang et al., 2023). Liu G. et al. (2021) trained a differential convolutional neural network to predict the gaze difference between two input eye images of a same subject, and the gaze direction of a new eye sample was predicted by inferring the gaze differences of a set of subject-specific calibration images. Krafka et al. (2016) used the data from 13 fixed locations for calibration to train the SVR to predict the gaze location. The performance was improved significantly.

Overall, the roles of personal calibration are mainly reflected in three aspects: (1) calibrate some user-specific parameters in the gaze estimation model to meet the condition for gaze estimation; (2) calibrate the user-specific parameters using multiple calibration benchmarks to enhance its robustness; (3) introduce some user-specific features to improve the gaze accuracy. At present, eye tracking products on the market also require personal calibration at the beginning of use, by looking at several calibration points or images. However, this explicit user calibration limits the convenience of product use and provides a poor user experience. To help improve the personal calibration process for gaze tracking systems and promote the implementation of instant use that an ideal gaze tracking system should have, we specifically explore the issues of personal calibration for gaze tracking. This paper provides a comprehensive overview of calibration information, calibration settings, and calibration modes involved in personal calibration, clarifies the characteristics of typical personal calibration methods under different calibration settings through simulation experiments, and discusses some key issues for designing personal calibration. In view of the research status, some main directions for future studies on personal calibration are provided. To the best of our knowledge, there is no literature specifically exploring the issues of personal calibration for gaze tracking. Most literature only provides a short description of its own calibration process. The main contributions of this work are as follows:

(1) This is the first review on the issues of personal calibration for gaze tracking. Some calibration information, calibration settings, and calibration modes in existing methods are summarized, reflecting the characteristics of current personal calibration.

(2) Simulation experiments of several typical personal calibration methods are conducted to clarify the key issues of personal calibration, which are helpful in determining the calibration information, setting the calibration process, and selecting the calibration mode.

(3) The current limitations and further research directions of personal calibration are discussed, providing guidance for researchers to seek more convenient and natural personal calibration methods, and promoting the upgrading of personal calibration in eye-movement interaction applications.

This paper is organized as shown in Figure 2. Section 2 provides an overview of the user-specific information obtained through personal calibration in different methods. Section 3 analyzes some of the main settings in the existing personal calibration process. We summarize several existing personal calibration modes for gaze tracking and compared their characteristics in Section 4. Section 5 compares the performance of typical personal calibration methods under different settings through simulation experiments, reflecting the characteristics of different personal calibration settings intuitively. A discussion of the key issues for designing personal calibration is given in Section 6. Finally, this paper is concluded and the development trends of personal calibration are prospected in the conclusion.

Figure 2
www.frontiersin.org

Figure 2. Framework of this paper.

2 Personal calibration information

The primary role of personal calibration is to provide useful personal information for gaze estimation. According to the classification of gaze estimation methods, this section discusses the commonly calibrated personal information in different methods, as Figure 3 shows.

Figure 3
www.frontiersin.org

Figure 3. Personal calibration information in different methods.

2.1 Calibration information in 2D mapping-based methods

2D mapping-based methods mainly include pupil/iris-corner technique (PCT/ICT)-based methods, pupil/iris-corneal reflection technique (PCRT/ICRT)-based methods, cross-ratio (CR)-based methods, and homography normalization (HN)-based methods.

PCT/ICT-based methods construct a 2D mapping model between the vector formed by pupil/iris center and eye corner point and the POR. To determine the mapping model, the users need to sequentially stare at multiple on-screen calibration points during personal calibration to collect multiple sets of corresponding pupil/iris-corner vector and POR, so as to calibrate the user-specific coefficients of mapping model through regression (Cheung and Peng, 2015; George and Routray, 2016; Hu et al., 2019). PCRT/ICRT-based methods utilize the glint as the reference point, and construct a 2D mapping model between the vector formed by pupil/iris center and glint and the POR (Sigut and Sidha, 2011; Blignaut, 2013; Rattarom et al., 2017; Mestre et al., 2018). Similar to PCT/ICT-based methods, after determining the coefficients of mapping model through personal calibration, the POR can be estimated by substituting the pupil/iris-glint vector into the determined mapping model. The complexity of personal calibration for polynomial mapping methods such as PCT/ICT-based methods and PCRT/ICRT-based methods is related to the mapping model used. The most commonly used mapping model is the quadratic polynomial of the two components of the eye movement feature vector. There are 12 coefficients in this quadratic polynomial model, which uses nine calibration points for personal calibration. Cerrolaza et al. (2008) analyzed more than 400,000 mapping models, and demonstrated that higher-order polynomials cannot significantly improve the performance. A simple mapping model can still obtain ideal gaze accuracy (Jen et al., 2016; Xia et al., 2016). The simpler the polynomial, the fewer coefficients that need to be calibrated, and the fewer calibration points required for personal calibration. Xia et al. (2016) used a simple linear mapping between the pupil center and the screen point, and determined the mapping model using two calibration points.

CR-based methods use the invariance property of cross-ratios in projective transformations. The corneal reflection plane formed by the corneal reflections of light sources attached to four corners of the screen is taken as the medium, and the on-screen point corresponding to the pupil imaging center is calculated due to the equal CR of corresponding edges on the image plane and the screen, which is regarded as the gaze point (Coutinho and Morimoto, 2013; Arar et al., 2015, 2017; Cheng et al., 2017). Since the fact that the 3D pupil center and the corneal reflection points are not coplanar, it is necessary to determine a scale factor α through personal calibration to make the pupil center coplanar to the corneal reflection plane (Coutinho and Morimoto, 2013). In addition, the mapping point of the pupil imaging center on the screen is regarded as the gaze point, which is essentially the intersection point of the OA and the screen (POA), rather than the actual POR defined as the intersection of the VA and the screen. Therefore, it is also common to calibrate the kappa angle between the OA and the VA of the eyeball. Arar et al. (2015) introduced a personal calibration method using regularized least squares regression to compensate the kappa angle after using the conventional CR-based method to estimate the initial gaze point.

HN-based methods utilize two homography projection transformations from the image plane to the corneal reflection plane, as well as from the corneal reflection plane to the screen, to convert the pupil imaging center to a point on the screen, which is the gaze point. Due to the uncalibrated system, the corneal reflection plane is unknown, so it is necessary to define a normalized plane to replace it. The homography matrix from the image plane to the normalized plane is calculated by four glints generated by corneal reflection and four corner points of normalized plane. And the homography matrix from the normalized plane to the screen is determined through personal calibration using not less than four calibration points (Hansen et al., 2010; Shin et al., 2015; Morimoto et al., 2020).

2.2 Calibration information in 3D geometry-based methods

According to the type of camera used, 3D geometry-based methods can be divided into common-camera (CC)-based methods and depth-sensor (DS)-based methods.

Among the CC-based methods, the most common one is based on corneal reflection and pupil refraction (CRPR-based method). When a single-camera system is used, it is necessary to calibrate the corneal radius, the distance between 3D corneal center and 3D pupil center, and the kappa angle. In this way, the 3D corneal center and 3D pupil center can be estimated using the calibrated corneal radius and distance between 3D corneal center and 3D pupil center. The OA of the eyeball is represented by the line connecting 3D corneal center and 3D pupil center. Then the VA of the eyeball, that is the 3D gaze, can be converted from the OA using the calibrated kappa angle (Guestrin and Eizenman, 2006; Brousseau et al., 2018; O'Reilly et al., 2019). When a multi-camera system is used, 3D corneal center estimation can be simplified as the intersection of the reflection planes of two light sources of two cameras, and the OA of the eyeball can be obtained by the intersection of the refraction planes composed of 3D corneal center, camera optical center, and corresponding pupil imaging center of two cameras. Therefore, only the kappa angle is necessary to be calibrated to estimate the 3D gaze (Villanueva and Cabeza, 2008; Lidegaard et al., 2014). The personal calibration process of this method can be simplified to single-point calibration by fully utilizing the eyeball structure and the geometric imaging model (Villanueva and Cabeza, 2008; Lai et al., 2015). In addition, since the iris radius is an eye invariant parameter and the iris is less affected by corneal refraction (Hansen and Ji, 2010), 3D gaze estimation can be achieved using a single-camera system by calibrating the iris radius, corneal radius, and the kappa angle (Liu J. et al., 2020). In the absence of active light source, Cristina and Camilleri (2016) estimated the 3D gaze using a single camera based on a cylindrical head and spherical eye model, where the personal calibration information was detecting an initial frontal eye and head pose.

DS-based methods usually only require a depth camera such as Kinect. Due to the ability to obtain depth information, this method can estimate the head pose to obtain the transformation matrix between head and camera coordinate systems. According to the property that the eyeball center remains fixed relative to the head, the eyeball center in the head coordinate system can be calculated through personal calibration, and the 3D eyeball center in the camera coordinate system can be transformed. Combined with the 3D pupil or iris center calculated using the geometric imaging model, 3D corneal center can be determined and the OA of the eyeball can be constructed. By adding the kappa angle obtained from personal calibration, the VA of the eyeball can be calculated. The 3D POR can be obtained by intersecting the VA and the screen (Sun et al., 2014, 2015; Wang and Ji, 2016; Zhou et al., 2017). To fully consider individual differences, in addition to calibrating the eyeball center in the head coordinate system and the kappa angle, Sun et al. (2015, 2014) calibrated the eyeball radius, and the vector from eyeball center to inner eye corner in the head coordinate system. Wang and Ji (2016) additionally calibrated the eyeball radius and the distance between eyeball center and corneal center, where the distance between eyeball center and corneal center was used to estimate the 3D corneal center to represent the real gaze.

2.3 Calibration information in appearance-based methods

Appearance-based methods construct a mapping model between facial or eye appearance and the gaze. With the increase of the datasets available online, researchers can directly use existing datasets to conduct research on appearance-based gaze estimation. Due to the use of a large number of labeled training samples, this method usually does not require personal calibration to train a gaze estimation model. However, due to differences in eye appearance, illumination conditions, head pose, and distributions, this method requires sufficiently large and diverse data to produce accurate results (Uhm et al., 2020). Sugano et al. (2014) compared the mean estimation errors of random forest regression with person-specific training and cross-person training, which were 3.9° and 6.5°, respectively. This reflects the significant impact of individual differences and the limitations of the trained gaze estimation model. Therefore, some researchers used some calibration samples to train user-specific gaze estimation models (Zhang et al., 2015, 2019). In addition, some researchers used a few calibration samples for model calibration to improve the gaze estimation performance (Krafka et al., 2016; Park et al., 2018; He et al., 2019; Lindén et al., 2019; Chen and Shi, 2020; Gu et al., 2021; Liu G. et al., 2021; Wang et al., 2023). Lindén et al. (2019) modeled personal variations as a low-dimensional latent parameter space, and captured the range of personal variations by calibrating a spatial adaptive gaze estimator for a new person. Chen and Shi (2020) proposed to decompose the gaze angle into a subject-dependent bias term and a subject-independent gaze angle. During the test, the subject-dependent bias was estimated using a few images of the subject staring at a certain point, and then the gaze was represented by adding it to the calculated gaze angle. To alleviate the problem of information loss in the low-resolution gaze estimation task, Zhu et al. (2023) used the relatively fixed structure and components of human faces as prior knowledge and constructed a residual branch to recover the residual information between the low- and high-resolution images.

2.4 Discussion

Personal calibration is included in 2D mapping-based methods, 3D geometry-based methods, and appearance-based methods. The personal calibration process in appearance-based methods is the simplest, as it only requires the collection of calibration samples through the user's staring process, without the need for fine segmentation of image features. 2D mapping-based methods and 3D geometry-based methods need to detect visible features from the image and extract visible feature parameters, so as to calibrate specific parameters. 2D mapping-based methods use some feature center coordinates or feature points, such as the pupil center and glint, while 3D geometry-based methods often use some edge information of visual features on this basis, such as pupil imaging ellipse or pupil edge points. Therefore, comparatively, the personal calibration process in 3D geometry-based methods has the highest requirements on image processing, followed by 2D mapping-based methods. The personal calibration process in appearance-based methods has the lowest requirements.

From a performance perspective, compared to appearance-based methods, the most significant advantage of 2D mapping-based methods and 3D geometry-based methods is their high accuracy. Although both the 2D mapping-based method and the appearance-based method are mapping models, the 2D mapping-based method studies the essence of eye movement and determines the specific relationship between eye movement changes and gaze point through personal calibration. 3D geometry-based methods fully utilize visual features that reflect individual differences and eye movement changes. The calibration information mainly consists of some eye invariant parameters, which are relatively stable and not affected by head movement. It can be better adapted to the influence of head movement than 2D mapping-based methods that calibrate some mapping coefficients.

3 Personal calibration settings

Personal calibration is a process that is conducted before gaze tracking. Figure 4 shows the schematic diagram of personal calibration when users use different gaze tracking systems. When using a remote system (e.g., a laptop or a desktop eye-tracker combined with an external screen), the system camera, light sources, and screen are in close distance or nearly coplanar. The user needs to sit at a certain distance from the system and stare at some calibration points. During the user's staring process, the system will synchronously collect the user's face images, from which the user's facial or eye features can be obtained. The remote system has low interference to the user. When using smart glasses or a near-eye gaze tracker with a chin rest, the camera's optical axis deviates significantly from the gaze direction. The user is asked to stare at some calibration points a certain distance away while wearing the smart glasses or putting the head on the chin rest. During the staring process, the images of each eye are captured by a corresponding camera, and the features including pupil and glints can be extracted. When using a VR helmet, the helmet is equipped with a built-in display, and users can select the calibration function using the handle. The calibration process usually involves the user staring at several green dots that appear in sequence on the display, which is less affected by head movement. Alternatively, the user is asked to look at some displayed images (Chen and Ji, 2015; Wang et al., 2016; Alnajar et al., 2017; Hiroe et al., 2018). Personal calibration settings mainly involve system configuration, subject situation, experimental distance, calibration benchmark, and head movement. Table 1 lists the personal calibration settings mentioned in some typical methods, where “NA” represents those that are not mentioned or not available. In summary, the characteristics of personal calibration settings are reflected in:

Figure 4
www.frontiersin.org

Figure 4. Schematic diagram of personal calibration.

Table 1
www.frontiersin.org

Table 1. Personal calibration settings in some typical methods.

3.1 System configuration

The system configuration used for personal calibration is consistent with the studied gaze tracking system, and is usually determined by the gaze tracking algorithm. Except for some scenarios that require high gaze accuracy using multi-camera systems, most current methods use single-camera systems as the researchers continue to explore the simplest system configuration for gaze tracking. To ensure sufficient conditions for personal calibration and gaze estimation, many methods require the use of a single-camera-multi-light-source system. The methods that can estimate gaze by using a single-camera system include PCT/ICT-based methods (Cheung and Peng, 2015; Xia et al., 2016; Eom et al., 2019; Hu et al., 2019), DS-based methods (Sun et al., 2014, 2015; Xiong et al., 2014; Li and Li, 2016; Wang and Ji, 2016; Zhou et al., 2016, 2017), and appearance-based methods (Alnajar et al., 2017; Kellnhofer et al., 2019; Zhang et al., 2019; Chen and Shi, 2023). For example, Hu et al. (2019) used a single camera to construct a polynomial mapping function from the eye-movement vector to the gaze point, where the eye-movement vector was represented by the average of four vectors formed by the inner and outer corners of both eyes and the iris center. The coefficients of mapping function were determined through nine-point calibration, and the obtained function was used for gaze estimation. Sun et al. (2014) used a Kinect to calibrated the eyeball radius and the vector from eyeball center to eye inner corner in the head coordinate system online, thereby estimating the eyeball center and the iris center in the camera coordinate system to indicate the gaze. They also calibrated the kappa angle to consider the individual differences (Sun et al., 2015). Zhang et al. (2019) proposed the GazeNet architecture to train a mapping model from 2D head angle and eye image to the gaze angles. The performance of cross-dataset evaluation improved by 22% on MPIIGaze, whose data was collected during laptop use.

3.2 Subject situation

To verify the generality of the algorithm, individual differences among subjects should be fully considered, such as race, gender, age, memory, and experience. At present, most methods do not consider the diversity of subjects, and usually analyze based on experimental data of six to 20 subjects in the experimental validation. Some methods provide the gender and age of the subjects (Ma et al., 2014; Mestre et al., 2018; Wang and Ji, 2018; Uhm et al., 2020; Chen and Shi, 2023). Whether glasses are allowed to be worn is also an issue that should be considered, as wearing glasses not only blocks the eyes, but also causes reflection from the glasses. Some methods explicitly state that wearing glasses is not allowed (Lai et al., 2015; Wang and Ji, 2016; Xia et al., 2016; Liu J. et al., 2021). Eom et al. (2019) studied the gaze estimation method when wearing glasses, and learned a neural network from the center of black eye, inner and outer corners, and gaze direction using the samples collected when the user looked at nine calibration points on the screen. To reduce the impact of individual differences and enable the model to perform well on more individuals, the subject situation should be fully considered.

3.3 Experimental distance

The experimental distance is usually determined by the imaging range of the camera. To enable the required features to be imaged and ensure the imaging quality, there are specific requirements for the experimental distance. In most personal calibration processes, the distance between the user and the screen is within the range of 300–800 mm (Blignaut, 2013; Ma et al., 2014; Chen and Ji, 2015; Lai et al., 2015; Shin et al., 2015; Sun et al., 2015; Li and Li, 2016; Wang and Ji, 2016, 2018; Cheng et al., 2017; Hu et al., 2019; O'Reilly et al., 2019; Uhm et al., 2020; Liu J. et al., 2021), which is consistent with the scenario of using a computer. Some personal calibration processes are conducted in scenarios larger than 1 m, such as experiments in head-mounted systems (Mansouryar et al., 2016; Liu M. et al., 2020), or in vehicle driving scenarios (Yuan et al., 2022). Mansouryar et al. (2016) set the experimental distance to 1 m/1.25 m/1.5 m/1.75 m/2 m to investigate whether calibration at a single depth is sufficient, and concluded that the estimation performance can be improved with multiple calibration depths.

3.4 Calibration benchmark

The calibration benchmark here refers to the reference information used for calibration. In most cases, the set calibration benchmark is several on-screen calibration points with known coordinates (Blignaut, 2013; Ma et al., 2014; Lai et al., 2015; Shin et al., 2015; Sun et al., 2015; Wang and Ji, 2016; Xia et al., 2016; Cheng et al., 2017; Mestre et al., 2018; Hu et al., 2019; O'Reilly et al., 2019; Uhm et al., 2020; Liu J. et al., 2021). Blignaut (2013) compared the gaze estimation performance using a mapping model from pupil-glint vector to 2D POR when the number of calibration points was 5/9/14/18/23/135. He found that the gaze accuracy can reach 0.5° when the number of calibration points was not < 14. In contrast, Li and Li (2016) used an RGB camera as a calibration benchmark, and asked subjects to gaze at the RGB camera from different head positions to calibrate the eyeball center in the head coordinate system. Wang and Ji (2018) utilized complementary gaze constraint, center prior constraint, display boundary constraint, and angular constraint to obtain the kappa angle of left and right eyes through implicit personal calibration, without the need for known fixed calibration points. There are also some methods that can obtain saliency maps as the calibration benchmark for subjects to view images (Chen and Ji, 2015; Yan et al., 2017; Liu M. et al., 2020).

3.5 Head movement

In addition to being constrained by the camera's imaging range, the setting of head movement is mainly determined by the gaze tracking algorithm. For example, when constructing a mapping model from pupil/iris-corner vector or pupil/iris-glint vector to 2D POR using personal samples, the gaze accuracy would decrease significantly with head movement (Morimoto and Mimica, 2005; Sigut and Sidha, 2011). In 3D geometry-based methods, head rolling will generate a rotational component of the eye's OA around itself, which cannot be characterized by the eye visual features or the centers located on the OA. Even using a system with light sources, the glint formed by corneal reflection of the light source cannot reflect the rotational component due to the approximate spherical structure of the cornea. Therefore, some 3D gaze estimation methods limit head rotation (Li and Li, 2016; Wang and Ji, 2018; Liu J. et al., 2021). Reasonable setting of head movement can ensure the accuracy of personal calibration and provide conditions for accurate gaze estimation.

4 Personal calibration modes

Personal calibration plays an important role in gaze tracking as it can obtain user-specific parameters and improve gaze estimation performance. However, complex personal calibration not only increases the burden on users, but also limits its application range. This section discusses four personal calibration modes based on their complexity: explicit multi-point calibration, explicit single-point calibration, implicit/automatic calibration, and calibration-free.

4.1 Explicit multi-point calibration

Explicit multi-point calibration is the most common calibration mode. 2D mapping-based methods usually conduct an explicit calibration process where users stare at multiple calibration points to obtain sufficient data to calibrate the required personal information, thereby determining the gaze estimation model (Blignaut, 2013; Ma et al., 2014; Cheung and Peng, 2015; Shin et al., 2015; Jen et al., 2016; Mansouryar et al., 2016; Xia et al., 2016; Arar et al., 2017; Cheng et al., 2017; Mestre et al., 2018; Sasaki et al., 2018, 2019; Hu et al., 2019; Luo et al., 2020; Morimoto et al., 2020; Uhm et al., 2020). Nine calibration points are the most commonly used. Cheung and Peng (2015) used nine calibration points to determine the mapping model from iris-corner vector to POR. Subsequently, the AWPOSIT algorithm was used to compensate for the displacement of head movement based on the static POR. During personal calibration, the subjects were allowed to wear glasses. Luo et al. (2020) verified the proposed mapping equation based on homography transformation by using nine calibration points in a single-camera-single-light-source system with a head entrust stent. Compared with the calibrated classical quadratic polynomial mapping equation (with an accuracy of 0.99°), the gaze accuracy was within 0.5°. Although simplifying the mapping model can reduce the number of calibration points, at least two calibration points are required (Xia et al., 2016; Morimoto et al., 2020). For 3D geometry-based methods, if a single-camera system or depth sensor is used, at least five calibration points are usually used to calibrate the required eye parameters (Xiong et al., 2014; Wang and Ji, 2016; Brousseau et al., 2018; O'Reilly et al., 2019; Liu J. et al., 2020). Wang and Ji (2016) estimated the kappa angle, eyeball radius, distance between 3D corneal center and 3D eyeball center, as well as 3D eyeball center in the head coordinate system using five calibration points. Liu J. et al. (2020) determined the iris radius and kappa angle of subjects through five-point calibration, and the head roll was not allowed during calibration. In contrast, appearance-based methods typically do not require an explicit multi-point calibration process to obtain calibration samples.

4.2 Explicit single-point calibration

To avoid staring at multiple calibration points, researchers attempted to improve the gaze tracking algorithm by using a single calibration point to obtain personal information. The implementation of explicit single-point calibration for 2D mapping-based methods usually requires the use of some prior knowledge. Choi et al. (2014) constructed a user calibration database based on multi-point calibration in advance, which can then achieve gaze tracking by looking at a single calibration point. Yoon et al. (2019) combined the prior knowledge of average calibration with single-point calibration when determining the mapping model. Most explicit single-point calibration processes are conducted in 3D geometry-based methods. Using a single-camera-multi-light-source system or multi-camera system, CC-based method can achieve 3D gaze estimation with free head movement through single-point calibration by fully utilizing the geometric imaging model (Guestrin and Eizenman, 2007; Nagamatsu et al., 2008, 2021; Villanueva and Cabeza, 2008; Ebisawa and Fukumoto, 2013; Lai et al., 2015). DS-based methods can also calibrate the eyeball center and kappa angle in the head coordinate system through explicit single-point calibration. However, unlike the above calibration process, users need to focus on this calibration point at multiple head positions. Zhou et al. (2017) pointed out that each subject required to look at a given calibration point with two different head poses during calibration, while they asked the subject to gaze at the calibration point with 10 different head poses or gaze directions to obtain a more accurate estimation. For appearance-based methods, explicit single-point calibration is used to obtain several calibration images for gaze compensation (Liu G. et al., 2021; Chen and Shi, 2023).

4.3 Implicit/automatic calibration

Explicit calibration requires users to stare at known calibration points, and multi-point calibration is time consuming. Although the number of calibration points can be reduced to one, multi-point calibration shows stronger robustness than single-point calibration (Guestrin and Eizenman, 2007; Morimoto et al., 2020), and the naturalness of interaction is more pursued by people. Therefore, studying implicit/automatic calibration methods has become a major research trend.

On one hand, to avoid the need for known calibration points, the natural constraint that the gazes of left and right eyes converge at one point is often used (Model and Eizenman, 2010; Wang and Ji, 2018; Wen et al., 2020). Model and Eizenman (2010) automatically calibrated the kappa angle in a two-camera-four-light-source system by minimizing the distance between the intersections of the VAs of left and right eyes with one or more observation surfaces. Wen et al. (2020) proposed a personal calibration process that requires users to focus on a specific yet unknown point and move their heads (rotate and translate) while maintaining focus. Based on the constraint of minimizing the distance between the PORs of left and right eyes, the parameters including eyeball radius, 3D eyeball center, and kappa angle were calibrated.

On the other hand, utilizing visual saliency to obtain gaze information is also an effective way (Chen and Ji, 2015; Wang et al., 2016; Alnajar et al., 2017; Hiroe et al., 2018; Liu M. et al., 2020). Hiroe et al. (2018) proposed an implicit personal calibration method using the saliency map around the OA of the eye, where the peak of the mean saliency map was used to represent the VA in the eyeball coordinate system. Liu M. et al. (2020) conducted a calibration process where users can randomly scan the surrounding environment, and then established a mapping model using gaze vectors and possible 3D calibration targets in the scene calculated using saliency maps. Especially, Sugano et al. (2015) used a monocular camera to capture the user's head pose and eye images during mouse clicks, and incrementally updated the local reconstruction-based gaze estimation model by clustering the head poses, to learn the mapping between eye appearance and gaze direction continuously and adaptively. Yuan et al. (2022) automatically calibrated the gaze estimation model through gaze pattern learning in driving scenarios, where the representative time samples of forward-view gaze zone, left-side mirror, right-side mirror, and rear-view mirror, speedometer, and center stack were implicitly selected as calibration points. Implicit/automatic calibration methods can gradually adapt the system to users and improve the performance during interaction (Sun et al., 2014; Chen and Ji, 2015).

4.4 Calibration-free

Most calibration-free methods are based on facial or eye appearance, and researchers directly use existing datasets to train user-independent gaze estimation models (Wood et al., 2015, 2016; Wen et al., 2017a,b; Li and Busso, 2018; Cheng et al., 2020a,b, 2021; Liu S. et al., 2020; Zhang et al., 2020, 2022; Bao et al., 2021; Cai et al., 2021; Murthy and Biswas, 2021; Abdelrahman et al., 2022; Donuk et al., 2022; Hu et al., 2022; Zhao et al., 2022; Huang et al., 2023; Ren et al., 2023). Hu et al. (2022) trained the gaze estimation model using saliency features and semantic features from the DR(eye)VE dataset. Liu S. et al. (2020) took a full-face image as input, and estimated the 3D gaze by using a multi-scale channel unit and a spatial attention unit for selecting and increasing important features, respectively. Appearance-based methods mainly use the 3D eyeball center to estimate the gaze, however, the VA of the eyeball passes through the corneal center, not the eyeball center, so this is an approximate estimation. For 2D mapping-based methods and 3D geometry-based methods, the implementation of calibration-free mainly relies on parameter settings or model approximation. The former uses the parameters from classical eyeball model or population averages to set some parameters that originally need to be calibrated (Morimoto et al., 2002; Coutinho and Morimoto, 2006, 2013), such as setting the scale factor to 2 in the CR-based method (Coutinho and Morimoto, 2006); setting the corneal radius to 7.8 mm (Coutinho and Morimoto, 2013). The latter does not consider the kappa angle and directly approximates the OA of the eyeball to represent the VA of the eyeball (Shih et al., 2000). To reduce the error caused by this approximation, researchers propose the binocular model and estimate the POR using the midpoint of the POAs of left and right eyes (Hennessey and Lawrence, 2009; Nagamatsu et al., 2010, 2011). Nagamatsu et al. (2011) used four cameras and three light sources to calculate the corneal center and the OA direction based on a geometric imaging model, thereby determining the OA of a single eye. They used the midpoint of the POAs of left and right eyes as a good approximation of the POR.

4.5 Discussion

The characteristics of different personal calibration modes are shown in Figure 5. To obtain personal information, explicit multi-point calibration requires users to focus on multiple known calibration points on the screen in sequence, while explicit single-point calibration requires users to focus on a known calibration point. Implicit/automatic calibration collects personal information of users while looking at unknown points or images, or browsing freely. The calibration-free mode does not require the user to perform any specific operations, it is mainly implemented based on parameter settings or model approximation. From the perspective of convenience and user experience, calibration-free is the best mode, followed by implicit/automatic calibration, while explicit multi-point calibration leads to a heavier burden on the user. From the perspective of robustness and accuracy, explicit multi-point calibration and implicit/automatic calibration have stronger robustness and higher accuracy due to the use of more personal information, which is superior to explicit single-point calibration and calibration-free. Nagamatsu et al. (2008) pointed out that the accuracy of a calibration-free gaze tracking system is generally lower than that of a tracking system with a calibration, and there is a trade-off between accuracy and calibration mode. Each personal calibration mode has its own characteristics, and the selection of calibration mode should be based on practical application requirements. For example, virtual reality helmets currently use explicit multi-point calibration to ensure the high accuracy and robustness of the device. Taking the Pico G2 4K virtual reality all-in-one headset as an example, the calibration mode used is for users to gaze at five calibration points that appear sequentially on the virtual screen.

Figure 5
www.frontiersin.org

Figure 5. Characteristics of different calibration modes.

Through the analysis of different personal calibration modes, it can be seen that the personal calibration modes used in different gaze tracking methods have certain regularity. Table 2 lists the personal calibration modes in some gaze tracking methods. 2D mapping-based methods mainly adopt explicit multi-point calibration. Explicit single-point calibration can be achieved by simplifying the mapping model, but prior knowledge is required. 3D geometry-based methods can simplify the explicit calibration process by using a more complex system or improving the gaze tracking algorithm. They can also utilize natural constraints such as “the gazes of left and right eyes converge at one point” for implicit/automatic calibration. To avoid personal calibration, some user-specific parameters can be set using the parameters from classical eyeball model or population averages, or the OA of the eyeball can be approximated to represent the VA of the eyeball, thus avoiding kappa angle calibration in some single-point calibration methods. With the continuous increase of datasets, there has been a significant increase in calibration-free gaze tracking methods proposed based on appearance in recent years. However, due to individual differences, head pose differences, and environmental differences, etc., the generalization ability of appearance-based gaze estimation models is far from reaching the level of universal applications. Table 3 lists the accuracy of some state-of-the-art methods. It can be seen that under different system configurations and calibration modes, the gaze accuracy of 2D mapping-based methods and 3D geometry-based methods is generally within 2°, while the gaze accuracy of appearance-based methods is usually above 3°, even if calibration samples are used.

Table 2
www.frontiersin.org

Table 2. Personal calibration modes in some gaze tracking methods.

Table 3
www.frontiersin.org

Table 3. Comparison of the state-of-the-art methods using different personal calibration modes.

5 Performance simulations of typical personal calibration methods

To indicate the characteristics of different settings in personal calibration intuitively, this section conducted some comparative simulation experiments on typical personal calibration methods for 2D and 3D gaze estimation, to reflect several key issues of personal calibration and provide reference for researchers to design personal calibration.

5.1 Personal calibration simulations in PCRT/ICRT-based methods

The PCRT/ICRT-based method is the most commonly used method for 2D gaze estimation, and is the core eye tracking technology used by many eye tracking device manufacturers. It essentially establishes a mapping model from pupil/iris-glint vector to 2D POR through personal calibration, and uses the calibrated mapping model for 2D gaze estimation and tracking. The mapping effect of the calibrated model determines the gaze estimation performance. There are two main factors that affect the mapping effect: the 2D mapping model used and the number of calibration points used for mapping model calibration. Therefore, we discussed the performance of using different 2D mapping models and using different numbers of calibration points in personal calibration, even the distribution of calibration points.

5.1.1 Comparison of 2D mapping models

Various mapping models can be selected, such as linear, quadratic, or higher-order. Some researchers have specifically studied different 2D mapping models and stated that higher order polynomials do not noticeably improve system behavior (Cerrolaza et al., 2008; Blignaut, 2013), so we focused on performance analysis of polynomials below fourth order, and selected six commonly used mapping models from existing literature that cover different orders and forms, as shown in Figure 6. In the models, sx, sy are the horizontal and vertical coordinates of the 2D POR, vxg, vyg are the horizontal and vertical components of the corresponding vector from pupil center to glint, and ai, bi are the coefficients of mapping models.

Figure 6
www.frontiersin.org

Figure 6. Six mapping models for analysis.

To compare these six models, we set simulation parameters based on the eyeball structure and the geometric imaging model. The simulation parameters are as follows: the light source was set at (–120, –3, 10), the camera focal length was 6 mm, and the corneal radius was 7.8 mm. The 3D corneal center was set at (–42.7208, –60.4287, 332.4721), and the coordinates of glint were (0.7968, 1.0908, –6). Nine evenly distributed screen points were located at: (100, 75), (200, 75), (300, 75), (100, 150), (200, 150), (300, 150), (100, 225), (200, 225), (300, 225), and the coordinates of their corresponding pupil imaging centers were (0.801, 1.0201, –6), (0.7718, 1.0185, –6), (0.745, 1.0233, –6), (0.8031, 1.0375, –6), (0.7739, 1.0358, –6), (0.7471, 1.0392, –6), (0.8036, 1.0568, –6), (0.7765, 1.0554, –6), (0.75, 1.0567, –6). Using the coordinates of pupil imaging centers and glint, the vector from pupil center to glint corresponding to each screen calibration point was calculated. By taking the same nine sets of pupil-glint vector and the screen calibration point as input, the coefficients (ai, bi) in the six different models can be fitted to determine the mapping models. To analyze the performance of determined mapping models, the pupil-glint vector corresponding to each screen point was substituted into each mapping model to predict the 2D POR. The Euclidean distance between the predicted POR and the ground-truth of screen point was calculated to represent the POR error, as Figure 7 shows. When using different 2D mapping models, the root mean squared errors (RMSEs) of the nine PORs were also calculated. It can be seen that increasing the model order can reduce the POR error to a certain extent, but more complex mapping model does not have better gaze estimation performance absolutely. On the contrary, more complex models have more coefficients to be calibrated, which requires more calibration points and increases the complexity of personal calibration. Therefore, on the basis of meeting certain performance requirements, it is superior to select a simpler model. At present, the most widely used is Model 4.

Figure 7
www.frontiersin.org

Figure 7. Performance comparison of calibrating different 2D mapping models.

5.1.2 Comparison of the number of calibration points

The selection of the number of calibration points should consider the minimum number required for the mapping model. For Model 4, there are six coefficients in each polynomial that need to be determined through personal calibration, so the number of calibration points used for this model is generally not less than six. To improve the performance, more calibration points may be used. Here, we analyzed the impact of the number of calibration points used in personal calibration on its performance.

The simulation parameters are set as follows: the light source, camera focal length, and corneal radius were the same as before. Twenty evenly distributed screen points (five rows and four columns) were set. Under the constraint that the distance between the 3D corneal center and the screen is 300 mm, the 3D corneal center was randomly generated, and the glint was determined accordingly by the corneal reflection of light source. Then the corresponding pupil imaging center for each screen point was determined based on the eyeball structure and the pupil refraction. Since Model 4 is the most widely used model, here, we used it to compare the gaze estimation performance when the number of calibration points is 5/6/9/16/20, respectively. Twenty evenly distributed screen points were numbered clockwise in a row. When studying 20 calibration points, the pupil-glint vectors and screen point coordinates corresponding to all screen points were used to calibrate the mapping model. When studying 5/6/9/16 calibration points, the calibration data was selected from the data of 20 screen points to ensure a single variable. Considering the generalization ability of the mapping model, the scattered calibration points on the screen were used as much as possible in our simulation. When studying five upper calibration points, the selected screen points were [1, 3, 5, 7, 9]; When studying five uniformly distributed calibration points, the selected screen points were [1, 4, 10, 17, 20]. When studying six calibration points, the selected screen points were [1, 7, 15, 17, 19, 20]. When studying nine calibration points, the selected screen points were [1–3, 10–12, 17–19]. When studying 16 calibration points, the selected screen points were [1–8, 13–20]. Using the pupil-glint vectors and screen point coordinates corresponding to these screen points as calibration data, the coefficients of Model 4 were fitted using different amounts of calibration data to determine the 2D mapping model. Then the 2D POR was predicted by substituting the pupil-glint vector corresponding to each screen point into the calibrated mapping model. The Euclidean distance between the predicted 2D POR and the ground-truth of screen point was calculated to represent the POR error, and the RMSEs of the 20 PORs when the mapping model was calibrated using different numbers of calibration points were also calculated, as shown in Figure 8. When using five calibration points, if all the points are in the upper part of the screen, the POR error was relatively large. In contrast, if the calibration points are uniformly distributed, the POR error was relatively small. When using six calibration points, small POR errors only exist at the screen points used as calibration points. This is because fewer calibration points cannot accurately represent the mapping relationship between eye features and any screen position. When there are less than six calibration points, polynomial fitting is underdetermined; When using six calibration points, there may be significant errors at certain points due to the inconsistency between the mapping relationship and the fitted mapping model. In contrast, when using more than six calibration points, the calibrated mapping model achieved high gaze accuracy. Therefore, the gaze estimation performance can be improved by increasing the number of calibration points, but it comes at the cost of increasing the complexity of personal calibration.

Figure 8
www.frontiersin.org

Figure 8. Performance comparison of 2D mapping models calibrated with different numbers of calibration points.

5.2 Personal calibration simulations in CC-based methods

Most 3D gaze estimation methods are based on common cameras, which estimate eye parameters such as corneal center and pupil center according to the basic structure and the geometric imaging model of the eyeball, in order to construct the OA, and then use the kappa angle to convert the 3D gaze. In this process, to take individual differences into consideration, user-specific parameters such as corneal radius and kappa angle are usually obtained through a personal calibration process. Therefore, based on our experience, we explored the two most prominent issues in the personal calibration for 3D gaze estimation. One is the setting of calibration parameters when calibrating the user-specific corneal radius, and the other is the selection of calibration modes when calibrating the kappa angle.

5.2.1 Comparison of calibration parameters

When using the CC-based method in a single-camera system, it is usually necessary to calibrate the user-specific corneal radius. Here, the importance of selecting calibration parameters is emphasized by analyzing several calibration parameter settings when calibrating the corneal radius. To replicate the typical CC-based method (Guestrin and Eizenman, 2006), the simulation parameters are set as follows: we set two light sources, located at (–120, –3, 10) and (120, –3, 10). The camera focal length was 6 mm. The ground-truth of corneal radius was 7.8 mm, and the distance between the 3D corneal center and the 3D pupil center was 4.2 mm. The horizontal and vertical components of kappa angle were 5 and 1.5°, respectively. We set nine evenly distributed screen points, and randomly generated the 3D corneal center corresponding to each screen point based on the constraint of 500 mm distance from the 3D corneal center to the screen. Using the 3D corneal center and corneal radius, the glints formed by two light sources were calculated based on the corneal reflection.

During corneal radius calibration, the parameters used include the positions of two light sources and their corresponding glints. We used the same nine sets of data (corresponding to nine evenly distributed screen points) for testing, with only the set calibration parameters being different. According to the corneal reflection of the light source, there are: ① the distance between the reflection point and the 3D corneal center is equal to the corneal radius; ② the incident light, normal, and reflected light are coplanar; ③ the incident angle is equal to the reflection angle. Using these properties, we studied three calibration parameter settings: (1) the calibration parameters were directly corneal radius R and 3D corneal center C. According to ①, the reflection points were represented by R and C. Then, four equations were constructed using ② and ③ to optimize the solutions of R and C; (2) According to ②, the reflection planes of two light sources contain the 3D corneal center C and the camera optical center O, so the vector OC was obtained by intersecting the reflection planes of two light sources. Then, the 3D corneal center C was represented by the product of a proportional coefficient t and the vector OC. The reflection point Gi(i = 1, 2) was represented by the product of the proportional coefficients ui and the corresponding glint gi. Four unknowns were R, t, ui, and they were optimized by constructing four equations using ① and ③; (3) On the basis of (2), ui was represented by using R and C according to ①, thereby the reflection point Gi was represented by ui and gi. At this time, there were only two unknowns: R and t. Two equations were constructed using ③ for optimization.

We used the least square method to solve the above calibration parameters. To test the robustness of the calibration results, we set three different initial values for each calibration parameter setting. The corneal radius error obtained through calibration is shown in Table 4. The results of directly calibrating R and C depend on the setting of initial values, and it is not easy to set the initial coordinates of 3D corneal center. The latter two calibration parameter settings first use the intersection of two corneal reflection planes to determine the vector OC, ensuring the absoluteness of the 3D corneal center C on the line OC and eliminating the deviation of 3D corneal center C. Within a certain range of initial values, the corneal radius can be accurately calibrated. However, it is superior to reduce the calibration parameters to R and t to avoid the initial setting of some non-intuitive parameters (e.g., ui needs to be determined by roughly estimating the ratio of the reflection point Gi to the glint gi).

Table 4
www.frontiersin.org

Table 4. Comparison of corneal radius calibration under different calibration parameter settings.

5.2.2 Comparison of calibration modes

Kappa angle is a significant parameter for converting from the OA to the VA of the eyeball. We compared the performance under different calibration modes by analyzing kappa angle calibration. The simulation parameters are set as follows: Using the parameters set in Section 5.2.1, the pupil imaging ellipse parameters corresponding to each screen point were calculated based on eyeball structure and pupil refraction. They are necessary in implicit multi-point calibration.

We evaluated the following five methods: (1) Explicit multi-point calibration: the OA vectors and VA vectors corresponding to nine screen points were used to calibrate a transformation matrix for converting the OA to the VA of the eyeball (Zhu and Ji, 2008; Wan et al., 2022); (2) Explicit single-point calibration: Due to the fact that the kappa angle is an eye invariant parameter, each screen point corresponds to an equal kappa angle. That is to say, any screen point can be selected as the calibration point. We selected the fifth screen point here. After calculating the horizontal and vertical angles of OA and VA in the eyeball coordinate system when staring at the fifth screen point respectively, the horizontal and vertical angles of kappa angle were calculated by subtracting the corresponding angles of OA and VA (Guestrin and Eizenman, 2006; Zhou et al., 2016); (3) Implicit multi-point calibration: the kappa angle was automatically calibrated by complementary gaze constraint, center prior constraint, display boundary constraint and angular constraint using data from nine screen points (Wang and Ji, 2018); (4) Calibration-free (set kappa): Considering that the set kappa angle deviates from the ground-truth in most cases, the horizontal and vertical components of the kappa angle were set to 5 and 1°, respectively. The difference between the vertical angle and the ground-truth was 0.5°; (5) Calibration-free (POA average): the kappa angle was not considered, and the average of the POAs of left and right eyes was used as the estimated POR (Hennessey and Lawrence, 2009; Nagamatsu et al., 2010, 2011).

Using the same simulation data, the kappa angle calibrated using the above five methods are shown in Table 5. Then, the calibrated kappa angle parameters were used to estimate the POR when staring at each screen point. The Euclidean distance between the predicted POR and the ground-truth of screen point was calculated to represent the POR error, as Figure 9 shows. The RMSEs of the nine PORs when using different calibration modes were calculated and labeled. It can be seen that when the kappa angle obtained from explicit single-point calibration is accurate, the accurate POR can be estimated. Explicit or implicit multi-point calibration includes the optimization process, resulting in small kappa angle errors. The estimated POR error was < 1.2 mm, and the RMSE was < 0.2 mm. When using the calibration-free method with the set kappa angle, the POR error was about 5 mm due to a deviation of 0.5° in the vertical. When using the calibration-free method of averaging the POAs of left and right eyes, the POR error was about 15 mm. This is because the average operation can only offset the horizontal component of the kappa angle of left and right eyes. In our simulation model, the vertical component of the kappa angle was 1.5°, which introduces a significant error. Therefore, personal calibration is necessary in most cases.

Table 5
www.frontiersin.org

Table 5. Comparison of kappa angle results under different calibration modes.

Figure 9
www.frontiersin.org

Figure 9. Comparison of 3D gaze estimation performance under different calibration modes. (A) Explicit/implicit calibration. (B) Calibration-free (using preset values/using the average of POAs of both eyes).

6 Key issues of personal calibration

Personal calibration is a key factor affecting the performance of gaze estimation. This section discusses several key issues related to personal calibration.

6.1 Determination of personal calibration information

Except for appearance-based methods that can directly use the calibration images to improve the gaze estimation performance, other methods require the calculation of user-specific information through personal calibration. When determining the required personal calibration information, it is advisable to select some eye invariant parameters or parameters with specific theoretical ranges as the information that needs to be calibrated. In addition, the amount of calibration information should be minimized as much as possible. If parameters can be converted to each other, it is meaningless to set them all as the calibration information, which will also affect the accuracy of calibration results. For example, when reconstructing the eye pose in space based on the geometric imaging model, some proportional coefficients are often used. There are certain positional relationships between spatial points represented by these proportional coefficients, such as perpendicular or equidistant. In this case, these proportional coefficients can be represented by some eyeball parameters (3D corneal center and corneal radius) based on these positional relationships, to avoid error introduction when optimizing these proportional coefficients. Selecting eye invariant parameters or parameters with specific theoretical ranges as personal calibration information can also provide a basis for setting initial values during solving.

6.2 Settings of personal calibration process

The settings such as system configuration, experimental distance, and head movement are generally related to gaze tracking systems or gaze estimation algorithms. The experimental distance determines the imaging quality, so its variation range is relatively fixed. According to the proposed gaze tracking algorithm, there is usually a minimum standard for system configuration. For example, at least one camera and one light source are required to construct the pupil-glint vector for PCRT/ICRT-based methods. The performance can be improved by using additional system configuration. Reducing head movement can ensure the accuracy of personal calibration, but it comes at the cost of increasing user burden. For some methods that allow for free head movement or the use of a head-mounted system, head movement setting do not need to be restricted. The subject situation and calibration benchmark are relatively flexible settings. The subject situation has certain limitations in existing methods. For unconstrained gaze estimation methods, it is necessary to expand the coverage range of subjects or increase the diversity of subjects as much as possible to verify the generality of the algorithm. The setting of calibration benchmark is actually to obtain personal calibration samples. If setting more calibration benchmarks and collecting more personal samples, it is usually easier to calculate accurate personal information. But in this way, the personal calibration process that users participate in becomes more complex. Therefore, the setting of calibration benchmark needs to balance the practical requirements for performance with the user operational complexity.

6.3 Selection of personal calibration mode

Simplifying or exempting personal calibration has always been a research trend in gaze tracking. Although many single-point calibration, implicit/automatic calibration, and calibration-free methods have been proposed, they have only been proven theoretically feasible to a large extent. In practical applications, explicit multi-point calibration is mainly used. The current technical level is not sufficient to achieve accurate calibration-free yet universal gaze estimation. Therefore, it is necessary to conduct personal calibration. In addition, it is of great importance to consider the comfort level of user participation in personal calibration in the new situation of pursuing natural interaction. Researchers are attempting to improve the conventional explicit multi-point calibration used in gaze tracking to simpler explicit single-point calibration or more natural implicit/automatic calibration. Through the analysis of operation mode and performance, implicit/automatic calibration is a relatively optimal personal calibration mode. It obtains gaze information during the calibration process through other methods, or directly establishes connections between calibration information using some natural constraints, thereby eliminating the need for explicit calibration points. This implicit/automatic calibration mode can also update user-specific features continuously, making personal calibration information more accurate and stable.

Overall, personal calibration is a necessary means to achieve accurate gaze tracking. The operability and real-time performance of gaze tracking can be improved by calibrating the necessary personal information, setting a reasonable personal calibration process, and selecting an optimal personal calibration mode. This helps to promote gaze tracking as an important channel for natural human-computer interaction.

7 Conclusion

This paper provides a comprehensive overview on personal calibration for VOG-based gaze tracking. It summarizes the personal calibration information required in different gaze estimation methods, and analyzes the settings of personal calibration from five aspects: system configuration, subject situation, experimental distance, calibration benchmark, and head movement. Subsequently, several existing personal calibration modes are analyzed. The characteristics of different personal calibration settings for typical 2D and 3D methods were analyzed through simulation experiments. Finally, three key issues for designing personal calibration are discussed, namely: How to determine personal calibration information? How to set some items in personal calibration? How to select personal calibration mode? This paper has guiding significance for conducting personal calibration and promoting natural interaction.

Although many convenient personal calibration methods have been proposed in theoretical research, they are rarely implemented in industrial applications. On one hand, this is because the generality verification in theoretical research is insufficient, such as limitations on head movement and wearing glasses, which limit the application scenarios of theoretical methods. On the other hand, this is related to product requirements, and a simple personal calibration mode may not meet the performance indicators such as accuracy. To better integrate with practical application requirements, the following directions can be studied:

(1) Research on unconstrained personal calibration scenarios: Unconstrained personal calibration is a more user-friendly experience that requires minimizing the limitations in personal calibration settings to enable subjects to naturally complete the calibration process. It not only requires the calibration algorithm to be feasible in the staring state, but also requires reasonable processing of personal information collected in specific or extreme situations such as blink and saccade.

(2) Research on some general criteria or constraints in gaze patterns: implicit/automatic calibration is a relatively natural calibration mode. It is recommended to conduct refined research on some basic criteria or natural constraints and verify their universality, which can broaden the thinking of designing personal calibration and promote the application of natural calibration mode.

(3) Research on personal calibration modes in specific application scenarios: Conducting research on personal calibration modes for specific application scenarios makes it easier to balance various indicators and improve some important indicators in a targeted manner. Taking head-mounted devices as an example, the factor of head movement can be approximately ignored, while high accuracy, strong robustness and real-time performance are the indicators we pursue.

Author contributions

JL: Conceptualization, Data curation, Funding acquisition, Investigation, Methodology, Writing – original draft, Writing – review & editing. JC: Formal analysis, Funding acquisition, Methodology, Supervision, Writing – review & editing. ZY: Conceptualization, Methodology, Validation, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This research was funded in part by the National Science Foundation for Young Scholars of China under Grant 62206016, the Guangdong Basic and Applied Basic Research Foundation under Grant 2022A1515140016 and 2023A1515140086, and the Fundamental Research Funds for the Central Universities under Grant FRF-GF-20-04A.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abdelrahman, A. A., Hempel, T., Khalifa, A., and Al-Hamadi, A. (2022). L2cs-net: fine-grained gaze estimation in unconstrained environments. arXiv [Preprint]. arXiv.2203.03339. doi: 10.48550/arXiv.2203.03339

Crossref Full Text | Google Scholar

Alnajar, F., Gevers, T., Valenti, R., and Ghebreab, S. (2017). Auto-calibrated gaze estimation using human gaze patterns. Int. J. Comput. Vis. 124, 223–236. doi: 10.1007/s11263-017-1014-x

Crossref Full Text | Google Scholar

Arar, N. M., Gao, H., and Thiran, J.-P. (2015). “Towards convenient calibration for cross-ratio based gaze estimation,” in 2015 IEEE Winter Conference on Applications of Computer Vision (Waikoloa, HI: IEEE), 642–648. doi: 10.1109/WACV.2015.91

Crossref Full Text | Google Scholar

Arar, N. M., Gao, H., and Thiran, J. P. (2017). A regression-based user calibration framework for real-time gaze estimation. IEEE Trans. Circuits Syst. Video Technol. 27, 2623–2638. doi: 10.1109/TCSVT.2016.2595322

Crossref Full Text | Google Scholar

Bao, Y., Cheng, Y., Liu, Y., and Lu, F. (2021). “Adaptive feature fusion network for gaze tracking in mobile tablets,” in 2020 25th International Conference on Pattern Recognition (ICPR) (Milan: IEEE), 9936–9943. doi: 10.1109/ICPR48806.2021.9412205

Crossref Full Text | Google Scholar

Blignaut, P. (2013). Mapping the pupil-glint vector to gaze coordinates in a simple video-based eye tracker. J. Eye Mov. Res. 7, 1–11. doi: 10.16910/jemr.7.1.4

Crossref Full Text | Google Scholar

Brousseau, B., Rose, J., and Eizenman, M. (2018). Accurate model-based point of gaze estimation on mobile devices. Vision 2:35. doi: 10.3390/vision2030035

PubMed Abstract | Crossref Full Text | Google Scholar

Cai, X., Chen, B., Zeng, J., Zhang, J., Sun, Y., Wang, X., et al. (2021). Gaze estimation with an ensemble of four architectures. arXiv [Preprint]. arXiv.2107.01980. doi: 10.48550/arXiv.2107.01980

Crossref Full Text | Google Scholar

Cerrolaza, J. J., Villanueva, A., and Cabeza, R. (2008). “Taxonomic study of polynomial regressions applied to the calibration of video-oculographic systems,” in Proceedings of the 2008 Symposium on Eye Tracking Research & Applications (New York, NY: Association for Computing Machinery), 259–266. doi: 10.1145/1344471.1344530

Crossref Full Text | Google Scholar

Chen, J., and Ji, Q. (2015). A probabilistic approach to online eye gaze tracking without explicit personal calibration. IEEE Trans. Image Process. 24, 1076–1086. doi: 10.1109/TIP.2014.2383326

PubMed Abstract | Crossref Full Text | Google Scholar

Chen, Z., and Shi, B. E. (2020). “Offset calibration for appearance-based gaze estimation via gaze decomposition,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (Snowmass, CO: IEEE), 270–279. doi: 10.1109/WACV45572.2020.9093419

Crossref Full Text | Google Scholar

Chen, Z., and Shi, B. E. (2023). Towards high performance low complexity calibration in appearance based gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 45, 1174–1188. doi: 10.1109/TPAMI.2022.3148386

PubMed Abstract | Crossref Full Text | Google Scholar

Cheng, H., Liu, Y., Fu, W., Ji, Y., Lu, Y., Yang, Z., et al. (2017). Gazing point dependent eye gaze estimation. Pattern Recognit. 71, 36–44. doi: 10.1016/j.patcog.2017.04.026

Crossref Full Text | Google Scholar

Cheng, Y., Bao, Y., and Lu, F. (2021). Puregaze: purifying gaze feature for generalizable gaze estimation. arXiv [Preprint]. arXiv.2103.13173. doi: 10.48550/arXiv.2103.13173

Crossref Full Text | Google Scholar

Cheng, Y., Huang, S., Wang, F., Qian, C., and Lu, F. (2020a). A coarse-to-fine adaptive network for appearance-based gaze estimation. Proc. AAAI Conf. Artif. Intell. 34, 10623–10630. doi: 10.1609/aaai.v34i07.6636

Crossref Full Text | Google Scholar

Cheng, Y., Zhang, X., Lu, F., and Sato, Y. (2020b). Gaze estimation by exploring two-eye asymmetry. IEEE Trans. Image Process. 29, 5259–5272. doi: 10.1109/TIP.2020.2982828

PubMed Abstract | Crossref Full Text | Google Scholar

Cheung, Y. M., and Peng, Q. M. (2015). Eye gaze tracking with a web camera in a desktop environment. IEEE Trans. Hum.-Mach. Syst. 45, 419–430. doi: 10.1109/THMS.2015.2400442

Crossref Full Text | Google Scholar

Choi, K. A., Ma, C., and Ko, S. J. (2014). Improving the usability of remote eye gaze tracking for human-device interaction. IEEE Trans. Consumer Electron. 60, 493–498. doi: 10.1109/TCE.2014.6937335

Crossref Full Text | Google Scholar

Coutinho, F. L., and Morimoto, C. H. (2006). “Free head motion eye gaze tracking using a single camera and multiple light sources,” in Proceedings of the 2006 19th Brazilian Symposium on Computer Graphics and Image Processing (Amazonas: IEEE). doi: 10.1109/SIBGRAPI.2006.21

Crossref Full Text | Google Scholar

Coutinho, F. L., and Morimoto, C. H. (2013). Improving head movement tolerance of cross-ratio based eye trackers. Int. J. Comput. Vis. 101, 459–481. doi: 10.1007/s11263-012-0541-8

Crossref Full Text | Google Scholar

Cristina, S., and Camilleri, K. P. (2016). Model-based head pose-free gaze estimation for assistive communication. Comput. Vis. Image Underst. 149, 157–170. doi: 10.1016/j.cviu.2016.02.012

Crossref Full Text | Google Scholar

Donuk, K., Ari, A., and Hanbay, D. (2022). A cnn based real-time eye tracker for web mining applications. Multimed. Tools Appl. 81, 39103–39120. doi: 10.1007/s11042-022-13085-7

Crossref Full Text | Google Scholar

Drakopoulos, P., alex Koulieris, G., and Mania, K. (2021). Eye tracking interaction on unmodified mobile VR headsets using the selfie camera. ACM Trans. Appl. Percept. 18:11. doi: 10.1145/3456875

Crossref Full Text | Google Scholar

Ebisawa, Y., and Fukumoto, K. (2013). Head-free, remote eye-gaze detection system based on pupil-corneal reflection method with easy calibration using two stereo-calibrated video cameras. IEEE Trans. Biomed. Eng. 60, 2952–2960. doi: 10.1109/TBME.2013.2266478

PubMed Abstract | Crossref Full Text | Google Scholar

Eom, Y., Mu, S., Satoru, S., and Liu, T. (2019). “A method to estimate eye gaze direction when wearing glasses,” in 2019 International Conference on Technologies and Applications of Artificial Intelligence (Kaohsiung: IEEE), 1–6. doi: 10.1109/TAAI48200.2019.8959824

Crossref Full Text | Google Scholar

George, A., and Routray, A. (2016). Fast and accurate algorithm for eye localisation for gaze tracking in low-resolution images. IET Comput. Vis. 10, 660–669. doi: 10.1049/iet-cvi.2015.0316

Crossref Full Text | Google Scholar

Gu, S., Wang, L., He, L., He, X., and Wang, J. (2021). Gaze estimation via a differential eyes' appearances network with a reference grid. Engineering 7, 777–786. doi: 10.1016/j.eng.2020.08.027

Crossref Full Text | Google Scholar

Guestrin, E. D., and Eizenman, M. (2006). General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Trans. Biomed. Eng. 53, 1124–1133. doi: 10.1109/TBME.2005.863952

PubMed Abstract | Crossref Full Text | Google Scholar

Guestrin, E. D., and Eizenman, M. (2007). “Remote point-of-gaze estimation with free head movements requiring a single-point calibration,” in Proceedings of the 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (Lyon: IEEE), 4556–4560. doi: 10.1109/IEMBS.2007.4353353

PubMed Abstract | Crossref Full Text | Google Scholar

Hansen, D. W., Agustin, J. S., and Villanueva, A. (2010). “Homography normalization for robust gaze estimation in uncalibrated setups,” in Proceedings of the 2010 Symposium on Eye-Tracking Research & Applications (New York, NY: Association for Computing Machinery), 13–20. doi: 10.1145/1743666.1743670

Crossref Full Text | Google Scholar

Hansen, D. W., and Ji, Q. (2010). In the eye of the beholder: a survey of models for eyes and gaze. IEEE Trans. Pattern Anal. Mach. Intell. 32, 478–500. doi: 10.1109/TPAMI.2009.30

PubMed Abstract | Crossref Full Text | Google Scholar

He, J., Pham, K., Valliappan, N., Xu, P., and Navalpakkam, V. (2019). “On-device few-shot personalization for real-time gaze estimation,” in 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) (Seoul: IEEE). doi: 10.1109/ICCVW.2019.00146

Crossref Full Text | Google Scholar

Hennessey, C. A., and Lawrence, P. D. (2009). Improving the accuracy and reliability of remote system-calibration-free eye-gaze tracking. IEEE Trans. Biomed. Eng. 56, 1891–1900. doi: 10.1109/TBME.2009.2015955

PubMed Abstract | Crossref Full Text | Google Scholar

Hiroe, M., Yamamoto, M., and Nagamatsu, T. (2018). “Implicit user calibration for gaze-tracking systems using an averaged saliency map around the optical axis of the eye,” in Proceedings of the 2018 Symposium on Eye Tracking Research and Applications (Warsaw: ACM). doi: 10.1145/3204493.3204572

Crossref Full Text | Google Scholar

Hu, D., Qin, H., Liu, H., and Zhang, S. (2019). “Gaze tracking algorithm based on projective mapping correction and gaze point compensation in natural light*,” in Proceedings of the 2019 IEEE 15th International Conference on Control and Automation (ICCA) (Edinburgh: IEEE). doi: 10.1109/ICCA.2019.8899597

Crossref Full Text | Google Scholar

Hu, Z., Lv, C., Hang, P., Huang, C., and Xing, Y. (2022). Data-driven estimation of driver attention using calibration-free eye gaze and scene features. IEEE Trans. Ind. Electron. 69, 1800–1808. doi: 10.1109/TIE.2021.3057033

Crossref Full Text | Google Scholar

Huang, G., Shi, J., Xu, J., Li, J., Chen, S., Du, Y., et al. (2023). Gaze estimation by attention-induced hierarchical variational auto-encoder. IEEE Trans. Cybern. 1–14. doi: 10.1109/TCYB.2023.3312392. Available online at: https://ieeexplore.ieee.org/document/10256055

PubMed Abstract | Crossref Full Text | Google Scholar

Jen, C.-L., Chen, Y.-L., Lin, Y.-J., Lee, C.-H., Tsai, A., Li, M.-T., et al. (2016). “Vision based wearable eye-gaze tracking system,” in 2016 IEEE International Conference on Consumer Electronics (ICCE) (Las Vegas, NV: IEEE), 202–203. doi: 10.1109/ICCE.2016.7430580

Crossref Full Text | Google Scholar

Kellnhofer, P., Recasens, A., Stent, S., Matusik, W., and Torralba, A. (2019). “Gaze360: physically unconstrained gaze estimation in the wild,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) (Seoul: IEEE), 6912–6921. doi: 10.1109/ICCV.2019.00701

Crossref Full Text | Google Scholar

Krafka, K., Khosla, A., Kellnhofer, P., Kannan, H., Bhandarkar, S., Matusik, W., et al. (2016). “Eye tracking for everyone,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Las Vegas, NV: IEEE), 2176–2184. doi: 10.1109/CVPR.2016.239

Crossref Full Text | Google Scholar

Lai, C. C., Shih, S. W., and Hung, Y. P. (2015). Hybrid method for 3-D gaze tracking using glint and contour features. IEEE Trans. Circ. Syst. Video Technol. 25, 24–37. doi: 10.1109/TCSVT.2014.2329362

Crossref Full Text | Google Scholar

Li, J., Chen, Z., Zhong, Y., Lam, H.-K., Han, J., Ouyang, G., et al. (2022). Appearance-based gaze estimation for asd diagnosis. IEEE Trans. Cybern. 52, 6504–6517. doi: 10.1109/TCYB.2022.3165063

PubMed Abstract | Crossref Full Text | Google Scholar

Li, J., and Li, S. (2016). Gaze estimation from color image based on the eye model with known head pose. IEEE Trans. Hum.-Mach. Syst. 46, 414–423. doi: 10.1109/THMS.2015.2477507

Crossref Full Text | Google Scholar

Li, N., and Busso, C. (2018). Calibration free, user-independent gaze estimation with tensor analysis. Image Vis. Comput. 74, 10–20. doi: 10.1016/j.imavis.2018.04.001

Crossref Full Text | Google Scholar

Lidegaard, M., Hansen, D. W., and Krüger, N. (2014). “Head mounted device for point-of-gaze estimation in three dimensions,” in Proceedings of the Symposium on Eye Tracking Research and Applications (New York, NY: Association for Computing Machinery), 83–86. doi: 10.1145/2578153.2578163

Crossref Full Text | Google Scholar

Lindén, E., Sjöstrand, J., and Proutiere, A. (2019). “Learning to personalize in appearance-based gaze tracking,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) (Seoul: IEEE).

Google Scholar

Liu, G., Yu, Y., Mora, K. A. F., and Odobez, J. M. (2021). A differential approach for gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 43, 1092–1099. doi: 10.1109/TPAMI.2019.2957373

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, J., Chi, J., Hu, W., and Wang, Z. (2021). 3D model-based gaze tracking via iris features with a single camera and a single light source. IEEE Trans. Hum.-Mach. Syst. 51, 75–86. doi: 10.1109/THMS.2020.3035176

Crossref Full Text | Google Scholar

Liu, J., Chi, J., Lu, N., Yang, Z., and Wang, Z. (2020). Iris feature-based 3-D gaze estimation method using a one-camera-one-light-source system. IEEE Trans. Instrum. Meas. 69, 4940–4954. doi: 10.1109/TIM.2019.2956612

Crossref Full Text | Google Scholar

Liu, M., Li, Y., and Liu, H. (2020). 3D gaze estimation for head-mounted eye tracking system with auto-calibration method. IEEE Access 8, 104207–104215. doi: 10.1109/ACCESS.2020.2999633

Crossref Full Text | Google Scholar

Liu, S., Liu, D., and Wu, H. (2020). “Gaze estimation with multi-scale channel and spatial attention,” in Proceedings of the 2020 9th International Conference on Computing and Pattern Recognition (New York, NY: ACM), 303–309. doi: 10.1145/3436369.3437438

Crossref Full Text | Google Scholar

Luo, K., Jia, X., Xiao, H., Liu, D., and Han, P. (2020). A new gaze estimation method based on homography transformation derived from geometric relationship. Appl. Sci. 10:9079. doi: 10.3390/app10249079

Crossref Full Text | Google Scholar

Ma, C., Baek, S. J., Choi, K. A., and Ko, S. J. (2014). Improved remote gaze estimation using corneal reflection-adaptive geometric transforms. Opt. Eng. 53:053112. doi: 10.1117/1.OE.53.5.053112

Crossref Full Text | Google Scholar

Mansouryar, M., Steil, J., Sugano, Y., and Bulling, A. (2016). “3D gaze estimation from 2D pupil positions on monocular head-mounted eye trackers,” in Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research & Applications (New York, NY: Association for Computing Machinery), 197–200. doi: 10.1145/2857491.2857530

Crossref Full Text | Google Scholar

Mestre, C., Gautier, J., and Pujol, J. (2018). Robust eye tracking based on multiple corneal reflections for clinical applications. J. Biomed. Opt. 23:035001. doi: 10.1117/1.JBO.23.3.035001

PubMed Abstract | Crossref Full Text | Google Scholar

Model, D., and Eizenman, M. (2010). An automatic personal calibration procedure for advanced gaze estimation systems. IEEE Trans. Biomed. Eng. 57, 1031–1039. doi: 10.1109/TBME.2009.2039351

PubMed Abstract | Crossref Full Text | Google Scholar

Morimoto, C. H., Amir, A., and Flickner, M. (2002). “Detecting eye position and gaze from a single camera and 2 light sources,” in 2002 International Conference on Pattern Recognition (Quebec City, QC: IEEE), 314–317. doi: 10.1109/ICPR.2002.1047459

Crossref Full Text | Google Scholar

Morimoto, C. H., Coutinho, F. L., and Dan, W. H. (2020). Screen-light decomposition framework for point-of-gaze estimation using a single uncalibrated camera and multiple light sources. J. Math. Imaging Vis. 62, 585–605. doi: 10.1007/s10851-020-00947-8

Crossref Full Text | Google Scholar

Morimoto, C. H., and Mimica, M. R. (2005). Eye gaze tracking techniques for interactive applications. Comput. Vis. Image Underst. 98, 4–24. doi: 10.1016/j.cviu.2004.07.010

Crossref Full Text | Google Scholar

Murthy, L. R. D., and Biswas, P. (2021). “Appearance-based gaze estimation using attention and difference mechanism,” in Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (Nashville, TN: IEEE), 3137–3146.

Google Scholar

Nagamatsu, T., Hiroe, M., and Arai, H. (2021). Extending the measurement angle of a gaze estimation method using an eye model expressed by a revolution about the optical axis of the eye. IEICE Trans. Inf. Syst. E104-D, 729–740. doi: 10.1587/transinf.2020EDP7072

Crossref Full Text | Google Scholar

Nagamatsu, T., Kamahara, J., and Tanaka, N. (2008). “3D gaze tracking with easy calibration using stereo cameras for robot and human communication,” in Proceedings of the 17th IEEE International Symposium on Robot and Human Interactive Communication (Munich: IEEE, Universität München).

Google Scholar

Nagamatsu, T., Sugano, R., Iwamoto, Y., Kamahara, J., and Tanaka, N. (2010). “User-calibration-free gaze tracking with estimation of the horizontal angles between the visual and the optical axes of both eyes,” in Proceedings of the 2010 Symposium on Eye-Tracking Research and Applications (Austin, TX: ACM), 251–254. doi: 10.1145/1743666.1743725

Crossref Full Text | Google Scholar

Nagamatsu, T., Sugano, R., Iwamoto, Y., Kamahara, J., and Tanaka, N. (2011). User-calibration-free gaze estimation method using a binocular 3D eye model. IEICE Trans. Inf. Syst. 94, 1817–1829. doi: 10.1587/transinf.E94.D.1817

Crossref Full Text | Google Scholar

O'Reilly, J., Khan, A. S., Li, Z., Cai, J., Hu, X., Chen, M., et al. (2019). “A novel remote eye gaze tracking system using line illumination sources,” in 2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR) (San Jose, CA: IEEE), 449–454. doi: 10.1109/MIPR.2019.00090

Crossref Full Text | Google Scholar

Park, S., Zhang, X., Bulling, A., and Hilliges, O. (2018). “Learning to find eye region landmarks for remote gaze estimation in unconstrained settings,” in Proceedings of the 2018 ACM Symposium on Eye Tracking Research & Applications (New York, NY: Association for Computing Machinery). doi: 10.1145/3204493.3204545

Crossref Full Text | Google Scholar

Rattarom, S., Aunsri, N., and Uttama, S. (2017). “A framework for polynomial model with head pose in low cost gaze estimation,” in 2017 International Conference on Digital Arts, Media and Technology (ICDAMT) (Chiang Mai: IEEE), 24–27. doi: 10.1109/ICDAMT.2017.7904927

Crossref Full Text | Google Scholar

Ren, Z., Fang, F., Hou, G., Li, Z., and Niu, R. (2023). Appearance-based gaze estimation with feature fusion of multi-level information elements. J. Comput. Des. Eng. 10, 1080–1109. doi: 10.1093/jcde/qwad038

Crossref Full Text | Google Scholar

Sasaki, M., Nagamatsu, T., and Takemura, K. (2018). “Cross-ratio based gaze estimation using polarization camera system,” in Proceedings of the 2018 ACM International Conference on Interactive Surfaces and Spaces (New York, NY: ACM), 333–338. doi: 10.1145/3279778.3279909

Crossref Full Text | Google Scholar

Sasaki, M., Nagamatsu, T., and Takemura, K. (2019). “Screen corner detection using polarization camera for cross-ratio based gaze estimation,” in Proceedings of the 11th ACM Symposium on Eye Tracking Research & Applications (New York, NY: Association for Computing Machinery). doi: 10.1145/3314111.3319814

Crossref Full Text | Google Scholar

Shih, S.-W., Wu, Y.-T., and Liu, J. (2000). “A calibration-free gaze tracking technique,” in Proceedings of the 15th International Conference on Pattern Recognition (Barcelona: IEEE).

Google Scholar

Shin, Y. G., Choi, K. A., Kim, S. T., and Ko, S. J. (2015). A novel single IR light based gaze estimation method using virtual glints. IEEE Trans. Consumer Electron. 61, 254–260. doi: 10.1109/TCE.2015.7150601

Crossref Full Text | Google Scholar

Sigut, J., and Sidha, S. A. (2011). Iris center corneal reflection method for gaze tracking using visible light. IEEE Trans. Biomed. Eng. 58, 411–419. doi: 10.1109/TBME.2010.2087330

PubMed Abstract | Crossref Full Text | Google Scholar

Sugano, Y., Matsushita, Y., and Sato, Y. (2014). “Learning-by-synthesis for appearance-based 3D gaze estimation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Columbus, OH: IEEE), 1821–1828. doi: 10.1109/CVPR.2014.235

Crossref Full Text | Google Scholar

Sugano, Y., Matsushita, Y., Sato, Y., and Koike, H. (2015). Appearance-based gaze estimation with online calibration from mouse operations. IEEE Trans. Hum.-Mach. Syst. 45, 750–760. doi: 10.1109/THMS.2015.2400434

Crossref Full Text | Google Scholar

Sun, L., Liu, Z., and Sun, M. T. (2015). Real time gaze estimation with a consumer depth camera. Inf. Sci. 320, 346–360. doi: 10.1016/j.ins.2015.02.004

Crossref Full Text | Google Scholar

Sun, L., Song, M., Liu, Z., and Sun, M. T. (2014). Real-time gaze estimation with online calibration. IEEE Multimedia 21, 28–37. doi: 10.1109/MMUL.2014.54

Crossref Full Text | Google Scholar

Uhm, K.-H., Kang, M.-C., Kim, J.-Y., and Ko, S.-J. (2020). Improving the robustness of gaze tracking under unconstrained illumination conditions. Multimed. Tools Appl. 79, 20603–20616. doi: 10.1007/s11042-020-08679-y

Crossref Full Text | Google Scholar

Villanueva, A., and Cabeza, R. (2008). A novel gaze estimation system with one calibration point. IEEE Trans. Syst. Man. Cybern. B Cybern. 38, 1123–1138. doi: 10.1109/TSMCB.2008.926606

PubMed Abstract | Crossref Full Text | Google Scholar

Wan, Z., Xiong, C., Chen, W., Zhang, H., and Wu, S. (2022). Pupil-contour-based gaze estimation with real pupil axes for head-mounted eye tracking. IEEE Trans. Ind. Inform. 18, 3640–3650. doi: 10.1109/TII.2021.3118022

Crossref Full Text | Google Scholar

Wang, K., and Ji, Q. (2016). “Real time eye gaze tracking with kinect,” in 2016 23rd International Conference on Pattern Recognition (ICPR) (Cancun: IEEE), 2752–2757. doi: 10.1109/ICPR.2016.7900052

Crossref Full Text | Google Scholar

Wang, K., and Ji, Q. (2018). 3D gaze estimation without explicit personal calibration. Pattern Recognit. 79, 216–227. doi: 10.1016/j.patcog.2018.01.031

Crossref Full Text | Google Scholar

Wang, K., Wang, S., and Ji, Q. (2016). “Deep eye fixation map learning for calibration-free eye gaze tracking,” in Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research and Applications (Charleston, SC: ACM), 47–55. doi: 10.1145/2857491.2857515

Crossref Full Text | Google Scholar

Wang, Q., Wang, H., Dang, R.-C., Zhu, G.-P., Pi, H.-F., Shic, F., et al. (2023). Style transformed synthetic images for real world gaze estimation by using residual neural network with embedded personal identities. Appl. Intell. 53, 2026–2041. doi: 10.1007/s10489-022-03481-9

Crossref Full Text | Google Scholar

Wen, Q., Bradley, D., Beeler, T., Park, S., and Xu, F. (2020). Accurate real-time 3D gaze tracking using a lightweight eyeball calibration. Comput. Graphi. Forum 39, 475–485. doi: 10.1111/cgf.13945

Crossref Full Text | Google Scholar

Wen, Q., Xu, F., Lu, M., and Yong, J.-H. (2017a). Real-time 3D eyelids tracking from semantic edges. ACM Trans. Graph. 36:193. doi: 10.1145/3130800.3130837

Crossref Full Text | Google Scholar

Wen, Q., Xu, F., and Yong, J.-H. (2017b). Real-time 3D eye performance reconstruction for rgbd cameras. IEEE Trans. Vis. Comput. Graph. 23, 2586–2598. doi: 10.1109/TVCG.2016.2641442

PubMed Abstract | Crossref Full Text | Google Scholar

Wood, E., Baltruaitis, T., Zhang, X., Sugano, Y., Robinson, P., Bulling, A., et al. (2015). “Rendering of eyes for eye-shape registration and gaze estimation,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV) (Santiago: IEEE), 3756–3764. doi: 10.1109/ICCV.2015.428

Crossref Full Text | Google Scholar

Wood, E., Baltrušaitis, T., Morency, L.-P., Robinson, P., and Bulling, A. (2016). “Learning an appearance-based gaze estimator from one million synthesised images,” in Proceedings of the Ninth Biennial ACM Symposium on Eye Tracking Research and Applications (New York, NY: Association for Computing Machinery), 131–138. doi: 10.1145/2857491.2857492

Crossref Full Text | Google Scholar

Wu, Y., Li, G., Liu, Z., Huang, M., and Wang, Y. (2022). Gaze estimation via modulation-based adaptive network with auxiliary self-learning. IEEE Trans. Circ. Syst. Video Technol. 32, 5510–5520. doi: 10.1109/TCSVT.2022.3152800

Crossref Full Text | Google Scholar

Xia, L., Sheng, B., Wu, W., Ma, L., and Li, P. (2016). Accurate gaze tracking from single camera using gabor corner detector. Multimed. Tools Appl. 75, 221–239. doi: 10.1007/s11042-014-2288-4

Crossref Full Text | Google Scholar

Xiong, X., Cai, Q., Liu, Z., and Zhang, Z. (2014). “Eye gaze tracking using an rgbd camera: a comparison with an rgb solution,” in Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct Publication (New York, NY: ACM), 1113–1121. doi: 10.1145/2638728.2641694

Crossref Full Text | Google Scholar

Yan, B., Yu, S., Pei, T., and Hu, Y. (2017). “Vision interaction method based on visual attention mechanism,” in 2017 12th IEEE Conference on Industrial Electronics and Applications (ICIEA) (Siem Reap: IEEE), 930–935. doi: 10.1109/ICIEA.2017.8282972

Crossref Full Text | Google Scholar

Yoon, H. S., Hong, H. G., Lee, D. E., and Park, K. R. (2019). Driver's eye-based gaze tracking system by one-point calibration. Multimed. Tools Appl. 78, 7155–7179. doi: 10.1007/s11042-018-6490-7

Crossref Full Text | Google Scholar

Yuan, G., Wang, Y., Yan, H., and Fu, X. (2022). Self-calibrated driver gaze estimation via gaze pattern learning. Knowl.-Based Syst. 235:107630. doi: 10.1016/j.knosys.2021.107630

Crossref Full Text | Google Scholar

Zhang, M., Liu, Y., and Lu, F. (2022). “Gazeonce: real-time multi-person gaze estimation,” in Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (New Orleans, LA: IEEE). doi: 10.1109/CVPR52688.2022.00416

Crossref Full Text | Google Scholar

Zhang, X., Sugano, Y., Bulling, A., and Hilliges, O. (2020). “Learning-based region selection for end-to-end gaze estimation,” in The 31st British Machine Vision Virtual Conference (Manchester). Available online at: https://www.bmvc2020-conference.com

Google Scholar

Zhang, X., Sugano, Y., Fritz, M., and Bulling, A. (2015). “Appearance-based gaze estimation in the wild,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Boston, MA IEEE), 4511–4520. doi: 10.1109/CVPR.2015.7299081

Crossref Full Text | Google Scholar

Zhang, X., Sugano, Y., Fritz, M., and Bulling, A. (2019). Mpiigaze: real-world dataset and deep appearance-based gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 41, 162–175. doi: 10.1109/TPAMI.2017.2778103

PubMed Abstract | Crossref Full Text | Google Scholar

Zhao, X., Huang, Y., Tian, Y., and Tian, M. (2022). Episode-based personalization network for gaze estimation without calibration. Neurocomputing 513, 36–45. doi: 10.1016/j.neucom.2022.09.050

Crossref Full Text | Google Scholar

Zhou, X., Cai, H., Li, Y., and Liu, H. (2017). “Two-eye model-based gaze estimation from a kinect sensor,” in IEEE International Conference on Robotics and Automation (ICRA) (Singapore: IEEE), 1646–1653. doi: 10.1109/ICRA.2017.7989194

Crossref Full Text | Google Scholar

Zhou, X., Cai, H., Shao, Z., Yu, H., and Liu, H. (2016). “3D eye model-based gaze estimation from a depth sensor,” in 2016 IEEE International Conference on Robotics and Biomimetics (ROBIO) (Qingdao: IEEE), 369–374. doi: 10.1109/ROBIO.2016.7866350

Crossref Full Text | Google Scholar

Zhu, Z., and Ji, Q. (2008). Novel eye gaze tracking techniques under natural head movement. IEEE Trans. Biomed. Eng. 54, 2246–2260. doi: 10.1109/TBME.2007.895750

PubMed Abstract | Crossref Full Text | Google Scholar

Zhu, Z., Zhang, D., Chi, C., Li, M., and Lee, D.-J. (2023). A complementary dual-branch network for appearance-based gaze estimation from low-resolution facial image. IEEE Trans. Cogn. Dev. Syst. 15, 1323–1334. doi: 10.1109/TCDS.2022.3210219

Crossref Full Text | Google Scholar

Keywords: personal calibration, video-oculographic, gaze tracking, eye-movement interaction, visual information

Citation: Liu J, Chi J and Yang Z (2024) A review on personal calibration issues for video-oculographic-based gaze tracking. Front. Psychol. 15:1309047. doi: 10.3389/fpsyg.2024.1309047

Received: 07 October 2023; Accepted: 08 March 2024;
Published: 20 March 2024.

Edited by:

Hantao Liu, Cardiff University, United Kingdom

Reviewed by:

Zhongxu Hu, Nanyang Technological University, Singapore
Yang Cao, University of Science and Technology of China, China

Copyright © 2024 Liu, Chi and Yang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jiannan Chi, dXN0YmpuYyYjeDAwMDQwO3VzdGIuZWR1LmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.