- 1Computer Science, Old Dominion University, Norfolk, VA, United States
- 2Physical Therapy, Movement & Rehabilitation Sciences, Northeastern University, Boston, MA, United States
- 3Art + Design, Northeastern University, Boston, MA, United States
- 4Institute for Neural Computation, UC San Diego, San Diego, CA, United States
Our subjective visual experiences involve complex interaction between our eyes, our brain, and the surrounding world. It gives us the sense of sight, color, stereopsis, distance, pattern recognition, motor coordination, and more. The increasing ubiquity of gaze-aware technology brings with it the ability to track gaze and pupil measures with varying degrees of fidelity. With this in mind, a review that considers the various gaze measures becomes increasingly relevant, especially considering our ability to make sense of these signals given different spatio-temporal sampling capacities. In this paper, we selectively review prior work on eye movements and pupil measures. We first describe the main oculomotor events studied in the literature, and their characteristics exploited by different measures. Next, we review various eye movement and pupil measures from prior literature. Finally, we discuss our observations based on applications of these measures, the benefits and practical challenges involving these measures, and our recommendations on future eye-tracking research directions.
1 Introduction
The five primary senses provide humans with a rich perceptual experience of the world, with vision as the dominant sense. Early studies of visual perception (Dodge, 1900; Buswell, 1935; Yarbus, 1967) and its physiological underpinnings (Hubel and Wiesel, 1979; Hubel, 1995), have provided a foundation for subtler and more sophisticated studies of the visual system and its dynamic interaction with the environment via the oculomotor system. The oculomotor system both maintains visual stability and controls gaze-orienting movements (Goldberg et al., 1991; Land and Furneaux, 1997). It is comprised of the efferent limb of the visual system and the vestibular system (Wade and Jones, 1997). The efferent limb is responsible for maintaining eye position and executing eye movements. The vestibular system, on the other hand, is responsible for providing our brain with information about motion, head position, and spatial orientation, which, in turn, facilitates motor functions, such as balance, stability during movement, and posture (Goldberg and Fernandez, 1984; Wade and Jones, 1997; Day and Fitzpatrick, 2005). The sense of hearing or touch also affects eye movements (Eberhard et al., 1995; Maier and Groh, 2009). There are five distinct types of eye movement, two gaze-stabilizing movements: vestibulo-ocular (VOR), opto-kinetic nystagmus (OKN); and three gaze-orienting movements: saccadic, smooth pursuit, and vergence (Duchowski, 2017; Hejtmancik et al., 2017). For the purposes of this review, we will focus on gaze-orienting eye movements that place the high-resolution fovea on selected objects of interest.
The existence of the fovea, a specialized high-acuity region of the central retina approximately 1–2 mm in diameter (Dodge, 1903), provides exceptionally detailed input in a small region of the visual field (Koster, 1895), approximately the size of a quarter held at arm’s length (Pumphrey, 1948; Hejtmancik et al., 2017). The role of gaze-orienting movements are to direct the fovea toward objects of interest. Our subjective perception of a stable world with uniform clarity is a marvel resulting from our visual and oculomotor systems working together seamlessly, allowing us to engage with a complex and dynamic environment.
Recent advancements in computing such as computer vision (Krafka et al., 2016; Lee et al., 2020) and image processing (Pan et al., 2017; Mahanama et al., 2020; Ansari et al., 2021) have led to the development of computing hardware and software that can extract these oculomotor events and measurable properties. This include eye tracking devices range from commodity hardware (Mania et al., 2021) capable of extracting few measures to reserach-grade eye trackers combined with sophisticated software capable of extracting various advance measures. As a result, eye movement and pupillometry have the potential for wide adoption for both in applications and research. There is a need for an aggregate body of knowledge on eye movement and pupillometry measures to provide, a.) a taxonomy of measures linking various oculomotor events, and b.) a quick reference guide for eye tracking and pupillometry measures. For application oriented literature of the eye tracking, interested reader is referred to (Duchowski, 2002) for a breadth-first survey of eye tracking applications.
In this paper, we review a selection of relevant prior research on gaze-orienting eye movements, the periods of visual stability between these movements, and pupil measures to address the aforementioned issue (see Figure 1). First, we describe the main oculomotor events and their measurable properties, and introduce common eye movement analysis methods. Next, we review various eye movement and pupil measures. Next, we discuss the applications of aforementioned measures in domains including, but not limited to, neuroscience, human-computer interaction, and psychology, and analyze their strengths and weaknesses. The paper concludes with a discussion on applications, recent developments, limitations, and practical challenges involving these measures, and our recommendations on future eye-tracking research directions.
2 Oculomotor Events
This section describes oculomotor events that function as the basis for several eye movement and pupil measures. These events are: 1) fixations and saccades, 2) smooth pursuit, 3) fixational eye movements (tremors, microsaccades, drifts), 4) blinks, and 5) ocular vergence.
2.1 Fixations and Saccadic Eye Movements
Eye movement information can be interpreted as a sequence of fixations and saccades. A fixation is a period where our visual gaze remains at a particular location. A saccade, on the other hand, is a rapid eye movement between two consecutive fixations. Typical humans perform 3–5 saccades per second, but this rate varies with current perceptual and cognitive demands (Fischer and Weber, 1993). Fixations and saccades are the primary means of interacting with and perceiving the visual world. During a fixation, our visual perceptual processes unfold. Saccades guide our fovea to selected regions of the visual field. We are effectively blind during saccades (Burr et al., 1994), which allows our gaze to remain relatively stable during saccadic reorientation. Saccadic eye movements are brief, and have a reliable amplitude-velocity relationship (see Figure 2) known as the main sequence (Bahill et al., 1975; Termsarasab et al., 2015). It shows that saccade velocity and saccade amplitude follow a linear relationship, up to 15°–20°. This relationship, however, varies with age and also in certain disorders (Choi et al., 2014; Reppert et al., 2015).
FIGURE 2. Saccade amplitude and peak velocity (A), main sequence (B) (Reppert et al., 2015).
Saccades are inhibited during engaged visual attention on stationary stimuli, and as a result, a (nearly) steady central fixation is obtained (Fischer and Breitmeyer, 1987). Previous studies (Schiller et al., 1980; Wang et al., 2015) have shown that saccade preparation processes can be analyzed via pupil size. In particular, Schiller et al. (1980) showed that distinct neural preparatory signals in Superior Colliculus (SC) and Frontal Eye Field (FEF) are vital for saccade preparation, and Wang et al. (2015) showed that the SC is associated with the pupil control circuit. Cortical processing is associated with saccade latency, with shorter latency indicating advanced motor preparation (Connolly et al., 2005). Reiterating this point, Jainta et al. (2011) showed a negative correlation between saccade latency and pupil size prior to a saccade. Thus, the analysis of measures such as pupil diameter during fixations, fixation duration, saccade rate, saccade accuracy, and saccade latency, provide important cumulative clues to understanding the underlying deployment of visual attention.
2.1.1 Identifying Fixations and Saccades
There exists several eye tracking technologies (Young and Sheena, 1975) that measure ocular features over time and transform them into a stream of gaze positions. These streams can be analyzed in different ways to identify periods of fixation and saccades. Salvucci and Goldberg (2000) describe five algorithms for such identification: Velocity Threshold Identification (I-VT), Hidden Markov Model Identification (I- HMM), Dispersion Threshold Identification (I-DT), Minimum Spanning Tree Identification (I-MST), and Area-of-Interest Identification (I-AOI). I-VT and I-HMM are velocity-based algorithms. In I-VT, consecutive points are identified as fixations or saccades, based on their point-to-point velocities (Findlay et al., 1995). I-HMM, on the other hand, uses a two-state Hidden Markov Model with hidden states representing the velocity distributions of saccade and fixation points. Compared to fixed-threshold methods like I-VT, I-HMM performs a more robust identification (Salvucci and Goldberg, 2000), since it employs a probabilistic model rather than a fixed velocity threshold, which allows more freedom in identifying points. I-DT and I-MST are dispersion-based algorithms that use a moving window to calculate the dispersion of points. Based on whether the dispersion is above or below the threshold, points are classified as fixations or saccades (Widdel, 1984). In I-MST, a minimum-spanning tree is constructed from gaze points. Edges with lengths exceeding a predefined ratio are labeled as saccades and clusters of points connected by saccades are labeled as fixations (Salvucci and Goldberg, 2000). I-AOI is an area-based algorithm that only identifies fixations within specified target areas (AOIs) (Salvucci and Goldberg, 2000). If a point falls within a target area, it is labeled as a fixation point, and if not, it is labeled as a saccade point. Consecutive fixation points are then grouped together, and groups that do not span a minimum duration are re-labeled as saccade points. A systematic evaluation of the performance of these algorithms are available at Komogortsev et al. (2010).
2.2 Smooth Pursuit Eye Movements
The smooth pursuit system is a different gaze-orienting movement that is deployed to keep a moving object in foveal vision (Carl and Gellman, 1987; Barnes, 2008). Smooth pursuit eye movements are generally made when tracking an object moving in the visual environment (Carl and Gellman, 1987; Barnes and Asselman, 1991). A typical smooth pursuit movement is usually initiated by an saccadic eye movement to orient to the tracked object. The pursuit system subsequently matches the eye velocity to target velocity (Robinson, 1965; Barnes and Asselman, 1991). This smooth movement is punctuated by additional saccadic movements that eliminate retinal error between the current gaze position and target. The smooth pursuit system has a functional architecture very similar to that of the saccadic system (Lisberger et al., 1987); however, smooth pursuit has a lower latency (100–125 ms) than saccades (200–250 ms) (Meyer et al., 1985; Krauzlis, 2004). Due to the underlying similarity between saccades and smooth pursuits, metrics used to characterize saccades could also be used to characterize smooth pursuit behavior (Lisberger et al., 1987). According to a classic smooth pursuit behavior model Robinson et al. (1986), there are three aspects of pursuit to characterize: onset, offset and motor learning. Pursuit onset is the response time of the pursuit system to a target which moves for a certain period of time. Since it occurs after the target starts to move, this time delay reflects the response of the pursuit system to the target motion while the eyes were still (Jiang, 1996). Pursuit offset is the response time of the pursuit system to turn off as a response when target stops its motion. When the target is seen to stop, the pursuit system is turned off and replaced by fixation (Robinson et al., 1986). Motor plasticity or motor learning, as a gradual process that makes small, adaptive steps in a consistent direction, was also incorporated and simulated in this model.
2.3 Fixational Eye Movements
The process of visual exploration is characterized by alternating fixations and saccades (Salvucci and Goldberg, 2000). However, a fixation does not imply a stationary eye; our eyes are continually in (involuntary) motion, even during periods of fixation (Adler and Fliegelman, 1934; Ratliff and Riggs, 1950; Ditchburn and Ginsborg, 1953). These fixational eye movements fall into one of three classes: 1) tremors, 2) micro-saccades, and 3) drifts (Martinez-Conde et al., 2004; Rucci and Poletti, 2015).
2.3.1 Tremor
A tremor (or ocular micro-tremor, or physiological nystagmus) is an aperiodic, wavelike eye movement with a high-frequency (
2.3.2 Microsaccade
A microsaccade (or flick, or flicker, or fixational saccade) is a small, fast, jerk–like eye movement that occurs during voluntary fixation (Ditchburn and Ginsborg, 1953; Martinez-Conde et al., 2004). They occur at a typical rate of 1–3 Hz, shifting the line of sight abruptly by a small amount (Ditchburn and Ginsborg, 1953). The average size of a microsaccade is about 6′ arc (i.e., the size of a thumb-tack head, held 2.5 m away from the eye) (Steinman et al., 1973). The dynamics of microsaccades vary with stimuli and viewing task. For instance, the difficulty of a task can be discerned by the number of microsaccades that occurred (Otero-Millan et al., 2008), and their magnitude (Krejtz et al., 2018). Microsaccades also have comparable spatio-temporal properties as saccades (Zuber et al., 1965; Otero-Millan et al., 2008). For instance, microsaccades lie on the saccadic main sequence (Zuber et al., 1965). The refractory periods between saccades and microsaccades are also equivalent (Otero-Millan et al., 2008). Moreover, microsaccades as small as 9′ generate a field potential over the occipital cortex and the mid-central scalp sites 100–140 ms after movement onset, which resembles the visual lambda response evoked by saccades (Dimigen et al., 2009). It is increasingly accepted that microsaccades play an important role in modulating attentional and perceptual processes (Hafed et al., 2015).
2.3.3 Drift
A drift (or slow drift) is a low-frequency eye movement that occurs during the intervals between microsaccades and saccades (Steinman et al., 1973). During a drift, the retinal image of a fixated object moves across photoreceptors (Ratliff and Riggs, 1950). Drifts have a compensatory role in maintaining accurate visual fixation; they occur either in the absence of microsaccades, or when the compensation by microsaccades is inadequate (Ratliff and Riggs, 1950). The average size of a drift is about a 6′ arc, with an average velocity of about a 1′ arc/sec (Ratliff and Riggs, 1950; Ditchburn and Ginsborg, 1953; Cornsweet, 1956; Nachmias, 1961).
2.4 Blinks
A blink is essentially the closing and reopening of the eyelids. Blinks are primitive, yet widely used, in eye tracking measures. When the blink originates from a voluntary action, the blink becomes a voluntary blink or a wink (Blount, 1927). In the case of non-voluntary blinks, they are of two types: spontaneous blinks and reflexive blinks. For reflexive blinks, external stimuli evoke reflexive blinks as a form of protection, while any involuntary blink not belonging to any of these categories is a spontaneous blink (Valls-Sole, 2019). The winks or voluntary blinks are not commonly adopted as a metric despite the usage as a form of interaction (Noronha et al., 2017). In contrast, involuntary blinks indicate the state of an individual (Stern et al., 1984) or a reflex action to a stimulus (Valls-Sole, 2019). Between involuntary blinks, spontaneous blinks are the most common type of blink used as a metric due to their correlation with one’s internal state (Shin et al., 2015; Maffei and Angrilli, 2019).
2.5 Ocular Vergence
Up until this point, the movements described are all referred to as conjugate or “yoked” eye movements, meaning that the eyes move in the same direction to fixate an object. Fortunately, we can choose to fixate on objects in different depth planes, during which binocular vision is maintained by opposite movements of the two eyes. These simultaneously directly opposing movements of the eyes result in Ocular vergence (Holmqvist et al., 2011). These vergence movements can occur in either direction, resulting in convergence or divergence. Far-to-near focus triggers convergent movements and near-to-far focus triggers divergent movements. The ubiquitous use of screen-based eye tracking results in more literature related to conjugate eye movements in a single depth plane.
3 Eye Movement Analysis
Eye movements are a result of complex cognitive processes, involving at the very least, target selection, movement planning, and execution. Analysis of eye movements (see Table 1 for a list of eye movement measures) can reveal objective and quantifiable information about the quality, predictability, and consistency of these covert processes (Van der Stigchel et al., 2007). Holmqvist et al. (2011) and Young and Sheena (1975) discuss several eye movement measures, eye movement measurement techniques, and key considerations for eye movement research. In this section, we introduce several eye movement analysis techniques from the literature.
3.1 Area of Interest Analysis
AOI analysis is a technique to analyze eye movements by assigning them to specific areas (or regions) of the visual scene (Holmqvist et al., 2011; Hessels et al., 2016). In contrast to obtaining eye movement measures across the entire scene, AOI analysis provides semantically localized eye movement measures that are particularly useful for attention-based research (e.g., User Interaction research, Marketing research, and Psychology research) (Hessels et al., 2016). In AOI analysis, defining the shape and bounds of an AOI can be difficult (Hessels et al., 2016). Ideally, each AOI should be defined with the same shape and bounds as the actual object. However, due to practical limitations, such as the difficulty of defining arbitrarily shaped (and sized) AOIs in eye-tracking software, AOIs are most commonly defined using simple shapes (rectangles, ellipses, etc.) (Holmqvist et al., 2011). Recent advancements in computer vision have given rise to models that automatically and reliably identify real world objects in visual scenes Ren et al. (2015), Redmon et al. (2016). This makes it possible to identify AOIs in real world images almost as readily as with pre-defined stimuli presented on a computer screen Jayawardena and Jayarathna (2021), Zhang et al. (2018).
3.2 Heat Map Analysis
Heat map analysis is a technique for analyzing the spatial distribution of eye movements across the visual scene. This technique can be used to analyze eye movements of individual participants, as well as aggregated eye movements of multiple participants. In general, heat maps are represented using Gaussian Mixture Models (GMMs) that indicate the frequency (or probability) of fixation localization. Heat map-based metrics generally involve a measure of overlap between two GMMs, indicating similarity of fixated regions of an image. In heat map analysis, the order of visitation is not captured (Grindinger et al., 2010); rather, it analyzes the spatial distribution of fixations. When visualizing heat maps, a color-coded, spatial distribution of fixations is overlaid on the stimuli that participants looked at. The color represents the quantity of fixations at each point on the heat map. Heat map analysis is particularly useful when analyzing the areas of the stimuli that participants paid more (or less) visual attention to, for example in driving research comparing different groups and conditions (Snider et al., 2021).
3.3 Scan Path Analysis
A scan path is a sequence of fixations and saccades that describe the pattern of eye movements during a task (Salvucci and Goldberg, 2000). Scan paths could appear quite complex, with frequent revisits and overlapping saccades. In general, scan paths are visualized as a sequence of connected nodes (a fixation centroid) and edges (a saccade between two successive fixations) displayed over the visual image of the scene (Heminghous and Duchowski, 2006; Goldberg and Helfman, 2010). Here, the diameter of each node is proportional to the fixation duration (Goldberg and Helfman, 2010). Scan path analyses have been widely used to model the dynamics of eye movement during visual search (Walker-Smith et al., 1977; Horley et al., 2003). It has also been used in areas like biometric identification (Holland and Komogortsev, 2011). In cognitive neuroscience, Parkhurst et al. (2002) and van Zoest et al. (2004) show that stimulus-driven, bottom-up attention dominates during the early phases of viewing. On the contrary, Nyström and Holmqvist (2008) show that top-down cognitive processes guide fixation selection throughout the course of viewing. They discovered that viewers eventually fixate on meaningful stimuli, regardless of whether that stimuli was obscured or reduced in contrast.
4 Eye Movement Measures
In this section, we discuss several metrics relevant to oculomotor behavior (Komogortsev et al., 2013) which are derived from fixations, saccades, smooth pursuit, blinks, vergence, and visual search paradigm.
4.1 Fixation Measures
Fixation-based measures are widely used in eye-tracking research. Here, fixations are first identified using algorithms such as I-VT, I-HMM (velocity-based), I-DT, I-MST (dispersion-based), and I-AOI (area-based) (Salvucci and Goldberg, 2000). This information is then used to obtain different fixation measures, as described below.
4.1.1 Count
Fixation count is the number of fixations identified within a given time period. Fixations can be counted either over the entire stimuli or within a single AOI (Holmqvist et al., 2011). Fixation count has been used to determine semantic importance (Buswell, 1935; Yarbus, 1967), the efficiency and difficulty in search (Goldberg and Kotval, 1999; Jacob and Karn, 2003), neurological dysfunctions (Brutten and Janssen, 1979; Coeckelbergh et al., 2002), and the impact of prior experience (Schoonahd et al., 1973; Megaw, 1979; Megaw and Richardson, 1979).
4.1.2 Duration
Fixation duration indicates a time period where the eyes stay still in one position (Salvucci and Goldberg, 2000). In general, fixation durations are around 200–300 ms long, and longer fixations indicate deeper cognitive processing (Rayner, 1978; Salthouse and Ellis, 1980). Also, fixations could last for several seconds (Young and Sheena, 1975; Karsh and Breitenbach, 1983) or be as short as 30–40 ms (Rayner, 1978; Rayner, 1979). Furthermore, the distribution of fixation duration is typically positively skewed, rather than Gaussian (Velichkovsky et al., 2000; Staub and Benatar, 2013). The average fixation duration is often used as a baseline to compare with fixation duration data at different levels (Salthouse and Ellis, 1980; Pavlović and Jensen, 2009). By comparing average fixation duration across AOIs, one could distinguish areas that were looked at for longer durations than others. In particular, if certain AOIs were looked at longer than others, then their average fixation duration would be higher (Goldberg and Kotval, 1999; Pavlović and Jensen, 2009).
4.2 Saccade Measures
4.2.1 Amplitude
The amplitude of a saccade (see Figure 2) is the distance travelled by a saccade during an eye movement. It is measured either by visual degrees (angular distance) or pixels, and can be approximated via the Euclidean distance between fixation points (Megaw and Richardson, 1979; Holmqvist et al., 2011). Saccade amplitudes are dependent on the nature of the visual task. For instance, in reading tasks, saccade amplitudes are limited to ≈ 2°, i.e., 7–8 letters in a standard font size (Rayner, 1978). They are further limited when oral reading is involved (Rayner et al., 2012). Furthermore, saccade amplitudes tend to decrease with increasing task difficulty (Zelinsky and Sheinberg, 1997; Phillips and Edelman, 2008), and with increasing cognitive load (May et al., 1990; Ceder, 1977).
4.2.2 Direction
The direction (or orientation, or trajectory) of a saccade or sequence of saccades is another useful descriptive measure. It can be represented as either an absolute value, a relative value, or a discretized value. Absolute saccade direction is calculated using the coordinates of consecutive fixations. Relative saccade direction is calculated using the difference of absolute saccade direction of two consecutive saccades. Discretized saccade directions are obtained by binning the absolute saccade direction into pre-defined angular segments (e.g., compass-like directions). In visual search studies, researchers have used different representations of saccade direction to analyze how visual conditions affect eye movement behavior. For instance, absolute saccade directions were used to analyze the effect of target predictability (Walker et al., 2006) and visual orientation (Foulsham et al., 2008) on eye movements. Similarly, discretized saccade directions were used to compare and contrast the visual search strategies followed by different subjects (Ponsoda et al., 1995; Gbadamosi and Zangemeister, 2001).
4.2.3 Velocity
Saccade velocity is calculated by taking the first derivative of time series of gaze position data. Average saccadic velocity is the average of velocities over the duration of a saccade. Peak saccadic velocity is the highest velocity reached during a saccade (Holmqvist et al., 2011). For a particular amplitude, saccade velocity has been found to decrease with tiredness (Becker and Fuchs, 1969; McGregor and Stern, 1996), sleep deprivation (Russo et al., 2003), and conditions such as Alzheimer’s (Boxer et al., 2012) and AIDS (Castello et al., 1998). In contrast, saccade velocity has been found to increase with increasing task difficulty (Galley, 1993), increasing intrinsic value of visual information (Xu-Wilson et al., 2009), and increasing task experience (McGregor and Stern, 1996). Many studies on neurological and behavioral effects of drugs and alcohol (Lehtinen et al., 1979; Griffiths et al., 1984; Abel and Hertle, 1988) have used peak saccade velocity as an oculomotor measure.
4.2.4 Latency
Saccadic latency measures the duration between the onset of a stimulus and the initiation of the saccade (Andersson et al., 2010). In practice, the measurement of the saccadic latency is affected by two main factors: The sampling frequency of the setup, and the saccade detection time. The sampling frequency refers to the operational frequency of the eye tracker, where the operational frequency negatively correlates with the introduced error. The saccade detection time is when the device detects a saccade by arriving at the qualifying velocity or the criteria in the saccade detection algorithm. In a study with young and older participants, researchers observed the saccadic latencies to increase significantly in older participants with the decrease in the stimulus size Warren et al. (2013). Further, in comparative studies between healthy subjects and subjects with Parkinson’s disease with and without medication, saccadic latency shows potential as a biomarker for the disease (Michell et al., 2006). A similar study with participants having amblyopia has found the interocular difference between saccadic latencies to correlate with the difference in Snellen acuity (McKee et al., 2016).
4.2.5 Rate
Saccade rate (or saccade frequency) is the number of saccadic eye movements per unit time (Ohtani, 1971). For static stimuli, the saccade rate is similar to the fixation rate. For dynamically moving stimuli, however, the saccade rate is a measure of catch-up saccades generated during smooth pursuit (Holmqvist et al., 2011). The saccade rate decreases with increasing task difficulty (Nakayama et al., 2002) and fatigue level (Van Orden et al., 2000). Moreover, subjects with neurological disorders exhibit higher saccadic rates during smooth pursuit (O’Driscoll and Callahan, 2008; van Tricht et al., 2010).
4.2.6 Gain
Saccade gain (or saccade accuracy) is the ratio between the initial saccade amplitude and the target amplitude (i.e., Euclidean distance between the two stimuli among which that saccade occurred) (Coëffé and O’regan, 1987). This measure indicates how accurately a saccadic movement landed on the target stimuli (Ettinger et al., 2002; Holmqvist et al., 2011). When the gain of a particular saccade is greater than 1.0, that saccade is called an overshoot, or hypermetric, and when it is less than 1.0, that saccade is called an undershoot, or hypometric. Saccadic gain is probabilistic at a per-individual level; Lisi et al. (2019) demonstrates that biased saccadic gains are an individualized probabilistic control strategy that adapts to different environmental conditions. Saccade gain is commonly used in neurological studies (Ettinger et al., 2002; Holmqvist et al., 2011). For instance, Crevits et al. (2003) used saccade gain to quantify the effects of severe sleep deprivation.
4.3 Smooth Pursuit Measures
4.3.1 Direction
Smooth pursuit direction (or smooth pursuit trajectory) indicates the direction of smooth pursuit movement as the eyes follow a moving stimulus (Holmqvist et al., 2011). The ability to pursue a moving object varies with its direction of motion. Rottach et al. (1996) showed that smooth pursuit gain is higher during horizontal pursuit than during vertical pursuit. This difference in gain can be attributed to most real-world objects naturally being in horizontal rather than vertical motion (Collewijn and Tamminga, 1984).
4.3.2 Velocity
Smooth pursuit velocity is first moment of gaze positions during a smooth pursuit. Compared to saccade velocity, the velocity of smooth pursuits is low, typically around 20°/s–40°/s (Young, 1971). However, when participants are specifically trained to follow moving stimuli, or are provided with accelerating stimuli, higher peak smooth pursuit velocities were observed (Meyer et al., 1985). For example, Barmack (1970) observed peak smooth pursuit velocities of 100°/s in typical participants, when provided with accelerating stimuli. However, Bahill and Laritz. (1984) observed peak smooth pursuit velocities of 130°/s on trained baseball players. As the velocity of moving stimuli increases, the frequency of catch-up saccades increases to compensate for retinal offset (De Brouwer et al., 2002).
4.3.3 Acceleration
The second moment of the gaze position trace provides the acceleration of a smooth pursuit eye movement. This acceleration is maintained until the eye velocity (smooth pursuit velocity) matches the visual target’s velocity (Kao and Morrow, 1994). Examination of acceleration is typically a part of determining smooth pursuit onset (see Figure 3). Smooth pursuit acceleration has been used to analyze how visual cues (Ladda et al., 2007) and prior knowledge of a visual target’s trajectory (Kao and Morrow, 1994) impact eye movements. Smooth pursuit acceleration is higher when a visual target’s motion was unpredictable.
FIGURE 3. Eye velocity (plotted line) and visual target velocity (dotted line) over time (x-axis), aligned to pursuit onset (0 ms). Smooth pursuit acceleration seen from 0 to 140 ms. Smooth pursuit latency seen from origin to 0 ms (Spering and Gegenfurtner, 2007).
4.3.4 Latency
Smooth pursuit latency is the delay between when a target object starts to move (i.e., target onset) and when the pursuit begins (i.e., smooth pursuit onset) (Holmqvist et al., 2011). When the direction and velocity of the target object are not predictable, the smooth pursuit latency varies between 100–200 ms (Burke and Barnes, 2006). In paradigms such as step-ramp (Rashbass, 1961) that allow for anticipation, smooth pursuit latency may drop to 0 ms (or less, if pursuit starts before target motion) when its direction and velocity are predictable (Burke and Barnes, 2006; de Hemptinne et al., 2006). If the luminance of the moving object is the same as the background, the smooth pursuit latency may be prolonged by
4.3.5 Retinal Position Error
Fixations are maintained more accurately on stationary targets rather than moving targets. During smooth pursuit, both the eye and the target are in motion, and lag between their positions is expected (Dell’Osso et al., 1992). This error is known as retinal position error, and is formally defined as the difference between eye and target positions measured during fixations.
4.3.6 Gain
Smooth pursuit gain (or smooth pursuit accuracy) is the ratio between smooth pursuit velocity and the target velocity (Zackon and Sharpe, 1987; Holmqvist et al., 2011). Typically, the smooth pursuit gain is lower than 1.0 and tends to fall even lower when the target velocity is high (Zackon and Sharpe, 1987). Moreover, smooth pursuit gain is modulated by on-line gain control (Robinson, 1965; Churchland and Lisberger, 2002). Smooth pursuit gain is decreased during conditions that distract user attention (Březinová and Kendell, 1977). Furthermore, smooth pursuit gain is also higher when tracking horizontal motion, compared to tracking vertical motion Rottach et al. (1996). Smooth pursuit gain has also been used in neurologically research. For instance, O’Driscoll and Callahan (2008) analyzed the smooth pursuit gain of individuals diagnosed with schizophrenia, and observed a low smooth pursuit gain in these populations. The smooth pursuit gain can be quantified by the root mean squared distance between the target point and the gaze position (θ) over the span of the experiment of n data samples.
4.4 Blink Measures
4.4.1 Rate
Blink rate (or spontaneous blink rate, or blink frequency) is typically measured in blinks per minute. In some studies, the time between blinks (or blink interval) is measured instead (Shin et al., 2015). Early studies (Peterson and Allison, 1931; Newhall, 1932) show that blink rate is subjected to factors such as lighting, time of the day (fatigue), temperature, wind, age, and sex. Moreover, while blinks are predominantly involuntary, they are inhibited during engaged visual attention to minimize any blink-induced interruption to visual information (Ranti et al., 2020). More recent studies explore the idea of standard spontaneous eye blink rate in a broader aspect; healthy and non-healthy individuals under single (Doughty, 2002; Doughty and Naase, 2006; Ranti et al., 2020) and multiple (Doughty, 2001; Doughty, 2019) experiment conditions. Their results demonstrate the lack of a common value for blink rate, as they are dependent on experimental conditions. Thus, a standard measure for blink rate remains illusive despite the vast body of studies. Many studies on cognition adopt eye blink rate as a metric due its relative ease of detection, and its ability to indicate aspects of one’s internal state. In a study on the impact of blink rate on a Stroop task, researchers found that blink rate increases during a Stroop task compared with a baseline task of resting (Oh et al., 2012). Further, a study involving movie watching found the blink rate a reliable biomarker for assessing the concentration level (Shin et al., 2015; Maffei and Angrilli, 2019). Studies in dopamine processes also use blink rate as a marker of dopamine function. A review on studies in cognitive dopamine functions and the relationship with blink rate indicates varying results on blink rate in dopaminergic studies, including primates (Jongkees and Colzato, 2016).
4.4.2 Amplitude
The blink amplitude is the measure of the distance traveled by (the downward distance of upper eyelid) in the event of a blink (Stevenson et al., 1986). The amplitude measures the relative distance of motion to the distance the eyelid travels in a complete blink. This measure can be obtained using a video-based eye tracker (Riggs et al., 1987) or using electrooculography (Morris and Miller, 1996). Blink amplitude is often used in conjunction with research associated with fatigue measurement (Morris and Miller, 1996; Galley et al., 2004) and task difficulty (Cardona et al., 2011; Chu et al., 2014). Cardona et al. (2011) assessed the characteristics of blink behavior during visual tasks requiring prolonged periods of demanding activities. During the experiment, the authors noted a larger percentage of incomplete blinks. Another study involving fatigued pilots in a flight simulator showed that blink amplitude served as a promising predictor for level of fatigue (Morris and Miller, 1996).
4.5 Visual Search Measures
Visual search behavior combines instances of saccades, fixations and possibly also smooth pursuit. The methods used to analyze such sequences of behavior are described below.
4.5 1 Scan Path Similarity
Vector and string-based editing approaches have been developed to compute the similarity of scan paths (Jarodzka et al., 2010; De Bruin et al., 2013). In particular, De Bruin et al. (2013) introduced three metrics: 1) Saccade Length Index (SLI), 2) Saccade Deviation Index (SDI)—to assist in faster analysis of eye tracking data, and 3) Benchmark Deviation Vectors (BDV)—to highlight repetitive path deviation in eye tracking data. The SLI is the sum of the distance of all the saccades during the experiment. The SLI can be obtained through following equation, where s is the starting position of the saccade, e ending position of the saccade, and n is the total number of saccades.
Jarodzka et al. (2010), on the other hand, proposed representing scan paths as geometrical vectors, and simplifying scan paths by clustering consecutive saccades directed at most Tϕ radians apart.
4.5.2 Time-to-First-Fixation on AOI
Time-to-First-Fixation on AOI (or time to first hit) refers to the time taken from stimulus onset up to the first fixation into a particular AOI (Holmqvist et al., 2011). This measure may be useful for both bottom-up stimulus-driven searches (e.g., a flashy company label) as well as top-down attention driven searches (e.g., when respondents actively decide to focus on certain elements or aspects on a website or picture). This metric is particularly useful for user interface evaluation, as a measure of visual search efficiency (Jacob and Karn, 2003). For instance, it has been used to evaluate the visual search efficiency of web pages (Ellis et al., 1998; Bojko, 2006). It is also influenced by prior knowledge, such as domain expertise. For instance, studies show that when analyzing medical images, expert radiologists exhibit a lower Time to First Fixation on AOIs (lesions, tumors, etc.) than novice radiologists (Krupinski, 1996; Venjakob et al., 2012; Donovan and Litchfield, 2013).
4.5.3 Revisit Count
Revisit (or re-fixation, or recheck) count indicates how often the gaze was returned to a particular AOI. It can be used to distinguish between the AOIs that were frequently revisited, and the AOIs that were less so. A participant may be drawn back to a particular AOI for different reasons, such as its semantic importance (Guo et al., 2016), to refresh the memory (Meghanathan et al., 2019), and for confirmatory purposes (Mello-Thoms et al., 2005). The emotion perceived through visual stimuli also affect the likelihood of subsequent revisits, and thereby, the revisit count (Motoki et al., 2021); yet this perceived emotion is difficult to interpret purely through revisit count. Revisits are particularly common in social scenes, where observers look back and forth between interacting characters to assess their interaction (Birmingham et al., 2008). Overall, revisit count is indicative of user interest towards an AOI, and can be used to optimize user experiences.
4.5.4 Dwell Time
Dwell time is the interval between one’s gaze entering an AOI and subsequently exiting it (Holmqvist et al., 2011). This includes the time spent on all fixations, saccades, and revisits during that visit (Tullis and Albert, 2013). For a typical English reading task, a lower dwell time (e.g.,
4.5.5 Gaze Transition Matrix
The gaze transition matrix is an adaptation of the transition matrix of Markov models into eye movement analysis. In a Markov model, a transition matrix (or probability matrix, or stochastic matrix, or substitution matrix) is a square matrix, where each entry represents the transition probability from one state to another. This concept was first applied for eye movement analysis by Ponsoda et al. (1995) to model the transition of saccade direction during visual search. Similarly, Bednarik et al. (2005) used this concept to model the transition of gaze position among AOIs and study the correlation between task performance and search pattern. The gaze transition matrix is calculated using the number of transitions from the Xth AOI to the Yth AOI. Based on the gaze transition matrix, several measures have been introduced to quantify different aspects of visual search.
4.5.6 Transition Matrix Density
Goldberg and Kotval (1999) defined transition matrix density (i.e., fraction of non-zero entries in the transition matrix) in order to analyze the efficiency of visual search. Here, a lower transition matrix density indicates an efficient and directed search, whereas a dense transition matrix indicates a random search.
4.5.7 Gaze Transition Probability
When comparing gaze transition matrices, Vandeberg et al. (2013) performed an element-wise comparison of transition probabilities by modelling eye movement transitions as a multi-level Hidden Markov Model. Similarly, Jayawardena et al. (2020) utilized gaze transition matrices to analyze the probabilities of transition of gaze between four AOIs. Figure 4A shows the AOIs used in this study. Figure 4B shows a sample scanpath of a participant, and Figure 4C shows its corresponding gaze transition matrix. Here, each cell in the matrix represents the probability of gaze transition from one AOI to another.
FIGURE 4. AOIs, Scanpaths, and Gaze Transition Matrices from Jayawardena et al. (2020). (A) AOI on face stimuli. (B) Scanpath with fixations on the AOIs. (C) Gaze transition matrix.
4.5.8 Gaze Transition Entropy
Gaze transition entropy is a measure of predictability in AOI transitions and overall distribution of eye movements over stimuli (Krejtz et al., 2014; Krejtz et al., 2015). Compared to transition matrix density, gaze transition entropy is a histogram-based estimation, which, in effect, takes into account the AOI size. The concept of entropy used here is that of information theory; it describes the amount of information required to generate a particular sequence, as a measure of its uncertainty. Krejtz et al. (2014), Krejtz et al. (2015) computes gaze transition entropy by first modelling eye movements between AOIs as a first-order Markov chain, and then obtaining its Shannon’s entropy (Shannon, 1948). They obtain two forms of entropy: 1) transition entropy Ht (calculated for individual subjects’ transition matrices) and 2) stationary entropy Hs (calculated for individual subjects’ stationary distributions).
Ht indicates the predictability of gaze transitions; a high Ht (∀ij|pij → 0.5) implies low predictability, whereas a low Ht (∀ij|pij → {0, 1}) implies high predictability. It is calculated by normalizing the transition matrix row-wise, replacing zero-sum rows by the uniform transition probability 1/s (Here, s is the number of AOIs), and obtaining its Shannon’s entropy. Hs, on the other hand, indicates the distribution of visual attention (Krejtz et al., 2014); a high Hs indicates that visual attention is equally distributed across all AOIs, whereas a low Hs indicates that visual attention is directed towards certain AOIs. It is estimated via eigen-analysis, and the Markov chain is assumed to be in a steady state where the transition probabilities converge. In one study, Krejtz et al. (2015) showed that participants who viewed artwork with a reportedly high curiosity, yielded a significantly less Ht (i.e., more predictable transitions) than others. Moreover, participants who viewed artwork with a reportedly high appreciation, yielded a significantly less Hs (i.e., more directed visual attention) than others. In another study, Krejtz et al. (2015) showed that participants who reportedly recognized a given artwork, yielded a significantly higher Ht and Hs (i.e., less predictable gaze) than others. Jayawardena et al. (2020) performed entropy-based eye movement analysis on neurotypical and ADHD-diagnosed subjects during an audiovisual speech-in-noise task. They found that ADHD-diagnosed participants made unpredictable gaze transitions (i.e., high entropy) at different levels of task difficulty, whereas the neurotypical group of participants made gaze transitions from any AOI to the mouth region (i.e., low entropy) regardless of task difficulty. These findings suggest that Ht and Hs are potential indicators of curiosity, interest, picture familiarity, and task difficulty.
4.6 Vergence Measures
Due to the association of ocular vergence with binocular vision, the binocular gaze data can be used to measure ocular vergence. The gaze vergence can be estimated (see Figure 5) using the distance between the individual gaze positions for each eye, the distance from the user to the screen, and the interocular distance (Daugherty et al., 2010; Wang et al., 2012). A common application of the ocular vergence is assessing stereoscopic perception (Essig et al., 2004; Daugherty et al., 2010; Wang et al., 2012).
FIGURE 5. Ocular vergence (Daugherty et al., 2010; Wang et al., 2012).
These studies use ocular vergence to assess the user’s depth perception under different stimuli conditions. Further, the ocular vergence also has been the subject of estimating 3D gaze positions (Mlot et al., 2016) based on 2D positions provided by eye trackers.
The perceived depth (Δd) can be computed using the distance between gaze positions (Δx), the distance between the left and right eye (a), and the distance between the user and the screen (D′).
5 Pupil Measures
Pupil measures (see Table 1 for a list of pupil measures) capture fluctuations in the pupil’s size and orientation to produce measurements that provide insights into one’s internal state. The pupil diameter is the result of tonic and phasic pupillary responses of the eye (Sun et al., 1983; Wass et al., 2015; Peysakhovich et al., 2017).
The tonic component refers to the pupil diameter changes caused by slow contractions, while the phasic refers to the quick or transient contractions. Most of the metrics use pupil dilation as the primary measure for the computations in order to determine the tonic and phasic components.
5.1 Pupil Diameter
The most primitive metric used in pupillometry is the average pupil diameter (Gray et al., 1993; Joshi et al., 2016). The measure captures both the tonic and the phasic components of the pupil dilation. Using the average dilation, some of the phasic and transient features can get smoothed out. An alternative measure to overcome the drawback of the average pupil dilation is to use the change in pupil dilation relative to a baseline. The baseline is assumed to correspond to the tonic component, while the relative change refers to the phasic component. The baseline can be either derived during the study or through a controlled environment.
5.2 Index of Cognitive Activity
Index of Cognitive Activity (ICA), introduced by Marshall (2002), is a measure of pupil diameter fluctuation as an instantaneous measure. Marshall (2000), Marshall (2007) describes the methodology followed in computing the ICA in an experiment setting. Furthermore, the publications also include parameter selections when performing experiments. The process starts by eliminating the pupil signal regions corresponding to blinks by either removing them or replacing them through interpolation as the preprocessing step. The signal is then passed through wavelet decomposition to capture the pupil signal’s abrupt changes through decomposing to the desired level. Finally, the decomposed signal subjects to thresholding, converting the decomposed signal coefficients to a binary vector of the same size; the thresholding stage acts as a de-noising stage here.
5.3 Index of Pupillary Activity
Index of Pupillary activity (IPA), introduced by Duchowski et al. (2018), is a metric inspired by ICA, with a similar underlying concept. Studies on ICA do not fully disclose the internals of ICA due to intellectual property reasons. IPA, however, discloses the internals of its process.
IPA computation starts by discarding pupil signals around blinks identified in the experiment. Duchowski et al. (2018) used a window of 200 ms in either direction during experiments. The procedure for computing the IPA measurement starts with a two-level Symlet-16 discrete wavelet decomposition of the pupil dilation signal by selecting a mother wavelet function ψj,k(t). The resulting dyadic wavelet upon the wavelet analysis of the signal x(t), generates a dyadic series representation. Then the process follows a multi-resolution signal analysis of the original signal x(t). A level is arbitrarily selected from the multi-resolution decomposition to produce a smoother approximation of the signal x(t). Finally, threshold the wavelet modulus maxima coefficients using a universal threshold defined by,
where
5.4 Low/High Index of Pupillary Activity
Low/High Index of Pupillary Activity (LHIPA), introduced by Duchowski et al. (2020), is a variation of the IPA measure identified earlier. The computation of the metric remains the same as in IPA, except for counting remnants. Instead of counting threshold remnants as in IPA (Duchowski et al., 2018), LHIPA counts the modulus maxima of the low and high-frequency bands contained in the wavelet (see Figure 6).
FIGURE 6. Visualization of the processed pupil diameter signal (yellow) and the threshold used for calculating the Low/High Index of Pupillary Activity (green) (Duchowski et al., 2020).
Using the Discrete Wavelet Transform to analyze the pupil diameter signal at multiple levels of resolution, the wavelet coefficients are found by:
where gk is the one-dimensional high-pass wavelet filter. Level j is chosen arbitrarily to select either high or low frequency wavelet coefficients. For the high frequency component, they have chosen j = 1 and for the low frequency component, they have chosen j = 1/2 log 2(n), the mid-level frequency octave where log 2(n) is the number of octaves. Thus, the low frequency/high frequency ratio is:
The pupillary response increases as the cognitive load increases. Since LHIPA is a ratio, an increase in the pupillary response reflects a decrease in the LHIPA reading. The authors demonstrate the metric’s applicability using a series of three experiments where they assess the relationship between the task difficulty and the corresponding measures of IPA and LHIPA. During the experiments, the authors determined that the LHIPA was able to identify between difficult tasks and easy or baseline tasks throughout the study, while IPA could do so during only one experiment. The experiments also revealed that the LHIPA demonstrated cognitive load earlier than IPA, indicating a faster response due to the measure being a ratio and a built-in ratio arising from the ratio’s computation.
6 Discussion
Eye movement and pupil measures have been applied in disciplines including, but not limited to, neuroscience (Hessels and Hooge, 2019), psychology (Mele and Federici, 2012), and human computer interaction (Duchowski, 2002). In this section, we collectively summarize the literature of eye movement and pupillometry applications under different domains, discuss the limitations that we observe, and by doing so, establish a vision for implementing eye movements and pupil measures in these application environments.
6.1 Applications
6.1.1 Neuroscience
In eye-tracking neuroscience, researchers have jointly analyzed neuronal activity and oculomotor activity to study the physiological organization of the vision system, and their effect on cognition and behavior. Studies of neuronal activity during phenomena such as attention (Blair et al., 2009; Kimble et al., 2010), scene perception (Duc et al., 2008), inattentional blindness (Simons and Chabris, 1999), visual engagement (Catherine and James, 2001), and covert attention processing (i.e., visually fixating on one location while diverting attention to another) (Posner et al., 1980) had revealed important facts about cognition and behavior.
For instance, studies on covert attention processing show that attention cannot be inferred solely from whether an object was looked at (Hafed and Clark, 2002; Ebitz and Moore, 2019), and doing so would lead to false positives (Posner et al., 1980). Similarly, studies on scene perception show that information is processed in two forms: top-down (based on semantic importance) and bottom-up (based on visual significance such as color, brightness, etc.) (Duc et al., 2008). Moreover, studies on visual attention show that attention is divided among AOIs through a sustained cognitive state, from which relevant visual objects become available to influence behavior Duncan et al. (1994).
Studies have also revealed that pupillary activity is correlated with cognitive load (Hess and Polt, 1964; Hyönä et al., 1995), and also with neural gain (Eldar et al., 2013) and cortical activity (Reimer et al., 2016). While cognitive load can be inferred from pupillary activity, studies show that such inference becomes challenging in fast-paced cognitive tasks (Wierda et al., 2012), temporally overlapping cognitive tasks (Wierda et al., 2012), and in surroundings with varying ambient luminance (Zénon, 2017). Cherng et al. (2020) showed that pupil diameter is regulated by sympathetic activation (arousal-induced pupil dilation) and parasympathetic inhibition (saccade-induced pupil dilation), both of which are affected by ambient luminance. Wierda et al. (2012) showed that a deconvolved pupil response signal is indicative of cognitive load, with a high temporal resolution. However, (Zénon, 2017), showed that this method does not account for low-frequency pupil fluctuations and inter-individual variability of pupil responses, and instead proposed using auto-regressive models (Cho et al., 2018) with exogenous inputs to analyze pupillary activity. Similarly, Watson and Yellott (2012) proposes a generalized formula to analyze pupillary activity, which accounts for ambient luminance, the size of the adapting field, the age of the observer, and whether both pupils are adapted.
Researchers of eye tracking neurscience have utilized various measures to study neurodevelopmental disorders. Some of these studies are exclusively based on fixation measures. For instance, He et al. (2019) examined the fixation count and fixation duration in preschoolers with Autism Spectrum Disorder (ASD), and found that preschoolers with ASD had atypical gaze patterns in a facial emotion expression task compared to typical individuals. Moreover, they found that deficits in recognizing emotions from facial expressions in ASD correlated with the social interaction and development quotient. In contrast, some studies are based exclusively on saccadic measures. For instance, MacAskill et al. (2002) found that the saccade amplitude was modified during adaption of memory-guided saccades and this adaptive ability was impaired in individuals with Parkinson’s disease. Similarly, Barbosa et al. (2019) identfied that saccadic direction errors are associated with impulsive compulsive behaviors of individuals with Parkinson’s Disease, as they had difficulty in suppressing automatic saccades to a given target. Patel et al. (2012) showed that saccade latency is correlated with the severity of Huntington Disease (HD), and suggested its possiblity of being a biomarker of disease severity in HD. In another study, Jensen et al. (2019) found that reduced saccade velocity is a key indicator of progressive supranuclear palsy and other disoders of mid-brain. Similarly, Biscaldi et al. (1998) found that individuals with Dyslexia had significantly higher regressive saccade rate in a sequential-target task. Similarly, Termsarasab et al. (2015) stated that abnormalities in saccade gain could aid diagnosis of hyperkinetic and hypokinetic movement disorders. Similarly, Fukushima et al. (2013) used a memory-based smooth pursuit task to examine working memory of smooth pursuit direction in individuals with PD.
Certain studies performed their analysis using visual search measures. Rutherford and Towns (2008), for instance, demonstrated scan path similarities and differences during emotion perception between typical individuals and individuals with ASD. Similarly, Bours et al. (2018) demonstrated that individuals with ASD had increased time to first fixation on the eyes of fearful faces during emotion recognition task. Duret et al. (1999) showed that refixation strategies in macular disorders was dependant on location of the target relative to the scotoma, spatial characteristics of the disease and the duration of the disorder. Guillon et al. (2015) showed that gaze transition matrices could potentially reveal new strategies of visual scanning followed by individuals with ASD. Moreover, Wainstein et al. (2017) showed that pupil diameter could be a biomarker in ADHD based on the results from a visuo-spatial working memory task.
6.1.2 Human Computer Interaction
In HCI research, eye tracking has been used to evaluate the usability of human-computer interfaces (both hardware and software). Here, the primary eye movement measures being analyzed are saccades, fixations, smooth pursuits, compensatory, vergence, micro-saccades, and nystagmus (Goldberg and Wichansky, 2003). For instance, Farbos et al. (2000) and Dobson (1977) had attempted to correlate eye tracking measures with software usability metrics such as time taken, completion rate, and other global metrics. Similarly, Du and MacDonald (2014) had analyzed how visual saliency is affected by icon size. In both scenarios, eye movements were recorded while users navigated human-computer interfaces, and were subsequently analyzed using fixation and scan-path measures.
Pupil diameter measures are widely used to assess the cognitive load of users when interacting with human-computer interfaces. For instance, Bailey and Iqbal (2008) used pupil diameter measures to demonstrate that cognitive load varies while completing a goal-directed task. Adamczyk and Bailey (2004) used pupil diameter measures to demonstrate that the disruption caused by user interface interruptions (e.g., notifications) is more pronounced during periods of high cognitive load. Iqbal et al. (2004) used pupil diameter measures to show that difficult tasks demand longer processing time, induces higher subjective ratings of cognitive load, and reliably evokes greater pupillary response at salient subtasks. In addition to pupil diameter measures, studies such as Chen and Epps (2014) have also utilized blink rate measures to analyze how cognitive load (Sweller, 2011) and perceptual load (Macdonald and Lavie, 2011) varies across different tasks. They claim that pupil diameter indicates cognitive load well for tasks with low perceptual load, and that blink rate indicates perceptual load better than cognitive load.
6.1.3 Psychology
In psychology research, eye tracking has been used to understand eye movements and visual cognition during naturalistic interactions such as reading, driving, and speaking. Rayner (2012), for instance, have analyzed eye movements during English reading, and observed a mean saccade duration and amplitude of 200–250 ms and 7–9 letters, respectively. They have also observed different eye movement patterns when reading silently vs aloud, and correlations of text complexity to both fixation duration (+) and saccade length (−). Three paradigms were commonly used to analyze eye movements during reading tasks: moving window—selecting a few characters before/after the fixated word (McConkie and Rayner, 1975), foveal mask—masking a region around the fixated word (Bertera and Rayner, 2000), and boundary—creating pre-defined boundaries (i.e., AOIs) to classify fixations (Rayner, 1975). Over time, these paradigms have been extended to tasks such as scene perception. Recarte and Nunes (2000) and Stapel et al. (2020), for instance, have used the boundary method to analyze eye movements during driving tasks and thereby assess the effect of driver awareness on gaze behavior. In particular, Recarte and Nunes (2000) have observed that distracted drivers had higher fixation durations, higher pupil dilations, and lower fixation counts in the mirror/speedometer AOIs compared to control subjects.
In psycholinguistic research, pupil measures have been used to analyze the cognitive load of participants in tasks such as simultaneous interpretation (Russell, 2005), speech shadowing (Marslen-Wilson, 1985), and lexical translation. For instance, Seeber and Kerzel (2012) measured the pupil diameter during simultaneous interpretation tasks, and observed a larger average pupil diameter (indicating a higher cognitive load) when translating between verb-final and verb-initial languages, compared to translating between languages of the same type. Moreover, Hyönä et al. (1995) measured the pupil diameter during simultaneous interpretation, speech shadowing, and lexical translation tasks. They observed a larger average pupil diameter (indicating a higher cognitive load) during simultaneous interpretation than speech shadowing, and also momentary variations in pupil diameter (corresponding to spikes in cognitive load) during lexical translation.
Moreover, studies in marketing and behavioral finance, have used eye-tracking and pupillometric measures to understand the relationship between presented information and the decision-making process. For instance, Rubaltelli et al. (2016) used pupil diameter to understand investor decision-making, while Ceravolo et al. (2019), Hüsser and Wirth (2014) used dwell time. Further, studies in marketing have identified relationships between the consumer decision process through fixations on different sections in the product description Ares et al. (2013), Menon et al. (2016). The utility of eye-tracking in marketing extends beyond product descriptions, to advertisements, brands, choice, and search patterns (Wedel and Pieters, 2008).
6.2 Recent Developments
Among the recent developments in eye tracking and pupil measures, the introduction of a series of pupillometry-based measures (ICA, IPA, and LHIPA) (Marshall, 2002; Duchowski et al., 2018; Duchowski et al., 2020) to assess the cognitive load is noteworthy. These measures, in general, compute the cognitive load by processing the pupil dilation as a signal, and thus require frequent and precise measurements of pupil dilation to function. We ascribe the success of these measures to the technological advancements in pupillometric devices, which led to higher sampling rates and more accurate measurement of pupil dilation than possible before. In the future, we anticipate the continued development of measures that exploit pupil dilation signals (El Haj and Moustafa, 2021; Maier and Grueschow, 2021).
Another noteworthy development is the emergence of commodity camera-based eye-tracking (Sewell and Komogortsev, 2010; Krafka et al., 2016; Semmelmann and Weigelt, 2018; Mahanama et al., 2020) systems to further democratize eye-tracking research and interaction. Since these systems require no dedicated/specialized hardware, one could explore eye-tracking at a significantly lower cost than otherwise possible. Yet the lack of specialized hardware, such as IR illumination or capture, restricts their capability to only eye tracking, and not pupillometry. The relatively low sampling rates of commodity cameras may negatively affect the quality of eye-tracking measures. Overall, eye tracking on commodity hardware provides a cost-effective means of incorporating other eye-tracking measures, despite its quality being heavily device-dependent. In the future, we anticipate cameras to have higher sampling frequencies (Wang et al., 2021) and resolutions, which, in turn, would bridge the gap between specialized eye trackers and commodity camera-based eye tracking systems.
6.3 Limitations
One of the major limitation we observed is that most studies use derivative measures of only a single oculomotor event instead of combinations of multiple events. For instance, studies that use fixational eye movements generally use only measures associated with fixations, despite the possible utility of saccadic information. This limiation is exacerbated with the confusion on the concepts of fixations and saccades (Hessels et al., 2018). Further, most studies rely on first-order statistical features (e.g., histogram-based features such as min, max, mean, median, sd, and skewness) or trend analysis (e.g., trajectory-based features such as sharpest decrease and sharpest increase between consecutive samples) on the features instead of employing advanced measures. Our study identified some measures to be popular in specific domains, despite their potential applicability into other domains; for instance, pupillary measures are extensively studied in neuroscience research (Schwalm and Jubal, 2017; Schwiedrzik and Sudmann, 2020), but not quite so in psychology research. One plausible reason is the lack of literature that aggregates eye movement and pupillometric measures to assist researchers in identifying additional measures. Through this paper, we attempt to provide this missing knowledge to researchers. Another reason could be the computational limitations of eye tracking hardware and software. For instance, micro-saccadic measures often require high-frequency eye trackers ( ≥300 Hz) (Krejtz et al., 2018) that are relatively expensive, thereby imposing hardware-level restrictions. Likewise, intellectual property restrictions (Marshall, 2000) and the lack of public implementations (Duchowski et al., 2018; Duchowski et al., 2020) (i.e., code libraries) of eye tracking solutions impose software-level restrictions. Both scenarios create an entry barrier for researchers into all applicable eye movement and pupillometry measures. Even though we suggest alternatives for patented measures in this paper, the lack of code libraries still remains unaddressed.
Another limitation is the relatively unexplored research avenues of eye-tracking and pupillometric measures in Extended Reality (XR) environments (Rappa et al., 2019; Renner and Pfeiffer, 2017; Clay et al., 2019; Mutasim et al., 2020; Heilmann and Witte, 2021), such as Virtual Reality (VR), Augmented Reality (AR), and Mixed Reality (MR). These extended realities have a broad utility in simulating real-world scenarios, often eliminating the requirement of complex experimental environments (i.e., VR driver behavior (Bozkir et al., 2019)). Further, the significant control of the reality vested to the experiment designer enables them to simulate complex and rare real-world events. However, these extended reality scenarios require additional hardware and software to perform eye-tracking. This poses entry barriers for practitioners in the form of budgetary constraints (i.e., cost of additional hardware, and software), knowledge constraints (i.e., experience/knowledge on XR toolsets), and experimental design (i.e., the structure of the experiment, how to simulate and integrate modalities such as touch or haptics). We suspect these factors collectively contribute to the less exploratory studies in XR eye tracking research. Considering the possibility of this technology being more commonplace (e.g., Microsoft Hololens 21 eye tracking optics, HTC VIVE Pro Eye2, VARJO VR eye tracking3), this could be an excellent opportunity for future studies.
7 Conclusion
In this paper, we have identified, discussed, and reviewed existing measures in eye-tracking and pupillometry. Further, we classified these measures based on the mechanism of vision; as eye movement, blink, and pupil-based measures. For each, we provided an overview of the measure, the underlying eye-mechanism exploited by it, and how it can be measured. Further, we identified a selected set of studies that use each type of measure and documented their findings.
This survey aims to help the researchers in two forms. First, due to the utility of eye-tracking and pupillometric measures in a broader range of domains, the implications of measures/results in other domains can easily be overlooked. Our study helps to overcome the issue by including applications and their indications along with each measure. Further, we believe the body of knowledge in the study would help researchers to choose appropriate measures for a future study. Researchers could adapt our taxonomy to classify eye-tracking and pupil measures based on the eye mechanism and vice versa. Moreover, the researchers can identify particular eye mechanisms or measures exploited for research through the presented classification. Finally, we expect this review to serve as a reference for researchers exploring eye tracking techniques using eye movements and pupil measures.
Author Contributions
BM: Planned the structure of the paper and content, carried out the review for the survey, organized the entire survey and contributed to writing the manuscript. YJ: Planned the structure of the paper and content, carried out the review for the survey, organized the entire survey and contributed to writing the manuscript. SR: Planned the structure of the paper and content, proofread, and contributed to writing the manuscript. GJ: Contributed to writing the Pupillometry measures. LC: Contributed to proofread, and supervision JS: Contributed to proofread, and supervision SJ: Idea formation, proofread and research supervision.
Funding
This work is supported in part by the U.S. National Science Foundation grant CAREER IIS-2045523. Any opinions, findings and conclusion or recommendations expressed in this material are the author(s) and do not necessarily reflect those of the sponsors.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
We thank the reviewers whose comments/suggestions helped improve and clarify this manuscript.
Footnotes
1https://www.microsoft.com/en-us/hololens/hardware
2https://www.vive.com/us/product/vive-pro-eye/overview/
3https://varjo.com/use-center/get-to-know-your-headset/eye-tracking/
References
Abel, L. A., and Hertle, R. W. (1988). “Effects of Psychoactive Drugs on Ocular Motor Behavior,” in Neuropsychology of Eye Movements (Hillsdale, NJ: Lawrence Erlbaum Associates), 81–114.
Adamczyk, P. D., and Bailey, B. P. (2004). “If Not Now, when? the Effects of Interruption at Different Moments within Task Execution,” in Proceedings of the SIGCHI conference on Human factors in computing systems, 271–278.
Adler, F. H., and Fliegelman, M. (1934). Influence of Fixation on the Visual Acuity. Arch. Ophthalmol. 12, 475–483. doi:10.1001/archopht.1934.00830170013002
Anderson, N. C., Bischof, W. F., Foulsham, T., and Kingstone, A. (2020). Turning the (Virtual) World Around: Patterns in Saccade Direction Vary with Picture Orientation and Shape in Virtual Reality. J. Vis. 20, 21. doi:10.1167/jov.20.8.21
Andersson, R., Nyström, M., and Holmqvist, K. (2010). Sampling Frequency and Eye-Tracking Measures: How Speed Affects Durations, Latencies, and More. J. Eye Move. Res. 3. doi:10.16910/jemr.3.3.6
Ansari, M. F., Kasprowski, P., and Obetkal, M. (2021). Gaze Tracking Using an Unmodified Web Camera and Convolutional Neural Network. Appl. Sci. 11, 9068. doi:10.3390/app11199068
Anson, E. R., Bigelow, R. T., Carey, J. P., Xue, Q.-L., Studenski, S., Schubert, M. C., et al. (2016). Aging Increases Compensatory Saccade Amplitude in the Video Head Impulse Test. Front. Neurol. 7, 113. doi:10.3389/fneur.2016.00113
Ares, G., Giménez, A., Bruzzone, F., Vidal, L., Antúnez, L., and Maiche, A. (2013). Consumer Visual Processing of Food Labels: Results from an Eye-Tracking Study. J. Sens Stud. 28, 138–153. doi:10.1111/joss.12031
Bahill, A. T., Clark, M. R., and Stark, L. (1975). The Main Sequence, a Tool for Studying Human Eye Movements. Math. biosciences 24, 191–204. doi:10.1016/0025-5564(75)90075-9
Bahill, A. T., and Laritz, T. (1984). Why Can’t Batters Keep Their Eyes on the ball. Am. Scientist 72, 249–253.
Bailey, B. P., and Iqbal, S. T. (2008). Understanding Changes in Mental Workload during Execution of Goal-Directed Tasks and its Application for Interruption Management. ACM Trans. Comput.-Hum. Interact. 14, 1–28. doi:10.1145/1314683.1314689
Barbosa, P., Kaski, D., Castro, P., Lees, A. J., Warner, T. T., and Djamshidian, A. (2019). Saccadic Direction Errors Are Associated with Impulsive Compulsive Behaviours in Parkinson's Disease Patients. Jpd 9, 625–630. doi:10.3233/jpd-181460
Barmack, N. H. (1970). Modification of Eye Movements by Instantaneous Changes in the Velocity of Visual Targets. Vis. Res. 10, 1431–1441. doi:10.1016/0042-6989(70)90093-3
Barnes, G. R., and Asselman, P. T. (1991). The Mechanism of Prediction in Human Smooth Pursuit Eye Movements. J. Physiol. 439, 439–461. doi:10.1113/jphysiol.1991.sp018675
Barnes, G. R. (2008). Cognitive Processes Involved in Smooth Pursuit Eye Movements. Brain Cogn. 68, 309–326. doi:10.1016/j.bandc.2008.08.020
Bartels, M., and Marshall, S. P. (2012). “Measuring Cognitive Workload across Different Eye Tracking Hardware Platforms,” in Proceedings of the symposium on eye tracking research and applications, 161–164. doi:10.1145/2168556.2168582
Becker, W., and Fuchs, A. F. (1969). Further Properties of the Human Saccadic System: Eye Movements and Correction Saccades with and without Visual Fixation Points. Vis. Res. 9, 1247–1258. doi:10.1016/0042-6989(69)90112-6
Bednarik, R., Myller, N., Sutinen, E., and Tukiainen, M. (2005). “Applying Eye-Movememt Tracking to Program Visualization,” in 2005 IEEE Symposium on Visual Languages and Human-Centric Computing (VL/HCC’05) (IEEE), 302–304.
Bertera, J. H., and Rayner, K. (2000). Eye Movements and the Span of the Effective Stimulus in Visual Search. Perception & Psychophysics 62, 576–585. doi:10.3758/bf03212109
Birmingham, E., Bischof, W. F., and Kingstone, A. (2008). Social Attention and Real-World Scenes: The Roles of Action, Competition and Social Content. Q. J. Exp. Psychol. 61, 986–998. doi:10.1080/17470210701410375
Biscaldi, M., Gezeck, S., and Stuhr, V. (1998). Poor Saccadic Control Correlates with Dyslexia. Neuropsychologia 36, 1189–1202. doi:10.1016/s0028-3932(97)00170-x
Blair, M. R., Watson, M. R., Walshe, R. C., and Maj, F. (2009). Extremely Selective Attention: Eye-Tracking Studies of the Dynamic Allocation of Attention to Stimulus Features in Categorization. J. Exp. Psychol. Learn. Mem. Cogn. 35, 1196–1206. doi:10.1037/a0016272
Blount, W. P. (1927). Studies of the Movements of the Eyelids of Animals: Blinking. Exp. Physiol. 18, 111–125. doi:10.1113/expphysiol.1927.sp000426
Bojko, A. (2006). Using Eye Tracking to Compare Web page Designs: A Case Study. J. Usability Stud. 1, 112–120.
Bours, C. C. A. H., Bakker-Huvenaars, M. J., Tramper, J., Bielczyk, N., Scheepers, F., Nijhof, K. S., et al. (2018). Emotional Face Recognition in Male Adolescents with Autism Spectrum Disorder or Disruptive Behavior Disorder: an Eye-Tracking Study. Eur. Child. Adolesc. Psychiatry 27, 1143–1157. doi:10.1007/s00787-018-1174-4
Boxer, A. L., Garbutt, S., Seeley, W. W., Jafari, A., Heuer, H. W., Mirsky, J., et al. (2012). Saccade Abnormalities in Autopsy-Confirmed Frontotemporal Lobar Degeneration and Alzheimer Disease. Arch. Neurol. 69, 509–517. doi:10.1001/archneurol.2011.1021
Bozkir, E., Geisler, D., and Kasneci, E. (2019). “Person Independent, Privacy Preserving, and Real Time Assessment of Cognitive Load Using Eye Tracking in a Virtual Reality Setup,” in 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR) (IEEE), 1834–1837. doi:10.1109/vr.2019.8797758
Braun, D. I., Pracejus, L., and Gegenfurtner, K. R. (2006). Motion Aftereffect Elicits Smooth Pursuit Eye Movements. J. Vis. 6, 1. doi:10.1167/6.7.1
Březinová, V., and Kendell, R. (1977). Smooth Pursuit Eye Movements of Schizophrenics and normal People under Stress. Br. J. Psychiatry 130, 59–63.
Brutten, G. J., and Janssen, P. (1979). An Eye-Marking Investigation of Anticipated and Observed Stuttering. J. Speech Lang. Hear. Res. 22, 20–28. doi:10.1044/jshr.2201.20
Buonocore, A., Chen, C.-Y., Tian, X., Idrees, S., Münch, T. A., and Hafed, Z. M. (2017). Alteration of the Microsaccadic Velocity-Amplitude Main Sequence Relationship after Visual Transients: Implications for Models of Saccade Control. J. Neurophysiol. 117, 1894–1910. doi:10.1152/jn.00811.2016
Buonocore, A., McIntosh, R. D., and Melcher, D. (2016). Beyond the point of No Return: Effects of Visual Distractors on Saccade Amplitude and Velocity. J. Neurophysiol. 115, 752–762. doi:10.1152/jn.00939.2015
Burke, M. R., and Barnes, G. R. (2006). Quantitative Differences in Smooth Pursuit and Saccadic Eye Movements. Exp. Brain Res. 175, 596–608. doi:10.1007/s00221-006-0576-6
Burr, D. C., Morrone, M. C., and Ross, J. (1994). Selective Suppression of the Magnocellular Visual Pathway during Saccadic Eye Movements. Nature 371, 511–513. doi:10.1038/371511a0
Buswell, G. T. (1935). How People Look at Pictures: A Study of the Psychology and Perception in Art. Chicago: Univ. Chicago Press.
Cardona, G., García, C., Serés, C., Vilaseca, M., and Gispets, J. (2011). Blink Rate, Blink Amplitude, and Tear Film Integrity during Dynamic Visual Display Terminal Tasks. Curr. Eye Res. 36, 190–197. doi:10.3109/02713683.2010.544442
Carl, J. R., and Gellman, R. S. (1987). Human Smooth Pursuit: Stimulus-dependent Responses. J. Neurophysiol. 57, 1446–1463. doi:10.1152/jn.1987.57.5.1446
Castello, E., Baroni, N., and Pallestrini, E. (1998). Neurotological and Auditory Brain Stem Response Findings in Human Immunodeficiency Virus-Positive Patients without Neurologic Manifestations. Ann. Otol Rhinol Laryngol. 107, 1054–1060. doi:10.1177/000348949810701210
Catherine, L., and James, M. (2001). Educating Children with Autism. Washington, DC: National Academies Press.
Ceder, A. (1977). Drivers' Eye Movements as Related to Attention in Simulated Traffic Flow Conditions. Hum. Factors 19, 571–581. doi:10.1177/001872087701900606
Ceravolo, M. G., Farina, V., Fattobene, L., Leonelli, L., and Raggetti, G. (2019). Presentational Format and Financial Consumers’ Behaviour: an Eye-Tracking Study. Int. J. Bank Marketing 37 (3), 821–837. doi:10.1108/ijbm-02-2018-0041
Chen, S., and Epps, J. (2014). Using Task-Induced Pupil Diameter and Blink Rate to Infer Cognitive Load. Human-Computer Interaction 29, 390–413. doi:10.1080/07370024.2014.892428
Cherng, Y. G., Baird, T., Chen, J. T., and Wang, C. A. (2020). Background Luminance Effects on Pupil Size Associated with Emotion and Saccade Preparation. Sci. Rep. 10, 1–11. doi:10.1038/s41598-020-72954-z
Cho, S.-J., Brown-Schmidt, S., and Lee, W.-y. (2018). Autoregressive Generalized Linear Mixed Effect Models with Crossed Random Effects: An Application to Intensive Binary Time Series Eye-Tracking Data. Psychometrika 83, 751–771. doi:10.1007/s11336-018-9604-2
Choi, J. E. S., Vaswani, P. A., and Shadmehr, R. (2014). Vigor of Movements and the Cost of Time in Decision Making. J. Neurosci. 34, 1212–1223. doi:10.1523/jneurosci.2798-13.2014
Chu, C. A., Rosenfield, M., and Portello, J. K. (2014). Blink Patterns. Optom. Vis. Sci. 91, 297–302. doi:10.1097/opx.0000000000000157
Churchland, A. K., and Lisberger, S. G. (2002). Gain Control in Human Smooth-Pursuit Eye Movements. J. Neurophysiol. 87, 2936–2945. doi:10.1152/jn.2002.87.6.2936
Clay, V., König, P., and König, S. (2019). Eye Tracking in Virtual Reality. J. Eye Mov Res. 12. doi:10.16910/jemr.12.1.3
Coeckelbergh, T. R. M., Brouwer, W. H., Cornelissen, F. W., Van Wolffelaar, P., and Kooijman, A. C. (2002). The Effect of Visual Field Defects on Driving Performance. Arch. Ophthalmol. 120, 1509–1516. doi:10.1001/archopht.120.11.1509
Coëffé, C., and O'regan, J. K. (1987). Reducing the Influence of Non-target Stimuli on Saccade Accuracy: Predictability and Latency Effects. Vis. Res. 27, 227–240. doi:10.1016/0042-6989(87)90185-4
Collewijn, H., and Tamminga, E. P. (1984). Human Smooth and Saccadic Eye Movements during Voluntary Pursuit of Different Target Motions on Different Backgrounds. J. Physiol. 351, 217–250. doi:10.1113/jphysiol.1984.sp015242
Connolly, J. D., Goodale, M. A., Goltz, H. C., and Munoz, D. P. (2005). Fmri Activation in the Human Frontal Eye Field Is Correlated with Saccadic Reaction Time. J. Neurophysiol. 94, 605–611. doi:10.1152/jn.00830.2004
Cornsweet, T. N. (1956). Determination of the Stimuli for Involuntary Drifts and Saccadic Eye Movements*. J. Opt. Soc. Am. 46, 987–993. doi:10.1364/josa.46.000987
Costa, M., Simone, A., Vignali, V., Lantieri, C., and Palena, N. (2018). Fixation Distance and Fixation Duration to Vertical Road Signs. Appl. Ergon. 69, 48–57. doi:10.1016/j.apergo.2017.12.017
Crevits, L., Simons, B., and Wildenbeest, J. (2003). Effect of Sleep Deprivation on Saccades and Eyelid Blinking. Eur. Neurol. 50, 176–180. doi:10.1159/000073060
Daugherty, B. C., Duchowski, A. T., House, D. H., and Ramasamy, C. (2010). “Measuring Vergence over Stereoscopic Video with a Remote Eye Tracker,” in Proceedings of the 2010 Symposium on Eye-Tracking Research & Applications, 97–100. doi:10.1145/1743666.1743690
Day, B. L., and Fitzpatrick, R. C. (2005). The Vestibular System. Curr. Biol. 15, R583–R586. doi:10.1016/j.cub.2005.07.053
de Brouwer, S., Missal, M., and Lefèvre, P. (2001). Role of Retinal Slip in the Prediction of Target Motion during Smooth and Saccadic Pursuit. J. Neurophysiol. 86, 550–558. doi:10.1152/jn.2001.86.2.550
De Brouwer, S., Yuksel, D., Blohm, G., Missal, M., and Lefèvre, P. (2002). What Triggers Catch-Up Saccades during Visual Tracking. J. Neurophysiol. 87, 1646–1650. doi:10.1152/jn.00432.2001
De Bruin, J., Malan, K., and Eloff, J. (2013). “Saccade Deviation Indicators for Automated Eye Tracking Analysis,” in Proceedings of the 2013 conference on eye tracking south africa, 47–54. doi:10.1145/2509315.2509324
de Hemptinne, C., Lefèvre, P., and Missal, M. (2006). Influence of Cognitive Expectation on the Initiation of Anticipatory and Visual Pursuit Eye Movements in the Rhesus Monkey. J. Neurophysiol. 95, 3770–3782. doi:10.1152/jn.00007.2006
Dell'Osso, L. F., Van der Steen, J., Steinman, R. M., and Collewijn, H. (1992). Foveation Dynamics in Congenital Nystagmus. II: Smooth Pursuit. Doc Ophthalmol. 79, 25–49. doi:10.1007/BF00160131
Demberg, V., Kiagia, E., and Sayeed, A. (2013). The index of Cognitive Activity as a Measure of Linguistic Processing. reading time 500, 1500.
Demberg, V. (2013). “Pupillometry: the index of Cognitive Activity in a Dual-Task Study,” in Proceedings of the Annual Meeting of the Cognitive Science Society, 2154–2159.35
Dimigen, O., Valsecchi, M., Sommer, W., and Kliegl, R. (2009). Human Microsaccade-Related Visual Brain Responses. J. Neurosci. 29, 12321–12331. doi:10.1523/jneurosci.0911-09.2009
Ditchburn, R. W., and Ginsborg, B. L. (1953). Involuntary Eye Movements during Fixation. J. Physiol. 119, 1–17. doi:10.1113/jphysiol.1953.sp004824
Dobson, M. W. (1977). Eye Movement Parameters and Map reading. The Am. Cartographer 4, 39–58. doi:10.1559/152304077784080022
Dodge, R. (1903). Five Types of Eye Movement in the Horizontal meridian Plane of the Field of Regard. Am. J. physiology-legacy content 8, 307–329. doi:10.1152/ajplegacy.1903.8.4.307
Dodge, R. (1900). Visual Perception during Eye Movement. Psychol. Rev. 7, 454–465. doi:10.1037/h0067215
Donovan, T., and Litchfield, D. (2013). Looking for Cancer: Expertise Related Differences in Searching and Decision Making. Appl. Cognit. Psychol. 27, 43–49. doi:10.1002/acp.2869
Doughty, M. J. (2001). Consideration of Three Types of Spontaneous Eyeblink Activity in normal Humans: during reading and Video Display Terminal Use, in Primary Gaze, and while in Conversation. Optom. Vis. Sci. 78, 712–725. doi:10.1097/00006324-200110000-00011
Doughty, M. J. (2019). Effect of Distance Vision and Refractive Error on the Spontaneous Eye Blink Activity in Human Subjects in Primary Eye Gaze. J. Optom. 12, 111–119. doi:10.1016/j.optom.2018.03.004
Doughty, M. J. (2002). Further Assessment of Gender- and Blink Pattern-Related Differences in the Spontaneous Eyeblink Activity in Primary Gaze in Young Adult Humans. Optom. Vis. Sci. 79, 439–447. doi:10.1097/00006324-200207000-00013
Doughty, M. J., and Naase, T. (2006). Further Analysis of the Human Spontaneous Eye Blink Rate by a Cluster Analysis-Based Approach to Categorize Individuals with 'Normal' versus 'Frequent' Eye Blink Activity. Eye & contact lens 32, 294–299. doi:10.1097/01.icl.0000224359.32709.4d
Du, P., and MacDonald, E. F. (2014). Eye-tracking Data Predict Importance of Product Features and Saliency of Size Change. J. Mech. Des. 136. doi:10.1115/1.4027387
Duchowski, A. T. (2002). A Breadth-First Survey of Eye-Tracking Applications. Behav. Res. Methods Instr. Comput. 34, 455–470. doi:10.3758/BF03195475
Duchowski, A. T., Krejtz, K., Gehrer, N. A., Bafna, T., and Bækgaard, P. (2020). “The Low/high index of Pupillary Activity,” in Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, 1–12. doi:10.1145/3313831.3376394
Duchowski, A. T., Krejtz, K., Krejtz, I., Biele, C., Niedzielska, A., Kiefer, P., et al. (2018). “The index of Pupillary Activity: Measuring Cognitive Load Vis-À-Vis Task Difficulty with Pupil Oscillation,” in Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, 1–13.
Duncan, J., Ward, R., and Shapiro, K. (1994). Direct Measurement of Attentional Dwell Time in Human Vision. Nature 369, 313–315. doi:10.1038/369313a0
Duret, F., Buquet, C., Charlier, J., Mermoud, C., Viviani, P., and Safran, A. B. (1999). Refixation Strategies in Four Patients with Macular Disorders. Neuro-Ophthalmology 22, 209–220. doi:10.1076/noph.22.4.209.3718
Eberhard, K. M., Spivey-Knowlton, M. J., Sedivy, J. C., and Tanenhaus, M. K. (1995). Eye Movements as a Window into Real-Time Spoken Language Comprehension in Natural Contexts. J. Psycholinguist Res. 24, 409–436. doi:10.1007/bf02143160
Ebitz, R. B., and Moore, T. (2019). Both a Gauge and a Filter: Cognitive Modulations of Pupil Size. Front. Neurol. 9, 1190. doi:10.3389/fneur.2018.01190
El Haj, M., and Moustafa, A. A. (2021). Pupil Dilation as an Indicator of Future Thinking. Neurol. Sci. 42, 647–653. doi:10.1007/s10072-020-04533-z
Eldar, E., Cohen, J. D., and Niv, Y. (2013). The Effects of Neural Gain on Attention and Learning. Nat. Neurosci. 16, 1146–1153. doi:10.1038/nn.3428
Ellis, S., Candrea, R., Misner, J., Craig, C. S., Lankford, C. P., and Hutchinson, T. E. (1998). “Windows to the Soul? what Eye Movements Tell Us about Software Usability,” in Proceedings of the usability professionals’ association conference, 151–178.
Essig, K., Pomplun, M., and Ritter, H. (2004). “Application of a Novel Neural Approach to 3d Gaze Tracking,” in Vergence Eye-Movements in Autostereograms, 26. eScholarship.
Ettinger, U., Kumari, V., Chitnis, X. A., Corr, P. J., Sumich, A. L., Rabe-Hesketh, S., et al. (2002). Relationship between Brain Structure and Saccadic Eye Movements in Healthy Humans. Neurosci. Lett. 328, 225–228. doi:10.1016/s0304-3940(02)00517-7
Farbos, B., Mollard, R., Cabon, P., and David, H. (2000). “Measurement of Fatigue and Adaptation in Large-Scale Real-Time Atc Simulation,” in Proceedings of the Human Factors and Ergonomics Society Annual Meeting, Los Angeles, CA (Los Angeles, CA: SAGE Publications Sage CA), 3–204. doi:10.1177/154193120004401905
Fehringer, B. C. O. F. (2021). Optimizing the Usage of Pupillary Based Indicators for Cognitive Workload. J. Eye Mov Res. 14. doi:10.16910/jemr.14.2.4
Fehringer, B. C. (2020). “One Threshold to Rule Them All? Modification of the index of Pupillary Activity to Optimize the Indication of Cognitive Load,” in ACM Symposium on Eye Tracking Research and Applications, 1–5.
Findlay, J. M., Walker, R., and Kentridge, R. W. (1995). Eye Movement Research: Mechanisms, Processes and Applications. Elsevier.
Fischer, B., and Breitmeyer, B. (1987). Mechanisms of Visual Attention Revealed by Saccadic Eye Movements. Neuropsychologia 25, 73–83. doi:10.1016/0028-3932(87)90044-3
Fischer, B., and Weber, H. (1993). Express Saccades and Visual Attention. Behav. Brain Sci. 16, 553–567. doi:10.1017/s0140525x00031575
Fisher, D. F., Monty, R. A., and Senders, J. W. (2017). Eye Movements: Cognition and Visual Perception, Vol. 8. Oxfordshire, England, UK: Routledge.
Foulsham, T., Kingstone, A., and Underwood, G. (2008). Turning the World Around: Patterns in Saccade Direction Vary with Picture Orientation. Vis. Res. 48, 1777–1790. doi:10.1016/j.visres.2008.05.018
Fukushima, K., Fukushima, J., Warabi, T., and Barnes, G. R. (2013). Cognitive Processes Involved in Smooth Pursuit Eye Movements: Behavioral Evidence, Neural Substrate and Clinical Correlation. Front. Syst. Neurosci. 7, 4. doi:10.3389/fnsys.2013.00004
Galley, N., Schleicher, R., and Galley, L. (2004). Blink Parameters as Indicators of Driver’s Sleepiness–Possibilities and Limitations. Vis. vehicles 10, 189–196.
Galley, N. (1993). The Evaluation of the Electrooculogram as a Psychophysiological Measuring Instrument in the Driver Study of Driver Behaviour. Ergonomics 36, 1063–1070. doi:10.1080/00140139308967978
Gbadamosi, J., and Zangemeister, W. H. (2001). Visual Imagery in Hemianopic Patients. J. Cogn. Neurosci. 13, 855–866. doi:10.1162/089892901753165782
Goldberg, J. H., and Helfman, J. I. (2010). “Scanpath Clustering and Aggregation,” in Proceedings of the 2010 symposium on eye-tracking research & applications, 227–234. doi:10.1145/1743666.1743721
Goldberg, J. H., and Kotval, X. P. (1999). Computer Interface Evaluation Using Eye Movements: Methods and Constructs. Int. J. Ind. Ergon. 24, 631–645. doi:10.1016/s0169-8141(98)00068-7
Goldberg, J. H., and Wichansky, A. M. (2003). “Eye Tracking in Usability Evaluation,” in The Mind’s Eye. Editors J. Hyönä, R. Radach, and H. Deubel (Amsterdam: North-Holland), 493–516. doi:10.1016/B978-044451020-4/50027-X
Goldberg, J. M., and Fernández, C. (1984). The Vestibular System. Handbook Physiol. 3, 977–1022. doi:10.1002/cphy.cp010321
Goldberg, M. E., Eggers, H., and Gouras, P. (1991). The Oculomotor System. Principles Neural Sci., 660–676.
Gray, L. S., Winn, B., and Gilmartin, B. (1993). Accommodative Microfluctuations and Pupil Diameter. Vis. Res. 33, 2083–2090. doi:10.1016/0042-6989(93)90007-j
Griffiths, A., Marshall, R., and Richens, A. (1984). Saccadic Eye Movement Analysis as a Measure of Drug Effects on Human Psychomotor Performance. Br. J. Clin. Pharmacol. 18, 73S–82S. doi:10.1111/j.1365-2125.1984.tb02584.x
Grindinger, T. J., Murali, V. N., Tetreault, S., Duchowski, A. T., Birchfield, S. T., and Orero, P. (2010). “Algorithm for Discriminating Aggregate Gaze Points: Comparison with Salient Regions-Of-Interest,” in Asian Conference on Computer Vision (Springer), 390–399.
Guillon, Q., Afzali, M. H., Rogé, B., Baduel, S., Kruck, J., and Hadjikhani, N. (2015). The Importance of Networking in Autism Gaze Analysis. PLoS one 10, e0141191. doi:10.1371/journal.pone.0141191
Guo, F., Ding, Y., Liu, W., Liu, C., and Zhang, X. (2016). Can Eye-Tracking Data Be Measured to Assess Product Design?: Visual Attention Mechanism Should Be Considered. Int. J. Ind. Ergon. 53, 229–235. doi:10.1016/j.ergon.2015.12.001
Hafed, Z. M., Chen, C.-Y., and Tian, X. (2015). Vision, Perception, and Attention through the Lens of Microsaccades: Mechanisms and Implications. Front. Syst. Neurosci. 9, 167. doi:10.3389/fnsys.2015.00167
Hafed, Z. M., and Clark, J. J. (2002). Microsaccades as an Overt Measure of covert Attention Shifts. Vis. Res. 42, 2533–2545. doi:10.1016/s0042-6989(02)00263-8
Hauland, G., and Duijm, N. (2002). “Eye Movement Based Measures of Team Situation Awareness (Tsa),” in Japan-halden MMS Workshop (Kyoto, Japan: Kyoto University), 82–85.
He, Y., Su, Q., Wang, L., He, W., Tan, C., Zhang, H., et al. (2019). The Characteristics of Intelligence Profile and Eye Gaze in Facial Emotion Recognition in Mild and Moderate Preschoolers with Autism Spectrum Disorder. Front. Psychiatry 10, 402. doi:10.3389/fpsyt.2019.00402
Heilmann, F., and Witte, K. (2021). Perception and Action under Different Stimulus Presentations: A Review of Eye-Tracking Studies with an Extended View on Possibilities of Virtual Reality. Appl. Sci. 11, 5546. doi:10.3390/app11125546
Hejtmancik, J. F., Cabrera, P., Chen, Y., M’Hamdi, O., and Nickerson, J. M. (2017). “Vision,” in Conn’s Translational Neuroscience. Editor P. M. Conn (San Diego: Academic Press), 399–438. doi:10.1016/B978-0-12-802381-5.00031-2
Heminghous, J., and Duchowski, A. T. (2006). “Icomp: a Tool for Scanpath Visualization and Comparison,” in Proceedings of the 3rd Symposium on Applied Perception in Graphics and Visualization, 152.
Henderson, J. M., Choi, W., Luke, S. G., and Schmidt, J. (2018). Neural Correlates of Individual Differences in Fixation Duration during Natural reading. Q. J. Exp. Psychol. 71, 314–323. doi:10.1080/17470218.2017.1329322
Hess, E. H., and Polt, J. M. (1964). Pupil Size in Relation to Mental Activity during Simple Problem-Solving. Science 143, 1190–1192. doi:10.1126/science.143.3611.1190
Hessels, R. S., and Hooge, I. T. C. (2019). Eye Tracking in Developmental Cognitive Neuroscience - the Good, the Bad and the Ugly. Dev. Cogn. Neurosci. 40, 100710. doi:10.1016/j.dcn.2019.100710
Hessels, R. S., Kemner, C., van den Boomen, C., and Hooge, I. T. C. (2016). The Area-Of-Interest Problem in Eyetracking Research: A Noise-Robust Solution for Face and Sparse Stimuli. Behav. Res. 48, 1694–1712. doi:10.3758/s13428-015-0676-y
Hessels, R. S., Niehorster, D. C., Nyström, M., Andersson, R., and Hooge, I. T. C. (2018). Is the Eye-Movement Field Confused about Fixations and Saccades? A Survey Among 124 Researchers. R. Soc. Open Sci., 180502. doi:10.1098/rsos.180502
Hoang Duc, A., Bays, P., and Husain, M. (2008). Eye Movements as a Probe of Attention. Prog. Brain Res. 171, 403–411. doi:10.1016/s0079-6123(08)00659-6
Holland, C., and Komogortsev, O. V. (2011). “Biometric Identification via Eye Movement Scanpaths in reading,” in 2011 International joint conference on biometrics (IJCB), 1–8. doi:10.1109/ijcb.2011.6117536IEEE
Holmqvist, K., Nyström, M., Andersson, R., Dewhurst, R., Jarodzka, H., and Van de Weijer, J. (2011). Eye Tracking: A Comprehensive Guide to Methods and Measures. Oxford, UK: OUP Oxford.
Horley, K., Williams, L. M., Gonsalvez, C., and Gordon, E. (2003). Social Phobics Do Not See Eye to Eye:. J. anxiety Disord. 17, 33–44. doi:10.1016/s0887-6185(02)00180-9
Hubel, D. H., and Wiesel, T. N. (1979). Brain Mechanisms of Vision. Sci. Am. 241, 150–162. doi:10.1038/scientificamerican0979-150
Hüsser, A., and Wirth, W. (2014). Do investors Show an Attentional Bias toward Past Performance? an Eye-Tracking experiment on Visual Attention to Mutual Fund Disclosures in Simplified Fund Prospectuses. J. Financ. Serv. Mark 19, 169–185. doi:10.1057/fsm.2014.20
Hyönä, J., Tommola, J., and Alaja, A.-M. (1995). Pupil Dilation as a Measure of Processing Load in Simultaneous Interpretation and Other Language Tasks. The Q. J. Exp. Psychol. Section A 48, 598–612. doi:10.1080/14640749508401407
Iqbal, S. T., Zheng, X. S., and Bailey, B. P. (2004). “Task-evoked Pupillary Response to Mental Workload in Human-Computer Interaction,” in CHI’04 Extended Abstracts on Human Factors in Computing Systems, 1477–1480. doi:10.1145/985921.986094
Jacob, R. J. K., and Karn, K. S. (2003). “Eye Tracking in Human-Computer Interaction and Usability Research,” in The Mind’s Eye. Editors J. Hyönä, R. Radach, and H. Deubel (Amsterdam: North-Holland), 573–605. doi:10.1016/B978-044451020-4/50031-1
Jainta, S., Vernet, M., Yang, Q., and Kapoula, Z. (2011). The Pupil Reflects Motor Preparation for Saccades - Even before the Eye Starts to Move. Front. Hum. Neurosci. 5, 97. doi:10.3389/fnhum.2011.00097
Jarodzka, H., Holmqvist, K., and Nyström, M. (2010). “A Vector-Based, Multidimensional Scanpath Similarity Measure,” in Proceedings of the 2010 symposium on eye-tracking research & applications, 211–218. doi:10.1145/1743666.1743718
Jayawardena, G., and Jayarathna, S. (2021). “Automated Filtering of Eye Movements Using Dynamic Aoi in Multiple Granularity Levels,” in International Journal of Multimedia Data Engineering and Management (IJMDEM), 12, 49–64. doi:10.4018/ijmdem.2021010104
Jayawardena, G., Michalek, A., Duchowski, A., and Jayarathna, S. (2020). “Pilot Study of Audiovisual Speech-In-Noise (Sin) Performance of Young Adults with Adhd,” in ACM Symposium on Eye Tracking Research and Applications, 1–5. doi:10.1145/3379156.3391373
Jensen, K., Beylergil, S. B., and Shaikh, A. G. (2019). Slow Saccades in Cerebellar Disease. cerebellum ataxias 6, 1–9. doi:10.1186/s40673-018-0095-9
Jiang, M.-Q. (1996). The Role of Attention Mechanisms in Smooth Pursuit Performance in normal and Schizophrenic Subjects. Iowa City, IA: The University of Iowa.
Jongkees, B. J., and Colzato, L. S. (2016). Spontaneous Eye Blink Rate as Predictor of Dopamine-Related Cognitive Function-A Review. Neurosci. Biobehavioral Rev. 71, 58–82. doi:10.1016/j.neubiorev.2016.08.020
Joshi, S., Li, Y., Kalwani, R. M., and Gold, J. I. (2016). Relationships between Pupil Diameter and Neuronal Activity in the Locus Coeruleus, Colliculi, and Cingulate Cortex. Neuron 89, 221–234. doi:10.1016/j.neuron.2015.11.028
Kao, G. W., and Morrow, M. J. (1994). The Relationship of Anticipatory Smooth Eye Movement to Smooth Pursuit Initiation. Vis. Res. 34, 3027–3036. doi:10.1016/0042-6989(94)90276-3
Karsh, R., and Breitenbach, F. (1983). Looking at Looking: The Amorphous Fixation Measure. Eye movements Psychol. functions: Int. views, 53–64. doi:10.4324/9781003165538-6
Killian, N. J., Potter, S. M., and Buffalo, E. A. (2015). Saccade Direction Encoding in the Primate Entorhinal Cortex during Visual Exploration. Proc. Natl. Acad. Sci. USA 112, 15743–15748. doi:10.1073/pnas.1417059112
Kimble, M. O., Fleming, K., Bandy, C., Kim, J., and Zambetti, A. (2010). Eye Tracking and Visual Attention to Threating Stimuli in Veterans of the iraq War. J. anxiety Disord. 24, 293–299. doi:10.1016/j.janxdis.2009.12.006
Knox, P. C., and Wolohan, F. D. A. (2014). Cultural Diversity and Saccade Similarities: Culture Does Not Explain Saccade Latency Differences between Chinese and Caucasian Participants. PloS one 9, e94424. doi:10.1371/journal.pone.0094424
Komogortsev, O., Holland, C., Jayarathna, S., and Karpov, A. (2013). 2D Linear Oculomotor Plant Mathematical Model. ACM Trans. Appl. Percept. 10, 1–18. doi:10.1145/2536764.2536774
Komogortsev, O. V., Jayarathna, S., Koh, D. H., and Gowda, S. M. (2010). “Qualitative and Quantitative Scoring and Evaluation of the Eye Movement Classification Algorithms,” in Proceedings of the 2010 Symposium on eye-tracking research & applications, 65–68. doi:10.1145/1743666.1743682
Korbach, A., Brünken, R., and Park, B. (2018). Differentiating Different Types of Cognitive Load: A Comparison of Different Measures. Educ. Psychol. Rev. 30, 503–529. doi:10.1007/s10648-017-9404-8
Korbach, A., Brünken, R., and Park, B. (2017). Measurement of Cognitive Load in Multimedia Learning: a Comparison of Different Objective Measures. Instr. Sci. 45, 515–536. doi:10.1007/s11251-017-9413-5
Koster, W. (1895). Étude sur les cônes et les bâtonnets dans la region de la fovea centralis de la rétine chez l’homme. Arch. D’opht. 15, 428.
Krafka, K., Khosla, A., Kellnhofer, P., Kannan, H., Bhandarkar, S., Matusik, W., et al. (2016). “Eye Tracking for Everyone,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2176–2184. doi:10.1109/cvpr.2016.239
Krauzlis, R. J. (2004). Recasting the Smooth Pursuit Eye Movement System. J. Neurophysiol. 91, 591–603. doi:10.1152/jn.00801.2003
Krejtz, K., Żurawska, J., Duchowski, A. T., and Wichary, S. (2020). Pupillary and Microsaccadic Responses to Cognitive Effort and Emotional Arousal during Complex Decision Making. J. Eye Mov Res. 13. doi:10.16910/jemr.13.5.2
Krejtz, K., Duchowski, A., Szmidt, T., Krejtz, I., González Perilli, F., Pires, A., et al. (2015). Gaze Transition Entropy. ACM Trans. Appl. Percept. 13, 1–20. doi:10.1145/2834121
Krejtz, K., Duchowski, A. T., Niedzielska, A., Biele, C., and Krejtz, I. (2018). Eye Tracking Cognitive Load Using Pupil Diameter and Microsaccades with Fixed Gaze. PloS one 13, e0203629. doi:10.1371/journal.pone.0203629
Krejtz, K., Szmidt, T., Duchowski, A. T., and Krejtz, I. (2014). “Entropy-based Statistical Analysis of Eye Movement Transitions,” in Proceedings of the Symposium on Eye Tracking Research and Applications, 159–166. doi:10.1145/2578153.2578176
Krupinski, E. A. (1996). Visual Scanning Patterns of Radiologists Searching Mammograms. Acad. Radiol. 3, 137–144. doi:10.1016/s1076-6332(05)80381-2
Ladda, J., Eggert, T., Glasauer, S., and Straube, A. (2007). Velocity Scaling of Cue-Induced Smooth Pursuit Acceleration Obeys Constraints of Natural Motion. Exp. Brain Res. 182, 343–356. doi:10.1007/s00221-007-0988-y
Lai, H. Y., Saavedra-Pena, G., Sodini, C. G., Sze, V., and Heldt, T. (2019). Measuring Saccade Latency Using Smartphone Cameras. IEEE J. Biomed. Health Inform. 24, 885–897. doi:10.1109/JBHI.2019.2913846
Land, M. F., and Furneaux, S. (1997). The Knowledge Base of the Oculomotor System. Phil. Trans. R. Soc. Lond. B 352, 1231–1239. doi:10.1098/rstb.1997.0105
Le Meur, O., Coutrot, A., Liu, Z., Rämä, P., Le Roch, A., and Helo, A. (2017). Visual Attention Saccadic Models Learn to Emulate Gaze Patterns from Childhood to Adulthood. IEEE Trans. Image Process. 26, 4777–4789. doi:10.1109/tip.2017.2722238
Lee, K. I., Jeon, J. H., and Song, B. C. (2020). “Deep Learning-Based Pupil center Detection for Fast and Accurate Eye Tracking System,” in European Conference on Computer Vision (Springer), 36–52. doi:10.1007/978-3-030-58529-7_3
Lehtinen, I., Lang, A. H., Jäntti, V., and Keskinen, E. (1979). Acute Effects of Alcohol on Saccadic Eye Movements. Psychopharmacology 63, 17–23. doi:10.1007/bf00426915
Lisberger, S. G., Morris, E., and Tychsen, L. (1987). Visual Motion Processing and Sensory-Motor Integration for Smooth Pursuit Eye Movements. Annu. Rev. Neurosci. 10 (1), 97–129. doi:10.1146/annurev.ne.10.030187.000525
Lisi, M., Solomon, J. A., and Morgan, M. J. (2019). Gain Control of Saccadic Eye Movements Is Probabilistic. Proc. Natl. Acad. Sci. USA 116, 16137–16142. doi:10.1073/pnas.1901963116
MacAskill, M. R., Anderson, T. J., and Jones, R. D. (2002). Adaptive Modification of Saccade Amplitude in Parkinson's Disease. Brain 125, 1570–1582. doi:10.1093/brain/awf168
Macdonald, J. S. P., and Lavie, N. (2011). Visual Perceptual Load Induces Inattentional Deafness. Atten Percept Psychophys 73, 1780–1789. doi:10.3758/s13414-011-0144-4
Maffei, A., and Angrilli, A. (2019). Spontaneous Blink Rate as an index of Attention and Emotion during Film Clips Viewing. Physiol. Behav. 204, 256–263. doi:10.1016/j.physbeh.2019.02.037
Mahanama, B., Jayawardana, Y., and Jayarathna, S. (2020). “Gaze-net: Appearance-Based Gaze Estimation Using Capsule Networks,” in Proceedings of the 11th Augmented Human International Conference, 1–4.
Maier, J. X., and Groh, J. M. (2009). Multisensory Guidance of Orienting Behavior. Hearing Res. 258, 106–112. doi:10.1016/j.heares.2009.05.008
Maier, S. U., and Grueschow, M. (2021). Pupil Dilation Predicts Individual Self-Regulation success across Domains. Sci. Rep. 11, 1–18. doi:10.1038/s41598-021-93121-y
Mania, K., McNamara, A., and Polychronakis, A. (2021). “Gaze-aware Displays and Interaction,” in ACM SIGGRAPH 2021 Courses, 1–67. doi:10.1145/3450508.3464606
Marshall, S. P. (2007). Identifying Cognitive State from Eye Metrics. Aviat Space Environ. Med. 78, B165–B175.
Marshall, S. P. (2000). Method and Apparatus for Eye Tracking and Monitoring Pupil Dilation to Evaluate Cognitive Activity. US Patent 6, 051–090. [Dataset].
Marshall, S. P. (2002). “The index of Cognitive Activity: Measuring Cognitive Workload,” in Proceedings of the IEEE 7th conference on Human Factors and Power Plants (IEEE), 7.
Marslen-Wilson, W. D. (1985). Speech Shadowing and Speech Comprehension. Speech Commun. 4, 55–73. doi:10.1016/0167-6393(85)90036-6
Martinez-Conde, S., Macknik, S. L., and Hubel, D. H. (2004). The Role of Fixational Eye Movements in Visual Perception. Nat. Rev. Neurosci. 5, 229–240. doi:10.1038/nrn1348
May, J. G., Kennedy, R. S., Williams, M. C., Dunlap, W. P., and Brannan, J. R. (1990). Eye Movement Indices of Mental Workload. Acta psychologica 75, 75–89. doi:10.1016/0001-6918(90)90067-p
McConkie, G. W., and Rayner, K. (1975). The Span of the Effective Stimulus during a Fixation in reading. Perception & Psychophysics 17, 578–586. doi:10.3758/bf03203972
McGregor, D. K., and Stern, J. A. (1996). Time on Task and Blink Effects on Saccade Duration. Ergonomics 39, 649–660. doi:10.1080/00140139608964487
McKee, S. P., Levi, D. M., Schor, C. M., and Movshon, J. A. (2016). Saccadic Latency in Amblyopia. J. Vis. 16, 3. doi:10.1167/16.5.3
McSorley, E., McCloy, R., and Lyne, C. (2012). The Spatial Impact of Visual Distractors on Saccade Latency. Vis. Res. 60, 61–72. doi:10.1016/j.visres.2012.03.007
Megaw, E. D. (1979). Factors Affecting Visual Inspection Accuracy. Appl. Ergon. 10, 27–32. doi:10.1016/0003-6870(79)90006-1
Megaw, E. D., and Richardson, J. (1979). Eye Movements and Industrial Inspection. Appl. Ergon. 10, 145–154. doi:10.1016/0003-6870(79)90138-8
Meghanathan, R. N., Nikolaev, A. R., and van Leeuwen, C. (2019). Refixation Patterns Reveal Memory-Encoding Strategies in Free Viewing. Atten Percept Psychophys 81, 2499–2516. doi:10.3758/s13414-019-01735-2
Mele, M. L., and Federici, S. (2012). Gaze and Eye-Tracking Solutions for Psychological Research. Cogn. Process. 13, 261–265. doi:10.1007/s10339-012-0499-z
Mello-Thoms, C., Hardesty, L., Sumkin, J., Ganott, M., Hakim, C., Britton, C., et al. (2005). Effects of Lesion Conspicuity on Visual Search in Mammogram reading1. Acad. Radiol. 12, 830–840. doi:10.1016/j.acra.2005.03.068
Menon, R. G. V., Sigurdsson, V., Larsen, N. M., Fagerstrøm, A., and Foxall, G. R. (2016). Consumer Attention to price in Social Commerce: Eye Tracking Patterns in Retail Clothing. J. Business Res. 69, 5008–5013. doi:10.1016/j.jbusres.2016.04.072
Meyer, C. H., Lasker, A. G., and Robinson, D. A. (1985). The Upper Limit of Human Smooth Pursuit Velocity. Vis. Res. 25, 561–563. doi:10.1016/0042-6989(85)90160-9
Michell, A. W., Xu, Z., Fritz, D., Lewis, S. J. G., Foltynie, T., Williams-Gray, C. H., et al. (2006). Saccadic Latency Distributions in Parkinson's Disease and the Effects of L-Dopa. Exp. Brain Res. 174, 7–18. doi:10.1007/s00221-006-0412-z
Mlot, E. G., Bahmani, H., Wahl, S., and Kasneci, E. (2016). “3d Gaze Estimation Using Eye Vergence,” in Proceedings of the 9th International Joint Conference on Biomedical Engineering Systems and Technologies (Rome, Italy: HEALTHINF), 125–131. doi:10.5220/0005821201250131
Mohanty, A., and Sussman, T. J. (2013). Top-down Modulation of Attention by Emotion. Front. Hum. Neurosci. 7, 102. doi:10.3389/fnhum.2013.00102
Morris, T. L., and Miller, J. C. (1996). Electrooculographic and Performance Indices of Fatigue during Simulated Flight. Biol. Psychol. 42, 343–360. doi:10.1016/0301-0511(95)05166-x
Mostofi, N., Zhao, Z., Intoy, J., Boi, M., Victor, J. D., and Rucci, M. (2020). Spatiotemporal Content of Saccade Transients. Curr. Biol. 30, 3999–4008. doi:10.1016/j.cub.2020.07.085
Motoki, K., Saito, T., and Onuma, T. (2021). Eye-tracking Research on Sensory and Consumer Science: A Review, Pitfalls and Future Directions. Food Res. Int. 145, 110389. doi:10.1016/j.foodres.2021.110389
Mulder, K., Klugkist, I., van Renswoude, D., and Visser, I. (2020). Mixtures of Peaked Power Batschelet Distributions for Circular Data with Application to Saccade Directions. J. Math. Psychol. 95, 102309. doi:10.1016/j.jmp.2019.102309
Mutasim, A. K., Stuerzlinger, W., and Batmaz, A. U. (2020). “Gaze Tracking for Eye-Hand Coordination Training Systems in Virtual Reality,” in Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, 1–9. doi:10.1145/3334480.3382924
Nachmias, J. (1961). Determiners of the Drift of the Eye during Monocular Fixation*. J. Opt. Soc. Am. 51, 761–766. doi:10.1364/josa.51.000761
Nakayama, M., Takahashi, K., and Shimizu, Y. (2002). “The Act of Task Difficulty and Eye-Movement Frequency for The’oculo-Motor Indices,” in Proceedings of the 2002 symposium on Eye tracking research & applications, 37–42. doi:10.1145/507072.507080
Newhall, S. M. (1932). The Control of Eyelid Movements in Visual Experiments. Am. J. Psychol. 44, 555–570. doi:10.2307/1415357
Noronha, B., Dziemian, S., Zito, G. A., Konnaris, C., and Faisal, A. A. (2017). “"Wink to Grasp" - Comparing Eye, Voice & EMG Gesture Control of Grasp with Soft-Robotic Gloves,” in 2017 International Conference on Rehabilitation Robotics (ICORR) (IEEE), 1043–1048. doi:10.1109/ICORR.2017.8009387IEEE Int. Conf. Rehabil. Robot2017
Nyström, M., and Holmqvist, K. (2008). Semantic Override of Low-Level Features in Image Viewing–Both Initially and Overall. J. Eye Move. Res. 2. doi:10.16910/jemr.2.2.2
O’Driscoll, G. A., and Callahan, B. L. (2008). Smooth Pursuit in Schizophrenia: a Meta-Analytic Review of Research since 1993. Brain Cogn. 68, 359–370.
Oh, J., Han, M., Peterson, B. S., and Jeong, J. (2012). Spontaneous Eyeblinks Are Correlated with Responses during the Stroop Task. PloS one 7, e34871. doi:10.1371/journal.pone.0034871
Ohtani, A. (1971). An Analysis of Eye Movements during a Visual Task. Ergonomics 14, 167–174. doi:10.1080/00140137108931235
Otero-Millan, J., Troncoso, X. G., Macknik, S. L., Serrano-Pedraza, I., and Martinez-Conde, S. (2008). Saccades and Microsaccades during Visual Fixation, Exploration, and Search: Foundations for a Common Saccadic Generator. J. Vis. 8, 21. doi:10.1167/8.14.21
Pan, J., Ferrer, C. C., McGuinness, K., O’Connor, N. E., Torres, J., Sayrol, E., et al. (2017). Salgan: Visual Saliency Prediction with Generative Adversarial Networks.arXiv preprint arXiv:1701.01081
Parkhurst, D., Law, K., and Niebur, E. (2002). Modeling the Role of Salience in the Allocation of Overt Visual Attention. Vis. Res. 42, 107–123. doi:10.1016/s0042-6989(01)00250-4
Patel, S. S., Jankovic, J., Hood, A. J., Jeter, C. B., and Sereno, A. B. (2012). Reflexive and Volitional Saccades: Biomarkers of huntington Disease Severity and Progression. J. Neurol. Sci. 313, 35–41. doi:10.1016/j.jns.2011.09.035
Pavlović, N., and Jensen, K. (2009). Eye Tracking Translation Directionality. Translation Res. projects 2, 93–109.
Peterson, J., and Allison, L. W. (1931). Controls of the Eye-Wink Mechanism. J. Exp. Psychol. 14, 144–154. doi:10.1037/h0070197
Peysakhovich, V., Vachon, F., and Dehais, F. (2017). The Impact of Luminance on Tonic and Phasic Pupillary Responses to Sustained Cognitive Load. Int. J. Psychophysiology 112, 40–45. doi:10.1016/j.ijpsycho.2016.12.003
Phillips, M. H., and Edelman, J. A. (2008). The Dependence of Visual Scanning Performance on Search Direction and Difficulty. Vis. Res. 48, 2184–2192. doi:10.1016/j.visres.2008.06.025
Ponsoda, V., Scott, D., and Findlay, J. M. (1995). A Probability Vector and Transition Matrix Analysis of Eye Movements during Visual Search. Acta psychologica 88, 167–185. doi:10.1016/0001-6918(95)94012-y
Posner, M. I., Snyder, C. R., and Davidson, B. J. (1980). Attention and the Detection of Signals. J. Exp. Psychol. Gen. 109, 160–174. doi:10.1037/0096-3445.109.2.160
Pumphrey, R. J. (1948). The Theory of the Fovea. J. Exp. Biol. 25, 299–312. doi:10.1242/jeb.25.3.299
Ranti, C., Jones, W., Klin, A., and Shultz, S. (2020). Blink Rate Patterns Provide a Reliable Measure of Individual Engagement with Scene Content. Sci. Rep. 10, 1–10. doi:10.1038/s41598-020-64999-x
Rashbass, C. (1961). The Relationship between Saccadic and Smooth Tracking Eye Movements. J. Physiol. 159, 326–338. doi:10.1113/jphysiol.1961.sp006811
Ratliff, F., and Riggs, L. A. (1950). Involuntary Motions of the Eye during Monocular Fixation. J. Exp. Psychol. 40, 687–701. doi:10.1037/h0057754
Rayner, K. (1979). Eye Guidance in reading: Fixation Locations within Words. Perception 8, 21–30. doi:10.1068/p080021
Rayner, K. (2012). Eye Movements and Visual Cognition: Scene Perception and reading. Springer Science & Business Media.
Rayner, K. (1978). Eye Movements in reading and Information Processing. Psychol. Bull. 85, 618–660. doi:10.1037/0033-2909.85.3.618
Rayner, K., Pollatsek, A., Ashby, J., and Clifton, C. (2012). Psychology of Reading. 1 edn. Hove, East Sussex, UK: Psychology Press.
Rayner, K. (1975). The Perceptual Span and Peripheral Cues in reading. Cogn. Psychol. 7, 65–81. doi:10.1016/0010-0285(75)90005-5
Recarte, M. A., and Nunes, L. M. (2000). Effects of Verbal and Spatial-Imagery Tasks on Eye Fixations while Driving. J. Exp. Psychol. Appl. 6, 31–43. doi:10.1037/1076-898x.6.1.31
Redmon, J., Divvala, S., Girshick, R., and Farhadi, A. (2016). “You Only Look once: Unified, Real-Time Object Detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 779–788. doi:10.1109/cvpr.2016.91
Reimer, J., McGinley, M. J., Liu, Y., Rodenkirch, C., Wang, Q., McCormick, D. A., et al. (2016). Pupil Fluctuations Track Rapid Changes in Adrenergic and Cholinergic Activity in Cortex. Nat. Commun. 7, 1–7. doi:10.1038/ncomms13289
Reingold, E. M., Reichle, E. D., Glaholt, M. G., and Sheridan, H. (2012). Direct Lexical Control of Eye Movements in reading: Evidence from a Survival Analysis of Fixation Durations. Cogn. Psychol. 65, 177–206. doi:10.1016/j.cogpsych.2012.03.001
Ren, S., He, K., Girshick, R., and Sun, J. (2015). Faster R-Cnn: Towards Real-Time Object Detection with Region Proposal Networks. arXiv preprint arXiv:1506.01497.
Renner, P., and Pfeiffer, T. (2017). “Attention Guiding Techniques Using Peripheral Vision and Eye Tracking for Feedback in Augmented-Reality-Based Assistance Systems,” in 2017 IEEE Symposium on 3D User Interfaces (3DUI) (IEEE), 186–194. doi:10.1109/3dui.2017.7893338
Reppert, T. R., Lempert, K. M., Glimcher, P. W., and Shadmehr, R. (2015). Modulation of Saccade Vigor during Value-Based Decision Making. J. Neurosci. 35, 15369–15378. doi:10.1523/jneurosci.2621-15.2015
Rerhaye, L., Blaser, T., and Alexander, T. (2018). “Evaluation of the index of Cognitive Activity (Ica) as an Instrument to Measure Cognitive Workload under Differing Light Conditions,” in Congress of the International Ergonomics Association (Springer), 350–359. doi:10.1007/978-3-319-96059-3_38
Riggs, L. A., Kelly, J. P., Manning, K. A., and Moore, R. K. (1987). Blink-related Eye Movements. Invest. Ophthalmol. Vis. Sci. 28, 334–342.
Riggs, L. A., Ratliff, F., Cornsweet, J. C., and Cornsweet, T. N. (1953). The Disappearance of Steadily Fixated Visual Test Objects*. J. Opt. Soc. Am. 43, 495–501. doi:10.1364/josa.43.000495
Riggs, L. A., and Ratliff, F. (1951). Visual Acuity and the normal Tremor of the Eyes. Science 114 (2949), 17–18. doi:10.1126/science.114.2949.17
Robinson, D. A., Gordon, J. L., and Gordon, S. E. (1986). A Model of the Smooth Pursuit Eye Movement System. Biol. Cybern. 55, 43–57. doi:10.1007/bf00363977
Robinson, D. A. (1965). The Mechanics of Human Smooth Pursuit Eye Movement. J. Physiol. 180, 569–591. doi:10.1113/jphysiol.1965.sp007718
Rottach, K. G., Zivotofsky, A. Z., Das, V. E., Averbuch-Heller, L., Discenna, A. O., Poonyathalang, A., et al. (1996). Comparison of Horizontal, Vertical and diagonal Smooth Pursuit Eye Movements in normal Human Subjects. Vis. Res. 36, 2189–2195. doi:10.1016/0042-6989(95)00302-9
Rubaltelli, E., Agnoli, S., and Franchin, L. (2016). Sensitivity to Affective Information and Investors' Evaluation of Past Performance: An Eye-Tracking Study. J. Behav. Dec. Making 29, 295–306. doi:10.1002/bdm.1885
Rucci, M., and Poletti, M. (2015). Control and Functions of Fixational Eye Movements. Annu. Rev. Vis. Sci. 1, 499–518. doi:10.1146/annurev-vision-082114-035742
Russell, D. (2005). Consecutive and Simultaneous Interpreting. Benjamins Translation Libr. 63, 135–164. doi:10.1075/btl.63.10rus
Russo, M., Thomas, M., Thorne, D., Sing, H., Redmond, D., Rowland, L., et al. (2003). Oculomotor Impairment during Chronic Partial Sleep Deprivation. Clin. Neurophysiol. 114, 723–736. doi:10.1016/s1388-2457(03)00008-7
Rutherford, M. D., and Towns, A. M. (2008). Scan Path Differences and Similarities during Emotion Perception in Those with and without Autism Spectrum Disorders. J. Autism Dev. Disord. 38, 1371–1381. doi:10.1007/s10803-007-0525-7
Salthouse, T. A., and Ellis, C. L. (1980). Determinants of Eye-Fixation Duration. Am. J. Psychol. 93, 207–234. doi:10.2307/1422228
Salvucci, D. D., and Goldberg, J. H. (2000). “Identifying Fixations and Saccades in Eye-Tracking Protocols,” in Proceedings of the 2000 Symposium on Eye Tracking Research & Applications, New York, NY, USA (Palm Beach Gardens, FL: Association for Computing Machinery), ETRA ’00), 71–78. doi:10.1145/355017.355028
Schiller, P. H., True, S. D., and Conway, J. L. (1980). Deficits in Eye Movements Following Frontal Eye-Field and superior Colliculus Ablations. J. Neurophysiol. 44, 1175–1189. doi:10.1152/jn.1980.44.6.1175
Schoonahd, J. W., Gould, J. D., and Miller, L. A. (1973). Studies of Visual Inspection. Ergonomics 16, 365–379. doi:10.1080/00140137308924528
Schwalm, M., and Rosales Jubal, E. (2017). Back to Pupillometry: How Cortical Network State Fluctuations Tracked by Pupil Dynamics Could Explain Neural Signal Variability in Human Cognitive Neuroscience. Eneuro 4. doi:10.1523/ENEURO.0293-16.2017
Schwiedrzik, C. M., and Sudmann, S. S. (2020). Pupil Diameter Tracks Statistical Structure in the Environment to Increase Visual Sensitivity. J. Neurosci. 40, 4565–4575. doi:10.1523/jneurosci.0216-20.2020
Seeber, K. G., and Kerzel, D. (2012). Cognitive Load in Simultaneous Interpreting: Model Meets Data. Int. J. Bilingualism 16, 228–242. doi:10.1177/1367006911402982
Semmelmann, K., and Weigelt, S. (2018). Online Webcam-Based Eye Tracking in Cognitive Science: A First Look. Behav. Res. 50, 451–465. doi:10.3758/s13428-017-0913-7
Sewell, W., and Komogortsev, O. (2010). “Real-time Eye Gaze Tracking with an Unmodified Commodity Webcam Employing a Neural Network,” in CHI’10 Extended Abstracts on Human Factors in Computing Systems, 3739–3744. doi:10.1145/1753846.1754048
Shannon, C. E. (1948). A Mathematical Theory of Communication. Bell Syst. Tech. J. 27, 379–423. doi:10.1002/j.1538-7305.1948.tb01338.x
Shiferaw, B., Downey, L., and Crewther, D. (2019). A Review of Gaze Entropy as a Measure of Visual Scanning Efficiency. Neurosci. Biobehavioral Rev. 96, 353–366. doi:10.1016/j.neubiorev.2018.12.007
Shin, Y. S., Chang, W.-d., Park, J., Im, C.-H., Lee, S. I., Kim, I. Y., et al. (2015). Correlation between Inter-blink Interval and Episodic Encoding during Movie Watching. PloS one 10, e0141242. doi:10.1371/journal.pone.0141242
Simons, D. J., and Chabris, C. F. (1999). Gorillas in Our Midst: Sustained Inattentional Blindness for Dynamic Events. perception 28, 1059–1074. doi:10.1068/p2952
Snider, J., Spence, R. J., Engler, A.-M., Moran, R., Hacker, S., Chukoskie, L., et al. (2021). Distraction “Hangover”: Characterization of the Delayed Return to Baseline Driving Risk after Distracting Behaviors. Thousand Oaks, CA: Human factors, 00187208211012218.
Spering, M., and Gegenfurtner, K. R. (2007). Contextual Effects on Smooth-Pursuit Eye Movements. J. Neurophysiol. 97, 1353–1367. doi:10.1152/jn.01087.2006
Stapel, J., El Hassnaoui, M., and Happee, R. (2020). Measuring Driver Perception: Combining Eye-Tracking and Automated Road Scene Perception. Thousand Oaks, CA: Human factors, 0018720820959958.
Staub, A., and Benatar, A. (2013). Individual Differences in Fixation Duration Distributions in reading. Psychon. Bull. Rev. 20, 1304–1311. doi:10.3758/s13423-013-0444-x
Steinman, R. M., Haddad, G. M., Skavenski, A. A., and Wyman, D. (1973). Miniature Eye Movement. Science 181, 810–819. doi:10.1126/science.181.4102.810
Stern, J. A., Walrath, L. C., and Goldstein, R. (1984). The Endogenous Eyeblink. Psychophysiology 21, 22–33. doi:10.1111/j.1469-8986.1984.tb02312.x
Stevenson, S. B., Volkmann, F. C., Kelly, J. P., and Riggs, L. A. (1986). Dependence of Visual Suppression on the Amplitudes of Saccades and Blinks. Vis. Res. 26, 1815–1824. doi:10.1016/0042-6989(86)90133-1
Sun, F., Tauchi, P., and Stark, L. (1983). Dynamic Pupillary Response Controlled by the Pupil Size Effect. Exp. Neurol. 82, 313–324. doi:10.1016/0014-4886(83)90404-1
Sweller, J. (2011). “Cognitive Load Theory,” in Psychology of Learning and Motivation (Elsevier), 55, 37–76. doi:10.1016/b978-0-12-387691-1.00002-8
Takeda, K., and Funahashi, S. (2002). Prefrontal Task-Related Activity Representing Visual Cue Location or Saccade Direction in Spatial Working Memory Tasks. J. Neurophysiol. 87, 567–588. doi:10.1152/jn.00249.2001
Termsarasab, P., Thammongkolchai, T., Rucker, J. C., and Frucht, S. J. (2015). The Diagnostic Value of Saccades in Movement Disorder Patients: a Practical Guide and Review. J. Clin. Mov Disord. 2, 14–10. doi:10.1186/s40734-015-0025-4
Tullis, T., and Albert, B. (2013). Measuring the User Experience. Elsevier, 163–186. chap. Behavioral and physiological metrics. doi:10.1016/b978-0-12-415781-1.00007-8Behavioral and Physiological Metrics
Valls-Sole, J. (2019). Spontaneous, Voluntary, and Reflex Blinking in Clinical Practice. J. Clin. Neurophysiol. 36, 415–421. doi:10.1097/wnp.0000000000000561
Van der Stigchel, S., Rommelse, N. N. J., Deijen, J. B., Geldof, C. J. A., Witlox, J., Oosterlaan, J., et al. (2007). Oculomotor Capture in Adhd. Cogn. Neuropsychol. 24, 535–549. doi:10.1080/02643290701523546
Van Orden, K. F., Jung, T.-P., and Makeig, S. (2000). Combined Eye Activity Measures Accurately Estimate Changes in Sustained Visual Task Performance. Biol. Psychol. 52, 221–240. doi:10.1016/s0301-0511(99)00043-5
van Tricht, M. J., Nieman, D. H., Bour, L. J., Boerée, T., Koelman, J. H. T. M., de Haan, L., et al. (2010). Increased Saccadic Rate during Smooth Pursuit Eye Movements in Patients at Ultra High Risk for Developing a Psychosis. Brain Cogn. 73, 215–221. doi:10.1016/j.bandc.2010.05.005
van Zoest, W., Donk, M., and Theeuwes, J. (2004). The Role of Stimulus-Driven and Goal-Driven Control in Saccadic Visual Selection. J. Exp. Psychol. Hum. perception Perform. 30, 746–759. doi:10.1037/0096-1523.30.4.749
Vandeberg, L., Bouwmeester, S., Bocanegra, B. R., and Zwaan, R. A. (2013). Detecting Cognitive Interactions through Eye Movement Transitions. J. Mem. Lang. 69, 445–460. doi:10.1016/j.jml.2013.05.006
Velichkovsky, B. B., Khromov, N., Korotin, A., Burnaev, E., and Somov, A. (2019). “Visual Fixations Duration as an Indicator of Skill Level in Esports,” in IFIP Conference on Human-Computer Interaction (Springer), 397–405. doi:10.1007/978-3-030-29381-9_25
Velichkovsky, B. M., Dornhoefer, S. M., Pannasch, S., and Unema, P. J. (2000). “Visual Fixations and Level of Attentional Processing,” in Proceedings of the 2000 symposium on eye tracking research & applications, 79–85. doi:10.1145/355017.355029
Venjakob, A., Marnitz, T., Mahler, J., Sechelmann, S., and Roetting, M. (2012). “Radiologists’ Eye Gaze when reading Cranial Ct Images,” in Medical imaging 2012: Image perception, observer performance, and technology assessment (San Diego: International Society for Optics and Photonics), 8318, 83180B. doi:10.1117/12.913611
Wade, M. G., and Jones, G. (1997). The Role of Vision and Spatial Orientation in the Maintenance of Posture. Phys. Ther. 77, 619–628. doi:10.1093/ptj/77.6.619
Wainstein, G., Rojas-Líbano, D., Crossley, N. A., Carrasco, X., Aboitiz, F., and Ossandón, T. (2017). Pupil Size Tracks Attentional Performance in Attention-Deficit/hyperactivity Disorder. Sci. Rep. 7, 8228–8229. doi:10.1038/s41598-017-08246-w
Walker, R., McSorley, E., and Haggard, P. (2006). The Control of Saccade Trajectories: Direction of Curvature Depends on Prior Knowledge of Target Location and Saccade Latency. Perception & Psychophysics 68, 129–138. doi:10.3758/bf03193663
Walker-Smith, G. J., Gale, A. G., and Findlay, J. M. (1977). Eye Movement Strategies Involved in Face Perception. Perception 6, 313–326. doi:10.1068/p060313
Wang, C.-A., Brien, D. C., and Munoz, D. P. (2015). Pupil Size Reveals Preparatory Processes in the Generation of Pro-saccades and Anti-saccades. Eur. J. Neurosci. 41, 1102–1110. doi:10.1111/ejn.12883
Wang, R. I., Pelfrey, B., Duchowski, A. T., and House, D. H. (2012). “Online Gaze Disparity via Bioncular Eye Tracking on Stereoscopic Displays,” in 2012 Second International Conference on 3D Imaging, Modeling, Processing (Visualization & TransmissionIEEE), 184–191. doi:10.1109/3dimpvt.2012.37
Wang, Y., Lu, S., and Harter, D. (2021). Multi-sensor Eye-Tracking Systems and Tools for Capturing Student Attention and Understanding Engagement in Learning: A Review. IEEE Sensors J. 21, 22402–22413. doi:10.1109/jsen.2021.3105706
Warren, D. E., Thurtell, M. J., Carroll, J. N., and Wall, M. (2013). Perimetric Evaluation of Saccadic Latency, Saccadic Accuracy, and Visual Threshold for Peripheral Visual Stimuli in Young Compared with Older Adults. Invest. Ophthalmol. Vis. Sci. 54, 5778–5787. doi:10.1167/iovs.13-12032
Wass, S. V., de Barbaro, K., and Clackson, K. (2015). Tonic and Phasic Co-variation of Peripheral Arousal Indices in Infants. Biol. Psychol. 111, 26–39. doi:10.1016/j.biopsycho.2015.08.006
Watson, A. B., and Yellott, J. I. (2012). A Unified Formula for Light-Adapted Pupil Size. J. Vis. 12, 12. doi:10.1167/12.10.12
Wedel, M., and Pieters, R. (2008). A Review of Eye-Tracking Research in Marketing. Rev. marketing Res., 123–147. doi:10.4324/9781351550932-5
Widdel, H. (1984). “Operational Problems in Analysing Eye Movements,” in Advances in Psychology (Elsevier), 22, 21–29. doi:10.1016/s0166-4115(08)61814-2
Wierda, S. M., van Rijn, H., Taatgen, N. A., and Martens, S. (2012). Pupil Dilation Deconvolution Reveals the Dynamics of Attention at High Temporal Resolution. Proc. Natl. Acad. Sci. 109, 8456–8460. doi:10.1073/pnas.1201858109
Xu-Wilson, M., Zee, D. S., and Shadmehr, R. (2009). The Intrinsic Value of Visual Information Affects Saccade Velocities. Exp. Brain Res. 196, 475–481. doi:10.1007/s00221-009-1879-1
Young, L. R. (1971). “Pursuit Eye Tracking Movements,” in The Control of Eye Movements (New York: Academic Press), 429–443. doi:10.1016/b978-0-12-071050-8.50019-7
Young, L. R., and Sheena, D. (1975). Survey of Eye Movement Recording Methods. Behav. Res. Methods Instrumentation 7, 397–429. doi:10.3758/bf03201553
Yu, G., Xu, B., Zhao, Y., Zhang, B., Yang, M., Kan, J. Y. Y., et al. (2016). Microsaccade Direction Reflects the Economic Value of Potential Saccade Goals and Predicts Saccade Choice. J. Neurophysiol. 115, 741–751. doi:10.1152/jn.00987.2015
Zackon, D. H., and Sharpe, J. A. (1987). Smooth Pursuit in senescence:Effects of Target Acceleration and Velocity. Acta oto-laryngologica 104, 290–297. doi:10.3109/00016488709107331
Zelinsky, G. J., and Sheinberg, D. L. (1997). Eye Movements during Parallel-Serial Visual Search. J. Exp. Psychol. Hum. perception Perform. 23, 244–262. doi:10.1037/0096-1523.23.1.244
Zénon, A. (2017). Time-domain Analysis for Extracting Fast-Paced Pupil Responses. Sci. Rep. 7, 1–10. doi:10.1038/srep41484
Zhang, X., Yuan, S.-M., Chen, M.-D., and Liu, X. (2018). A Complete System for Analysis of Video Lecture Based on Eye Tracking. IEEE Access 6, 49056–49066. doi:10.1109/access.2018.2865754
Keywords: eye tracking, pupillometry, visual perception, cognition, attention
Citation: Mahanama B, Jayawardana Y, Rengarajan S, Jayawardena G, Chukoskie L, Snider J and Jayarathna S (2022) Eye Movement and Pupil Measures: A Review. Front. Comput. Sci. 3:733531. doi: 10.3389/fcomp.2021.733531
Received: 30 June 2021; Accepted: 23 November 2021;
Published: 11 January 2022.
Edited by:
Kostas Karpouzis, Panteion University, GreeceReviewed by:
Ilias Maglogiannis, University of Piraeus, GreeceMichael J Proulx, University of Bath, United Kingdom
Copyright © 2022 Mahanama, Jayawardana, Rengarajan, Jayawardena, Chukoskie, Snider and Jayarathna. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Sampath Jayarathna, sampath@cs.odu.edu