- Vision and Control of Action (VISCA) Group, Department of Cognition, Development and Psychology of Education, Institut de Neurociències, Universitat de Barcelona, Barcelona, Spain
Catching a ball in a parabolic flight is a complex task in which the time and area of interception are strongly coupled, making interception possible for a short period. Although this makes the estimation of time-to-contact (TTC) from visual information in parabolic trajectories very useful, previous attempts to explain our precision in interceptive tasks circumvent the need to estimate TTC to guide our action. Obtaining TTC from optical variables alone in parabolic trajectories would imply very complex transformations from 2D retinal images to a 3D layout. We propose based on previous work and show by using simulations that exploiting prior distributions of gravity and known physical size makes these transformations much simpler, enabling predictive capacities from minimal early visual information. Optical information is inherently ambiguous, and therefore, it is necessary to explain how these prior distributions generate predictions. Here is where the role of prior information comes into play: it could help to interpret and calibrate visual information to yield meaningful predictions of the remaining TTC. The objective of this work is: (1) to describe the primary sources of information available to the observer in parabolic trajectories; (2) unveil how prior information can be used to disambiguate the sources of visual information within a Bayesian encoding-decoding framework; (3) show that such predictions might be robust against complex dynamic environments; and (4) indicate future lines of research to scrutinize the role of prior knowledge calibrating visual information and prediction for action control.
Introduction
Intercepting a ball in a parabolic trajectory before reaching ground level is a fundamental task in different sports: batting a baseball, hitting a high lob in tennis, or heading a football. In those situations, the time at which the interception is possible is very tight, yet our performance is astonishing. Time-to-contact (from now on TTC), that is, the time until an object reaches a location of interest, can provide very useful information that would help anticipate motor programs to solve those tasks.
In principle, to intercept a target, it would be enough to estimate its position and predict its future position based on speed estimates. Solutions based on this idea have been put forward for 2D motion (Kwon et al., 2015; Aguilar-Lleyda et al., 2018) but the generalization to 3D parabolic trajectories faces complex problems deeply rooted in the inverse-projection problem of Perception. The inverse problem of Perception refers to the ambiguous mapping between distal stimuli and final percept (Pizlo, 2001; Kersten et al., 2004). Unlike previous attempts where TTC is obtained from optical variables, in this article, we propose that some constants in the environment like gravity and size are considered and ease the otherwise complex transformation of optical variables to a 3D world to obtain relevant variables like TTC. The stance taken in this work will assume that we make implicit inferences (Von Helmholtz, 1867) about the present and future states of the world to act. However, the nature of the information guiding the control of action is an ongoing source of debate within the study of Perception.
Two Theories for Interceptive Control
Information-Based Control
The information-based control perspective, rooted in the Ecological or Gibsonian framework of Psychology (Gibson, 1966, 1979; Figure 1A), assumes that perceptual information is governed by certain physical regularities (Turvey et al., 1981) that can be captured and exploited to control our action. Under ecologically valid conditions (i.e., full-cue conditions), our perceptual system would be attuned to perceptual invariants directly specifying the characteristics of an event without the need to perform internal computations according to which humans act (Gibson, 1979). Thus, the information to solve a given task is directly specified in the optic flow (direct perception) explaining why only identifying task-relevant visual variables will determine an actor’s successful action from a perceptual perspective.
Figure 1. (A) Optic flow conforms to invariants that specify properties of the environment (direct perception), indicating the adequacy and availability of action within the task. (B) Sensory stimulation is combined with prior information to infer current or future states in the environment (read-out), providing the grounds to plan and adapt action.
Under this framework, mainstream interpretations argue that the role of the observer is to actively seek out invariants within certain task-relevant pieces of optic information and unfold a coupled action based on instantaneous information. Following this line, detecting and maintaining invariant stimulation requires reducing the difference or the error with an “ideal value.” Because of that, these strategies are also called error-nulling strategies (Fajen, 2005b). Based on this idea, different control laws have been proposed for visually guided actions such as intercepting a moving object (Warren et al., 2001; Wilkie and Wann, 2003; Bruggeman et al., 2007; Zhao et al., 2019), braking (Lee, 1976) or catching a ball on the fly (Chapman, 1968; Michaels and Oudejans, 1992; McLeod and Dienes, 1993; McBeath et al., 1995).
Among the previous examples, catching a ball in a parabolic flight is a paradigmatic interception case including locomotion and manual interceptive phases. The study of the underlying mechanisms regarding the locomotive phase is usually referred to as the outfielder problem (Todd, 1981). The outfielder problem studies a case of interception in which baseball players known as outfielders must move to catch a high-flying ball in a parabolic trajectory before it hits the ground (Chapman, 1968; Todd, 1981; Michaels and Oudejans, 1992).
The first catching error-nulling strategy put forward to explain action control within the outfielder problem was Chapman’s strategy (Chapman, 1968). S. Chapman noticed that when a ball follows a parabolic trajectory on a collision course with the observer, the elevation angle (γ; see Figure 2), that is, the vertical angle between the ball’s position and an observer’s eye level, increases during the whole trajectory. Therefore, all an observer would need to do to navigate towards the interception location is keeping the elevation angle increasing during the whole trajectory at a constant rate. However, others suggested it should increase at a decreasing rate (McLeod and Dienes, 1993, 1996).
Figure 2. Lateral view of a parabolic trajectory depicting the primary primitive monocular cues, that is, retinal size (green projection; θ) and elevation angle (orange projection; γ).
Later, Michaels and Oudejans (1992) described Chapman’s strategy in terms of the projected image in the vertical plane at launch distance. If the ball is in a collision course with the viewer, the vertical projection of the ball increases linearly through the trajectory. In any other case, the image of the ball would displace non-linearly, that is, accelerated. Therefore, to catch a ball in a parabolic trajectory, one needs to actively maintain the acceleration of the projected vertical position of the ball at zero. Motivated by that, this strategy was named Optic Acceleration Cancellation (from now on, OAC) in McBeath et al. (1995).
Although the error-nulling strategies emerged within the ecological framework, they conflict with a key concept at the core of the Ecological theory, the theory of affordances (Gibson, 1979). The affordance-based theory emphasizes the idea that observers are tuned to the availability of an action given a sensory array. This tuning would be a by-product of a gauging process that maps optic into movement information and even into optic correlates in object size units (Peper et al., 1994; Jacobs and Michaels, 2006). Hence, if a fielder is correctly calibrated, acting to keep certain variables of interest into a “safe region” would ensure interception. This notion provides the grounds for an affordance-based control strategy theory (Fajen, 2007).
Reformulating the OAC under the scope of the affordance-based control, either canceling out vertical acceleration or running at maximum speed without being able to cancel out vertical acceleration, would be required to perceive catchability. However, a series of studies (Postma et al., 2017, 2018) found that actual catchers did not need to cancel out acceleration nor run at their maximum speed to judge catchability. These results cast doubts on the informational nature of a catchability affordance.
Despite a lesser dependence on immediate visual information, an affordance-based control strategy is still dependent on instantaneous visual information. In this respect, simulations of the locomotion based on error-nulling strategies were irreconcilable with actual catches when accounting for human neuromotor acceleration and sensorimotor integration delays (McLeod and Dienes, 1996; Kistemaker et al., 2006, 2009). Consequently, a minimal prediction seems necessary to account for sensorimotor delays in the central nervous system (Nijhawan, 1994). Furthermore, the occlusion of visual information would result in a considerable impairment of the action. A possible solution to provide adequate responses would be to continue to do what has been done so far (Bootsma, 2009). Nevertheless, in temporally constrained tasks, another question arises: for how long? In the outfielder problem, catching the ball in-flight imposes tight temporal restrictions constraining the interception area. Hence, predicting parameters such as TTC or the interception area in dynamic contexts is key to planning our actions while looking around for a teammate, finding a safe path towards the goal, or modulating speed to reach the interception location in time. Following this reasoning, some alternatives have been proposed in the literature as solutions based on predictions about future states of the world.
Model-Based Control
The model-based control is framed within the constructivist framework (Von Helmholtz, 1867). It assumes that the information picked up by our senses is inherently ambiguous, to some degree corrupted by noise in the neural system and delayed at higher-order brain areas. Craik (1967) proposed that the brain tries to infer and replicate an external world model given the available sensory information. This replica results in an internal model of the environment, including an agent’s state that allows one to predict future states of the world and act accordingly avoiding sensorimotor delays.
Relying on predictions would allow us to divert our gaze from the immediate region of interest and consequently interrupt the sensory flow. For example, Hayhoe et al. (2005) and Diaz et al. (2013) showed anticipatory saccades towards future interest points to plan future goal states. The same applies to catching a ball in a parabolic trajectory for manual interception tasks. Despite distractors, parallel tasks, occlusions or head-turns that might divert our attention, we still manage to intercept a ball in flight (Dessing et al., 2009; López-Moliner et al., 2010; López-Moliner and Brenner, 2016; Binaee and Diaz, 2019) In fact, we can hit it even when the ball was visible for just a short time (Sharp and Whiting, 1974; Whiting and Sharp, 1974; Amazeen et al., 1999), revealing that actually, the major constraint for the use of a predictive strategy would be to obtain predictions early enough to overcome sensory-motor delays.
The trajectory prediction strategy (Saxberg, 1987a,b), framed within the model-based theory, assumes that an observer predicts where and when the ball will be within reach in Cartesian units, allowing to pre-program a minimal action plan since motion onset. However, the available optic information is egocentric and therefore ambiguous with respect to its source. Therefore, an observer must perform an inferential process to interpret optic information accurately.
In this line, Perception has been proposed as a Bayesian inferential process in which visual information is interpreted as a function of the most probable state of the world given prior knowledge. This inferential process has been formulated in terms of “encoding” and “decoding” (Knill and Pouget, 2004; Friston, 2010; Wei and Stocker, 2015). Encoding corresponds to the activity resulting from the transduction of external energy onto the sensory receptors. The encoded sensory information is then combined with prior knowledge through an inferential process called decoding (Figure 1B). The product of the decoding is an interpretation (read-out) of the currently available data resulting in a belief of the state of the world that provides the grounds to draw predictions.
In real life, we generally do not judge the parameters of a task for ambiguous targets in the environment. Instead, we have some prior knowledge of the elements to be judged that are stable and might help disambiguate optic information providing the grounds to extract valuable information for the task. In this line, a significant number of works highlighted the role of prior knowledge about contextual variables such as gravitational acceleration (McIntyre et al., 2001; Jörges and López-Moliner, 2017) or known/familiar size (López-Moliner et al., 2007; Hosking and Crassini, 2010), framing the interpretation of visual information for the control of timed actions.
Note that using a priori knowledge does not imply the availability of accurate Cartesian metrics or Newtonian laws within an internal model. A fully-featured 3D internal model replicating the external world has been repeatedly dismissed. For example, Shaffer and McBeath (2005) showed that even expert baseball players could not judge the apex of a ballistic trajectory on a collision course with the observer. In this situation, the apex was estimated to be 0.33 s before collision for flight durations of 4 s, that is, 1.66 s after the actual apex. These results indicate a tendency to judge the apex of the elevation angle as the physical apex of the trajectory. Also, Reed et al. (2010) showed that neither expert nor novel baseball players could reconstruct the visual trajectory of a parabolic trajectory in a head-on approach mixing up the ball’s movement in space with the visual trajectory it follows.
Unlike Craik (1967), however, we propose that the prior knowledge can be kept to a minimal number of components that help exploit the optic flow’s complexity. Thus, in line with a Bayesian framework, under our view, the use of prior knowledge just suggests the existence of a probabilistic and implicit knowledge acquired by repeated experience that helps infer the most probable sources of visual information in the external world (Zago et al., 2008; Gómez and López-Moliner, 2013). In this sense, an accurate representation of a priori parameters would suffice to obtain reliable estimates of the task’s parameters.
Here we propose using priors as internalized knowledge to translate optic variables into temporal estimates in a process we name calibration. Calibration would be the process by which optical or angular information is mapped into Cartesian ones with the assistance of different pieces of prior knowledge providing actionable predictions (López-Moliner et al., 2007, 2013). Calibrating optic cues into Cartesian allows us to test the correspondence between the prediction of a model and our actions in terms of accuracy. Furthermore, it may allow us to formulate a hypothesis based on known psychophysical precision levels for the different pieces of information in the sensory array and even check if integration rules apply (Wolpert et al., 1995; de la Malla and López-Moliner, 2015).
As an example, the GS model (Gómez and López-Moliner, 2013) is an algorithm that predicts the TTC for parabolic trajectories based on a combination of optic variables and prior knowledge information. Its predictions have been partially validated based on predictions about the accuracy and the precision of temporal estimations (de la Malla and López-Moliner, 2015; Aguado and López-Moliner, 2021). However, there is a lack of mathematical formulation to predict the interception location. This makes it harder to test experimentally predictive control strategies within the outfielder problem. Hence, in this work, we will limit ourselves to indicate the role of TTC estimates guiding the interceptive action.
It is essential to mention here that our definition of calibration is different from the definition of calibration made within the Ecologic framework (Fajen, 2005a; Jacobs and Michaels, 2007). Under our perspective, calibration is a process by which otherwise ambiguous optical information is directly mapped into kinematic and temporal estimates such as motion vectors or TTC highlighting the relevance of prior knowledge to provide predictions that may assist visually guided actions. In contrast, within the Ecological framework, calibration would be a by-product of a gauging process that maps visual information into movement information and even into optic correlates in object size units (Peper et al., 1994; Jacobs and Michaels, 2006).
Nevertheless, producing predictions does not necessarily mean that those predictions will be accurate or that the new visual information would be disregarded. Take the case of Fink et al. (2009) study. Participants had to catch a ball in a parabolic trajectory that suddenly would alter its motion towards the ground. As a reaction, the catchers changed their trajectory towards the ball as well which was taken as support for the information-based control perspective. Under Fink et al. (2009) rationale, a model-based control strategy would result in a consistent path towards the interception point despite mid-flight disturbances. This rationale assumes that new information would be dismissed or might be irrelevant because the prediction would remain invariant. However, predictions would also be subject to continuous evaluations to avoid errors or perceptually driven biases. In a similar line, Postma et al. (2014) reasoned that continuously gazing the ball through the trajectory would support the information-based control perspective. However, following the ball with our gaze does not necessarily imply that action guidance must be driven by instantaneous optic information. Periodically sampling visual input to correct the prediction made would be an alternative strategy to guide action with reasonable levels of accuracy in a more general framework (Brenner and Smeets, 2018). In fact, a simulation study conducted by Belousov et al. (2016) showed that predictive behavior would be indistinguishable from using error-nulling strategies if the ball is continuously monitored.
In the following sections, we will unveil how the interaction between perceptual information and prior knowledge contributes to interpreting and reliably predict TTC for gravitationally accelerated objects under parabolic trajectories. To do so, we will first analyze the available sources of visual information within the optic flow to judge TTC or time an interceptive action. Then, we will stress the role of prior information in calibrating visual information. After that, we will show the accuracy and reliability of the GS model which includes known gravity and size in complex environments. Finally, we will indicate future lines of research to address the role of predictions of TTC guiding interceptive behavior.
Available Visual Information
When an observer faces a ball in a parabolic trajectory, the projectile describes the following sequence of events (see Figure 2). The ball initially goes up at a decreasing speed until it reaches the peak of its trajectory. Then, it accelerates during the descent towards the ground. However, an observer cannot access the underlying dynamics of projectile motion using Cartesian metrics. Instead, they only have access to information based on egocentric angular variables that depend on their position and the kinematics of the ball (Shaffer and McBeath, 2005; McBeath et al., 2018). Hence, the ball’s kinematics and an observer’s movement influence the visual information being exposed throughout the trajectory. Figure 2 depicts the main primitive optic variables that we will consider and how both unfold over time (t) depending on the observer’s position and total flight time (T), which is unknown to the observer.
While interpreting sensory information is a central part of this work, we also need to consider the limitations of our sensory system to gather that visual information. Because of that, in the following subsections, we will elaborate on the conditions that render different visual cues helpful regarding detectability or discriminability, describing their precision as Weber fractions. To do so, we will assume that the observers keep their gaze on the projectile, which is usually the case in free viewing situations (Oudejans et al., 1999; Postma et al., 2014) and laboratory-controlled tasks when the position to hit a target is not pre-specified (Brenner and Smeets, 2007; Soechting et al., 2009; Cámara et al., 2018).
Retinal Size (θ)
The visual angle or retinal size is the angular size projected by an object on the retina (θ; see green projections in Figure 2). Retinal size (θ) is proportional to both object size and distance, being the prototypical example of an ambiguous optic variable. Previous studies have shown that human performance in discrimination tasks for angular size judgments is about a 3–6% Weber fraction (WF; kθ) for objects yielding >0.0009 radians, increasing steeply up to a 20–30% (WF) for smaller objects (Westheimer and McKee, 1977; Klein and Levi, 1987; McKee and Welch, 1992).
Retinal size is a zero-order variable; that is, it does not carry temporal information and thus, cannot be employed alone to estimate motion components or TTC. To do so, one needs to have access to the rate of expansion (), which is the speed at which the retinal image changes. The absolute detection threshold for has been reported to be about 0.0003 rad/s (McKee, 1981), while the discrimination threshold associated with this parameter is about an 8.5–14% of change (WF; k; Regan and Hamstra, 1993).
In some cases, such as baseball games, the players meet scenarios where the ball is at a considerable distance. In those cases, although mediated by physical size and ball’s horizontal speed, discriminability of the rate of expansion () is generally poor (20% of change (WF) for objects expanding at a rate of less than 0.004 rad/s (Regan and Beverley, 1978; Harris and Watamaniuk, 1995).
To show the effect of ball size on an observer’s ability to discriminate differences in the rate of expansion (), we computed the rate of expansion for two different balls (baseball and soccer balls) moving in a parabolic trajectory towards an observer from different initial distances. Figure 3 shows how retinal expansion unfolds as a function of time. Values below 0.004 rad/s (red dashed line) would fall below the optimal discriminability range, indicating that the observer’s ability to discriminate differences is inferior. As depicted in Figure 3, a player facing a baseball will not discriminate retinal expansion during more than half of the trajectory. Instead, differences in retinal expansion can be discriminated during most of the trajectory of the Soccer ball. This example aims to point out that even when visual cues are present in the optic flow, the resolution of our visual system might not allow us to use them to guide our actions. Therefore, initial estimates of TTC might be computed using alternative routes.
Figure 3. Rate of expansion () for two different ball sizes at five different initial distances. The total flight time is 2 s. Values under the red dashed line (0.004 rad/s) indicate that an observer cannot discriminate differences.
Tau (τ)
Lee’s seminal work (Lee, 1976) described Tau (from now on, τ) as the ratio between the visual angle (θ) and its rate of expansion (; for a review, see Hecht and Savelsbergh, 2004). Tau signals TTC and is directly accessible within the optic flow without prior knowledge or previous estimates of distance, size, or approaching speed.
Although τ can not be conceptualized as a primitive variable in the study of visual cues, it has shown different features that could allow us to consider it as such. Regan and Hamstra (1993) found that differences in τ could be distinguished independently of differences in θ or (kτ ≈ 0.07:0.13; WF). Because of this, Regan and Hamstra (1993) concluded that there might exist a mechanism sensitive to τ independently of θ or . Indeed, other studies have shown the existence of a neural mechanism tuned to τ (independently of retinal size and rate of expansion) or some of its modifications [such as the η-function (Judge and Rind, 1997) or τ_m-function (Keil and López-Moliner, 2012)] in various species such as pigeons and humans (Yonas et al., 1977; Sun and Frost, 1998; Rind and Simmons, 1999).
Tau (τ) has been indicated as a source of prospective information that might be used as a threshold (as a criterion) to perform different tasks such as hitting (Lee et al., 1983; Bootsma and van Wieringen, 1990) or catching (Savelsbergh et al., 1991). This threshold is usually referred to as Tau-margin (Wann, 1996). However, its applicability to time parabolic trajectories might be compromised due to several limitations.
First, τ would only generate accurate TTC predictions at launch when the ball travels in a collision course with the observer.
Second, the object should approach the observer at a constant speed. In a parabolic trajectory, from an allocentric viewpoint, Vx and Vz are constant (assuming no air resistance). Nevertheless, the approaching speed for the observer corresponds to radial velocity (Vr), which would carry the isotropic expansion of the retinal expansion. However, Vr is not constant through the trajectory and cannot be directly estimated from the optic flow (Gómez and López-Moliner, 2013). For instance, consider a trajectory launched 2 m far from an observer in a trajectory 10 m of height. In this situation, during the first half of the trajectory, Vr and retinal expansion would be negative; that is, the object moves further away from the observer, rendering meaningless estimates of TTC using Tau.
Third, even though Tau could be discriminated independently of the rate of expansion (), it is likely constrained to the same detection thresholds (Keil and López-Moliner, 2012). Therefore, rates of expansion () lower than 0.004 rad/s could result in a non-informative source to guide temporal estimations during an important section of parabolic trajectories. As reported above, the WF for Tau ranges between 7% and 13%. Therefore, we will be using a mean WF of 10% referring to Tau in the following sections.
Finally, it might not be directly implemented as a general-purpose mechanism because the object must be spherical and rigid, which is not the case for an interception in some sports such as Rugby or Frisbee.
Elevation Angle (γ)
The elevation angle (γ) is the position of an object in the vertical meridian of the retina (see orange projection in Figure 2. We conceptualize the elevation angle (γ) using a spherical projection because angular variables are assumed to be directly accessible to an observer (McBeath et al., 2018).
Even though it is often referred to as the vertical position of an object in the retina, when a projectile is visible, we tend to perform continuous visual follow-ups foveating the object, which might prevent perceptual biases (de la Malla et al., 2017). In those cases, the retinal angle of the elevation angle (γ) would tend to zero. Because of that, it has been suggested that this visual angle can be estimated as a combination of the displacement of the environment in the retina (Oudejans et al., 1999; Brenner and Smeets, 2015b), the movement of the eyes with respect to the observation axis (Crowell and Banks, 1996) and estimates of the heading angle generated at the otoliths of the vestibular system (Berthoz, 2000; Roy and Cullen, 2003) produced by the movement of the head and trunk (Crowell et al., 1998; Lewis et al., 1998). Previous literature has found that foveated objects require a difference of up to a 3–5% of change (WF) to be effectively discriminated (Regan and Kaushal, 1994; Crowell and Banks, 1996).
The first derivative with the time of the rate of change of the elevation angle () is the vertical rate of displacement of a target in the retina (see Figure 2). According to several studies (McKee, 1981; Orban et al., 1984; de Bruyn and Orban, 1988), the ability to judge differences in is about 5% (WF) for angular velocities between 0.03 rad/s (1.71 deg/s) and 1.2 rad/s (69 deg/s). Interestingly, Portfors-Yeomans and Regan (1996) suggest channels that process position and cardinal motion independently, which indicates that the noise for both estimates is independent.
Given that parabolic trajectories move accelerated by terrestrial gravity, it is reasonable to consider humans’ ability to detect acceleration. Calderone and Kaiser (1989) proposed that acceleration in the visual system can be studied as the rate of change in speed divided by the average object speed in a two-stage process carried out in about 200 ms (Werkhoven et al., 1992; Zaal et al., 2012). This delay would mean that the observer would not continuously monitor the adequacy of their actions. Furthermore, some studies found that it is necessary at least ~20% of the change in speed to detect acceleration (Gottsdanker et al., 1961; Werkhoven et al., 1992; Babler and Dannemiller, 1993; Brouwer et al., 2006; Zaal et al., 2012) indicating that humans are quite insensitive to changes in speed.
Disparity (δ)
Disparity (δ) is the vertical/horizontal difference between the position of an object in the retinal image of both eyes. An algorithm like Tau (τ) for the estimation of TTC has been proposed as a combination of the knowledge of interocular distance (I), the distance with the ball (D) and the rate of change of horizontal disparity) () (TTC ≈ I/(D*)). Furthermore, it can be used to estimate the lateral distance at which an object would pass an observer position.
A combination of Tau and the information contained in the above expression would assist the estimation of TTC to achieve our exceptional temporal precision batting fastballs (Gray and Regan, 1998). This solution may account for systematic underestimations of TTC by weighting the visual cue that indicates a shorter TTC to guide the final interceptive phase (Savelsbergh and Whiting, 1992; Gray and Regan, 1998; Rushton and Wann, 1999). However, Brenner et al. (2014) found no evidence that hitting a free fall ball uses the rate of change in disparity () to estimate TTC. In this sense, Brenner and Smeets (2018) argue that some studies that compare the performance between monocular and binocular conditions “ignore the benefit of having two estimates of the relevant monocular cues” instead of one.
Evidence of Prior Knowledge Calibrating Visual Information
Humans quickly acquire knowledge about regularities in their interaction with the environment. Regularities such as the light coming from above (Adams et al., 2004), that bigger means heavier (Peters et al., 2016) or the fact that object size is generally constant (López-Moliner and Keil, 2012) enhance predictability and reduce uncertainty about future states of the world. Indeed, the assumption of a stable world is at the heart of essential findings in different areas of vision science, such as speed perception (Stocker and Simoncelli, 2006), depth estimation (Glennerster et al., 2006), and manual interception (Brenner and Smeets, 2015a).
Such known regularities may also be referred to here as contextual information stressing the role of acquired knowledge by repeated experience within a specific context. In this work, we will focus on two pieces of internalized knowledge that usually remain stable in our world and might frame the interpretation of visual information for parabolic trajectories: known size and gravity.
Size
The assumption of constant size is likely one of the most critical assumptions about action because, in general, the objects around us do not change size unexpectedly. Under this assumption, known size calibrates visual information into distance estimations with an object as the ratio between known size and the retinal size projected (Ittelson, 1951; Hecht et al., 1996; Sousa et al., 2011) even though irregular objects such as rugby balls and frisbees can be problematic.
The previous expression provides a mechanism to scale the optic space into ball size units (Peper et al., 1994; Gómez and López-Moliner, 2013) for a broad range of contexts. Calibrating the optic space with known size provides estimates of relative distances in paintings, pictures, video games, and environments with low or incongruent pictorial detail (Todd, 1981; Saxberg, 1987b), sometimes at the cost of leading to systematic misperceptions (Battaglia et al., 2005; Tcheang et al., 2005; Battaglia et al., 2011).
In line with the use of known size calibrating visual information, López-Moliner et al. (2007) showed that when an object approaches at a constant speed and physical size is known; an observer can exploit the lawful relations between physical size and optic variables in the equation above to estimate approaching speed (Vz) as its first derivative:
In the above expression, known size (s) allows calibrating retinal size (θ) and rate of expansion () into an estimate of approaching speed (Vz) from otherwise spatiotemporally ambiguous optic variables. Then, an observer could use single optic variables (e.g., retinal size or expansion rate) to time the initiation of interceptive actions (Smith et al., 2001; López-Moliner et al., 2007; López-Moliner and Keil, 2012) depending on noise levels (Aguilar-Lleyda et al., 2018). However, what happens when it comes to obtaining parameters of more complex tasks such as parabolic trajectories?
In Todd (1981), the participants had to judge if a ball on a parabolic trajectory would fall in front or behind an observer in different experimental conditions. Todd’s work showed that, even though there might be enough information to estimate the final position qualitatively for parabolic trajectories only based on sensory information, prior knowledge of an object’s size helped the participants to judge the final position in depth accurately. In each block, the absolute size was either fixed, selected at random, or fixed to a single dot during the whole trajectory. Accuracy was significantly better when the absolute size was fixed than the condition in which size was selected randomly. Those results indicate that prior knowledge of the ball’s size aids the estimation of motion-in-depth.
Interestingly, the third condition yielded the worst performance of all three conditions, yet performance was slightly over chance level. In this condition, only the vertical movement was available to the observers to judge approaching speed. Participants judged landing position based on “the amount of vertical speed.” In line with these results, Jörges and López-Moliner (2017) showed that prior knowledge of gravity might be essential to calibrate estimates of the rate of change () into estimates of approaching speed (Vz) in parabolic motion.
Gravity
Since Lacquaniti and Maioli (1989) work, showing an anticipatory activity for gravitationally accelerated objects, there is evidence that an internal representation of gravity may play a key role in controlling interceptive actions and judging TTC. For example, McIntyre et al. (2001) found that astronauts react to moving objects as if they were accelerated by Earth gravity under micro-gravity conditions. That study showed that although the astronauts were immersed in an environment where visual and bodily cues indicated microgravity conditions, they could not adapt their interceptive actions completely. After 15 days, the astronauts were still anticipating their interceptive actions mimicking the conditions under terrestrial gravity. Subsequent work using virtual reality setups showed that participants could adapt to arbitrary gravities in a few trials. However, the performance is still lower than that under terrestrial gravity conditions (Zago and Lacquaniti, 2005; Zago et al., 2005).
Since then, an implicit representation of gravity has been found at a neurobiological level (Indovina et al., 2005; Miller et al., 2008) and in a broad range of tasks such as eye behavior (Bosco et al., 2012; Diaz et al., 2013; Jörges and López-Moliner, 2019) or the estimation of the duration of events (Hosking and Crassini, 2010; Moscatelli and Lacquaniti, 2011; Jörges and López-Moliner, 2020) despite our general insensibility to accelerations (Werkhoven et al., 1992).
However, the best example of a representation of gravity for sensorimotor control is that an observer does not need to see an ascending ball falling to intercept it (de la Malla and López-Moliner, 2015). In general, humans have an implicit expectation that upwards moving objects will eventually fall (López-Moliner et al., 2010; Reed et al., 2010). Nevertheless, this representation may not be available for every kind of task. For example, timing tasks for gravitationally accelerated objects in imagination show a bias towards the last visible motion speed (Gravano et al., 2017; Bratzke and Ulrich, 2021). In this line, some authors dismiss an internal model-based explanation favoring a prediction-free explanation (Baurès et al., 2007; Katsumata and Russell, 2012). However, the lack of adaptation under microgravity conditions and the need to account for sensorimotor delays pinpoint the relevance of gravity prior guiding predictive control (Zago et al., 2008).
A recent study (Jörges and López-Moliner, 2020) tried to derive the mean and standard deviation of the Gravity prior in a Bayesian framework. Their results found a prior with a standard deviation of 14% (WF). According to the authors, these results might correspond with an upper bound, as there seem to be theoretical reasons such as the lack of adaptation to arbitrary gravity values suggesting a relatively inflexible and robust gravity prior (Jörges and López-Moliner, 2017).
To test the use of different pieces of prior information for calibration, it is first necessary to put forward algorithms that require pieces of internalized knowledge. In the temporal domain, Gómez and López-Moliner (2013) showed that by using both prior knowledge of gravity and size, visual information could be calibrated, resulting in actionable estimates of TTC. The model was named GS model about the assumption of a priori known gravity and size.
Time-To-Contact Estimation
GS Model
The GS model (Gómez and López-Moliner, 2013) is an algorithm that relies on calibrated optic information using prior knowledge to obtain estimates of TTC for parabolic trajectories. It relies on a combination of contextual variables such as known ball size (s) and gravitational acceleration (g) along with monocular cues such as retinal size (θ), elevation angle (γ) and its first derivative (), providing accurate estimates of TTC.
Known ball size (s) and retinal size (θ) provide a mapping from retinal to Cartesian metrics. On its part, gravitational acceleration (g) calibrates and normalizes the rate of change of the elevation angle () to be interpreted into meaningful predictions of TTC under arbitrary gravitational accelerations. In addition, cos(γ) would act as a non-declarative internalized parameter linked to action expecting that the elevation angle (γ) would increase over time (Reed et al., 2010; Shaffer et al., 2013). Removing the internalized variables from the GS model, one can still obtain a correlate of TTC based on retinal size (θ), the elevation angle (γ) and the rate of change of the elevation angle (). However, its value is meaningless in signaling an actionable TTC and, therefore, not directly applicable. The following section will illustrate that envisioning those pieces of prior knowledge as priors within an encoding-decoding framework could calibrate ambiguous visual cues into accurate estimates of TTC for parabolic trajectories.
Simulating the Benefits of Using Gravity and Size Priors for the Decoding
Under the constructivist framework, visual information is underspecified, and many trajectories can originate the same stimulation. This simulation shows how entering the correct gravity and size values increases the chances of inferring the actual trajectory exposed to the system from a subset of the possible ones.
In the case of this simulation, the possible inferred parabolas are a combination of the nine different TTC (ranging from 1.8 to 2.2 s in steps of 0.05 s), five different conditions of gravity (8.826, 9.316, 9.807, 10.297, 10.787 m/s2) and five different sizes (0.0703, 0.07215, 0.074, 0.07585, 0.0777 m) launched at eye-level 30 meters away from the observer in a head-on approach. We chose those values of gravitational acceleration and size using differences of 5% the standard of Earth gravity and 2.5% of the standard size of a baseball to envision reliable values of prior knowledge of gravity and size (however, see Jörges and López-Moliner, 2020).
Figure 4 represents nine test trajectories exposed to the system (panel A), whereas panels B, C and D represent the range of possible values for each variable available for the observer: retinal size (θ), the elevation angle (γ) and the rate of change of the elevation angle () respectively.
Figure 4. Test parabolas used in the simulation. (A) Ball’s vertical vs. depth position. Panels (B–D) indicate retinal size (θ), elevation angle (γ) and rate of change of the elevation angle () as a function of time for the first 200 ms of the trajectory. Panels (E,F) depict the output of the GS model using visual information only and combined with prior information, respectively.
To reproduce the encoding process of a sensory stimulus, we simulated a set of tuning curves covering the range of possible stimulus strengths (stimulus values) for each optic variable 200 ms. after motion onset. In Figure 5, the reader can see an illustration of the tuning curves (black curves), the stimulus strength presented to the system (blue vertical line), and an example of the average response by a detector (red curve) for the average TTC under standard conditions of gravity and size (light blue trajectory in Figure 4). The detectors simulated for retinal size (θ) covered a range from 0.0024 to 0.0031 rad (SD = 0.00014; representing a 5% WF). The stimulus range covered for the elevation angle (γ) and its rate of change () was from 0.045 to 0.085 rad (SD = 0.00325 rad; 5% WF) and from 0.2 to 0.45 rad/s (SD = 0.01625 rad/s; 5% WF) respectively.
Figure 5. Simulated tuning curves of neurons specialized for different values of (A) retinal size (θ), (B) elevation angle (γ) and (C) rate of change of the elevation angle (). The blue vertical line indicates the true stimulus strength exposed to the system. The stimulus strength was selected from the standard condition 200 ms after motion onset (see main text). The red curve indicates the average activation per neuron in a single trial (Poisson noise added). The red dashed lines indicate the stimulus strength inferred by the encoding process.
After the detectors were exposed to the stimulus, we obtained the average response probability (r; solid red line in Figure 5) for corresponding neural detectors (Dayan and Abbott, 2001) as:
In this expression, (f) indicates the mean activity per detector, whereas (r!) corresponds to the factorial response for each detector. In our simulation, the resulting activation varies on each iteration by adding Poisson noise, representing random variability in neural activation.
We simulated 1,000 trials per TTC in which the size was the standard of a baseball (0.074 m), and gravity was the standard on Earth (9.807 m/s2). Once the optic variables were encoded by the detectors simulated, we recovered the most likely stimulus strength presented to the system for each optic variable using a Maximum Likelihood estimate (MLE) procedure. Then, we obtained a value corresponding to the GS model based only on the optic variables retrieved by the encoding procedure.
We compared that output to the GS model’s ideal (noiseless) output based only on optic variables for all the simulated trajectories. We obtained two possible sets of decoding responses to select all the potential trajectories that matched the model’s output. For the first set of responses, we only used sensory information without resorting to size or gravity priors to decode the correct TTC; that is, we used a Maximum Likelihood Estimation procedure. We assumed the correct size and gravity priors to decode the correct TTC for the second set of responses. Then, we selected all potential trajectories that would fall into a relatively low error margin of ± 5%.
Figure 6 depicts the average accuracy per procedure. Here, the performance of a system using only sensory information (MLE; red dashed line) is slightly over the chance level (blue dashed line). On the contrary, using the correct priors (central dots in panels A and B of Figure 6) improves the proportion of correctly estimated TTC’s substantially. Note that since the relative difference between the different simulated gravity values is higher than those of size, the procedure benefits more from an accurate representation of internalized knowledge of gravity. This example highlights that, despite the inherent ambiguity of sensory information in the optic flow, the use of prior information is a powerful calibration tool to interpret otherwise ambiguous visual information.
Figure 6. Average accuracy per procedure. The blue dashed line indicates chance level (11.1%), red dashed line indicates the performance of an maximum likelihood estimate (MLE) procedure. The black points indicate the performance assuming different size priors in panel (A) (g = 9.807 m/s2 assumed) and assuming different gravity priors in panel (B) (baseball size assumed).
Accuracy and Precision of Using Gravity and Known Size
So far, we have shown that calibrating visual information in the light of prior knowledge allows us to draw accurate predictions using the GS model. However, it could still be the case that even if the output is accurate, the visual information in the optic flow is so noisy that an estimation of TTC might not be available.
To evaluate an observer’s ability to estimate TTC accurately and precisely from the GS model, we simulated a series of typical spatio-temporal parameters for parabolic trajectories. We simulated parabolic trajectories launching at eye-level at five initial distances (Zinit = 15, 20, 30, 40, 50 m), one contact time (TTC = 2 s; Δt = 0.01 s) and a single radius corresponding to a baseball (0.037 m). In each case, the endpoint is the origin, that is, the position of the simulated observer.
To evaluate the precision of the output, we introduced independent Gaussian noise to θ, γ and according to their respective WFs (identified with the letter k).
To test the noise-suppression performance for the GS model, we ran 10,000 simulations for each condition using Equations (5)–(7). Then, we obtained a WF timewise as the ratio between the standard deviation of the signal and the mean predicted TTC. Note that the χ version of a variable denotes its noisified version.
Figure 7A depicts the output of the GS model for each initial distance (color code), whereas the inset indicates the predicted temporal error for the ideal (noise-free) output of the model. The GS model provides very accurate estimates for which the maximum error is lower than 10 ms. On its part, Figure 7B represents the WF estimated timewise for the GS model. As a comparison, the WF for Tau was envisioned as constant at 10% as reviewed above. The GS model presents a precise output during most of the trajectory (WF is always lower than 10%), resulting in an accurate and robust solution to the estimation of TTC in parabolic trajectories comparable to previous WFs found in the literature (Moscatelli and Lacquaniti, 2011; Jörges and López-Moliner, 2020).
Figure 7. (A) Noisified estimates of TTC using the GS model for different trajectories. The inset represents the temporal error for the noiseless output of the GS model. (B) Weber fraction (WF) computed as the ratio between standard deviation and mean of the GS model each temporal frame. The red dashed line indicates the mean WF of Tau (see main text). The translucid output indicates the WF of a combination of the GS model and Tau using an MLE procedure.
It is essential to mention here that just as others have already described in the literature, the sources of information to estimate TTC may vary depending on the segments of an approach visible (DeLucia, 2004; López-Moliner et al., 2013; DeLucia et al., 2016). While an initial temporal estimate would be available using the GS model, final interceptive actions would take advantage of more straightforward strategies such as a distance criterion (Wann, 1996; López-Moliner and Keil, 2012; Gómez and López-Moliner, 2013), Tau (Lee, 1976; Zago et al., 2004; Shaffer and McBeath, 2005) or correlates of binocular disparity (Rushton and Wann, 1999).
Following this reasoning, de la Malla and López-Moliner (2015) partially validated the use of the GS model, showing that early estimates of TTC based on the GS model could be integrated with the latest estimates derived from correlates of the rate of expansion resulting in an accurate and precise timing mechanism. Mimicking that context, we combined the predictions of Tau and the GS model using a maximum likelihood process (Ernst and Banks, 2002). The output results in a robust solution against sensory noise for the estimation of TTC and timing interceptive actions (see translucid lines in Figure 7B).
Generalization of the GS Model
The formalization of the GS model assumes that the ball moves in a collision course with the observer (Gómez and López-Moliner, 2013). However, this is not usually the case. Commonly an observer must move to intercept the ball. Therefore, the following question is to what extent the output of the GS model deviates from perfect accuracy for trajectories ending in locations other than the observer’s position?
To investigate those cases, we estimated the output of the GS model for trajectories ending at different interception locations. We simulated one initial distance (Zinit = 50 m) and eight interception points around the observer (see Figure 8A). Initially, the GS model provides accurate estimates of the TTC regardless of the position of the observer. Then, in contrast with trajectories on a collision course, the simulation reflects systematic errors in TTC estimation shortly after motion onset if the observer remains stationary (see Figure 8B). If the ball falls behind the observer, the rate of change of the predicted TTC decreases. Thus, the model’s output overestimates the remaining TTC and vice versa for balls falling ahead (see Figure 8B), pointing out that the errors depend on the interception location. In this context, a navigational strategy predicting where and when the ball would be within reach would initially guide the observer towards the wrong position.
Figure 8. (A) Ending positions for simulated trajectories around the observer. The lines represent the trajectories followed by the moving observer. Panels (B,C) depict the output of the GS model for a stationary and a moving observer. The line code indicates lateral ending position [XEnd = 0, 5 (m)]. The color code indicates the ending position in depth [ZEnd = −5, 0, 5 (m)] . Note that the GS model predicts an underestimation of TTC for balls falling ahead and an overestimation of TTC for those falling behind the observer.
Nevertheless, the simulations described so far result in predictions of TTC for an unlikely situation in which the observer is not at the interception location and remains stationary. Usually, an observer would control the ball moving towards the interception area. As a result, an observer’s movement would prompt changes in the optic flow. As we will show, simulating the observer’s movement, we found interesting properties in the output of the GS model that might indicate the availability of a navigational strategy.
To perform the following simulations, we replicated the previous trajectories. However, in this case, the observer started moving towards the interception location at a constant speed 500 ms after the ball’s launch. We chose this moment because catchers generally start running in the right direction 500 ms after the ball’s flight has started (Michaels and Oudejans, 1992; McLeod and Dienes, 1993, 1996; Hecht et al., 1996; Brouwer et al., 2006). The displacement speed of the observer was computed to reach the interception point just in time to catch the ball (see inset in Figure 8C). Although this pattern of displacement and speed does not correspond precisely to the found in real life (McLeod and Dienes, 1993; McLeod et al., 2006), it is essential to point out that it will be useful for an illustrative purpose.
In Figure 8B, the reader can see that when the observer remains stationary for some time in a position other than the interception location, the rate of change of the predicted TTC changes (Figure 8B). When the rate of change in TTC decreases, the ball will fall behind the observer and vice versa. This would signal the need to move and also provide information of the correct direction of movement in depth. Therefore, departures from the initial rate of change in TTC (slopes different from −1) could be used as a navigational strategy indicating if the observer must move forward or backwards. Instead, the movement of an observer in the correct direction provides the necessary change in the optic flow to linearize the predictions of the remaining TTC (see Figure 8C). Thus, keeping the prediction linear will ensure that the observer would end up at the interception position in time.
In sum, the above simulations indicate that the model’s output is accurate when the observer moves in the correct direction and speed providing the basis for a mechanism to navigate towards the interception location. However, these simulations were performed in a context in which the ball is only affected by the gravitational acceleration. Would a simulation of trajectories under air drag provide equally accurate temporal estimates?
Dynamic Effects: Air Drag
In real life, the ball is affected by external forces other than gravity, such as air drag, Magnus force or wind currents. These forces deviate the trajectory from a perfect parabola compared to motion in a vacuum in astonishing ways (McBeath et al., 2008). For instance, previous works indicate that air drag can reduce flight time and distance traveled by a flying ball up to 50% (Brancazio, 1985; Adair, 2002). Therefore, trajectories initially on a collision course with the observer are no longer so after a short period. This pattern would potentially preclude the use of different algorithms for estimating the TTC, such as the Tau or GS model.
It has been argued that the self-regulatory nature of information-based strategies can efficiently deal with dynamic effects in a parabolic trajectory, provided that continuous visual information is available. In contrast, it is commonly argued that an internal model assuming a constant gravitational acceleration would be insufficient to account for dynamic forces such as air drag (Fink et al., 2009). To account for air drag, an internal model would have to gain access to a drag coefficient, mass and size for every single object and environment dynamically, which limits a massive application (Craig et al., 2006). Furthermore, it seems at odds with the fact that most people think that objects fall at the same rate despite their mass or volume (Oberle et al., 2005). However, explicit knowledge of physics may not affect performance in action-related tasks (Reed et al., 2010; Flavell, 2014). Following this reasoning, in our view, predictions using priors would only include variables facilitating the interpretation of the most generic case of natural law or parameters for a given task. In the following, we will show how the GS model, which relies only on gravity and size priors, can predict the remaining TTC reliably for the general case of trajectories under gravity and air drag conditions.
Unlike the gravitational force, which exerts the same force for different projectiles, air resistance depends on several factors: ρ, cosity of the environment surrounding the object; Cd, a drag coefficient relative to the texture and shape of the projectile essentially; r object’s radius and v, the tangential speed of the object estimated dynamically. To simulate the effects of air drag on a parabolic trajectory, we followed the procedure described in Timmerman and van der Weele (1999) and Gómez and López-Moliner (2013). We simulated different trajectories under two different conditions: gravity only and gravity + air drag. Air viscosity around the ball (ρ) was set to 1.225 kg/m3 (value at sea level), and Cd was set to 0.346 or 0.4 for baseball and soccer balls, respectively (Alam et al., 2010; Kagan and Nathan, 2014). We introduced one initial vertical speed Vy0 = 9.807 m/s (2 s of flight time under gravity only conditions) for each trajectory and corresponding approaching speeds (Vy0 = 7.5,15,25 m/s) different balls launched at the origin. In this simulation, we did not include horizontal displacements.
In Figure 9A, the reader can see how air drag influences the trajectory described by the ball in both: the spatial and the temporal domain. In Figure 9A, the gray trajectory represents the trajectory followed by a ball under gravity-only conditions, whereas the blue and red trajectories indicate the trajectories followed by baseball and soccer balls including air drag in the simulation. For soccer balls, the effect of air drag is more pronounced mainly due to a larger cross-sectional area against the air. Note that since the initial vertical speed was the same for all the trajectories, the differences in flight time and distance traveled can be attributed to the different approaching speeds.
Figure 9. (A) Lateral view of different parabolic trajectories under gravity (gray lines) and gravity + air drag conditions for two different balls (red: soccer ball; blue: baseballs) and three initial approaching speeds (different panels). The figure annotates the difference in distance traveled (ZΔ) and flight duration (tΔ) compared to a trajectory only considering gravity. The black and orange dots in the third panel indicate the position of the corresponding simulated observer in panels (B) or (C,D), respectively. The green arrow indicates the displacement simulated in panel (D). Panels (B–D) indicate the predicted TTC using the GS model for different simulated observers in the worst-case scenario simulated. Insets depict the corresponding temporal errors using the predictions of the GS model.
Thus, how well can the GS model estimate the remaining TTC in trajectories, including air drag? To answer this question, we simulated the output of the GS model for the worst-case scenario previously simulated. In that case, the GS model will yield the least accurate predictions. As shown in Figure 9A, the trajectory most affected by air drag is when a Soccer ball moves at the highest horizontal speed (Vz0 = 25 m/s).
To test the model’s accuracy, we used three different situations. In the first one, the observer is stationary at the interception position under gravity + air drag conditions (black dot in Figure 9A). In the second, the observer is stationary at a midpoint between the fall point under gravity-only and gravity + air drag conditions (black dot in Figure 9B). Finally, we simulated a situation in which the observer is at the same “mid-point.” However, in this case, the observer moves towards the intercept point at a constant speed (8.13 m/s), 500 ms after motion onset (green arrow in Figure 9A).
The output of the model for corresponding situations is depicted in Figures 9B–D. In all cases, the GS model reflects initial temporal errors corresponding to the difference in flight time between trajectories under gravity-only and gravity + air drag conditions (see annotations within Figure 9A). When the observer stands still in the interception location (Figure 9B), the GS model presents a high degree of accuracy during most of the trajectory. For instance, 0.5 s before the collision, the output converges to temporal errors of about 10–20 ms. However, if the observer stands still at the midpoint (Figure 9C), the model’s output deviates severely. In this case, the rate of change in the predicted TTC remains consistently lower than −1. In principle, this pattern could inform the observer that the interception point would be ahead of their position. In contrast, when the observer heads towards the interception location (Figure 9D), the model yield accurate predictions.
These simulations provide evidence that the output of the information included within the GS model provides accurate and actionable predictions of the remaining TTC when the observer remains stationary in the interception location or displaces towards the interception location. Therefore, it could be used as a navigational strategy or to plan the final interceptive action even when air drag is present.
Benefits and Limitations of the Generalization of the GS Model
The GS model, like Tau, provides temporal information that may involve certain predictive benefits compared to the error-nulling strategies within the outfielder problem. Nevertheless, it also has some limitations that will be addressed in this section.
First, the GS model is much more robust to sensory noise than Tau (Gómez and López-Moliner, 2013). It uses the rate of change of the elevation angle () to estimate the TTC instead of a much noisier variable, the rate of expansion () upon which Tau relies.
Second, Tau only provides accurate estimates when the ball is moving at constant speed towards the observer. In contrast, the GS provides accurate estimates at launch independently of the observer’s position, which would provide an initial accurate temporal information useful for planning the action.
Third, the GS model overestimates the TTC for trajectories under air drag as a function of the difference in flight time between trajectories under gravity-only and gravity + air drag conditions. However, our simulations show that in combination with the observer movement, temporal errors of less than 50 ms are possible 1 s before collision (Figure 9D).
Fourth, in contrast with previous error-nulling strategies, the GS model provides temporal information and can help compensate for the temporal delays and occlusions due to its predictive nature. The TTC can be used to adjust locomotion speed, inform of the remaining TTC under restricted visibility conditions or plan the final manual interception. On its part, the rate of change of TTC could be used to adjust the direction of movement. In this sense, both signals are complementary. However, the viability of the latter requires being able to detect changes in the rate of change of the predicted TTC, which needs further research. Furthermore, the detection of variations in the rate of change of TTC will likely incur delays. Therefore, future studies should study to which extent the simultaneous TTC signal can compensate for these delays.
Finally, the prediction from the GS model shares some limitations with Tau; that is, dealing with non-spherical objects such as Rugby balls or Frisbees would need further elaboration. However, some studies found that those errors can be canceled out by adding binocular information at the latter stages of catching (Gray and Regan, 1998).
Evidence of Prediction in Eye Behavior and Manual Interception
The main problem to find support for model-based controlled behavior is that, when possible, the observer would keep track of the trajectory continuously (Oudejans et al., 1999; Postma et al., 2014). Indeed, this is the case of the outfielder problem, for which there is only anecdotal evidence of successful catchers directing away the gaze from the ball (Chodosh et al., 1995). In this context, accurate actions would not allow us to discriminate between information-based and model-based control directly (Belousov et al., 2016). Because of that, we need to scrutinize scenarios in which simple solutions such as heuristics or mappings between sensory information and temporal correlates for temporal estimation or action initiation are not available (Zhao and Warren, 2015).
One possibility to unveil the need for prediction in action control is to manipulate the target’s visibility. It is the most widely used experimental manipulation to study the predictive nature of behavior in interception (Sharp and Whiting, 1974; Whiting and Sharp, 1974; López-Moliner et al., 2010; Brenner and Smeets, 2011; Spering et al., 2011; de la Malla and López-Moliner, 2015). Nevertheless, more natural conditions are essential to understand how an observer could use temporal estimates to guide their action. In natural conditions, our gaze is often shifted to different locations to gather the information that may be relevant shortly (Hayhoe et al., 2005). In other contexts, a player would divert her gaze to check for deviations caused by balls’ bouncing (Diaz et al., 2013) or confirm whether the ball was appropriately hit (Mann et al., 2013). Some studies in manual interception and temporal estimations gave the observer complete freedom to decide which part of a trajectory they wanted to exploit visually while dealing with alternative tasks (Faisal and Wolpert, 2009; López-Moliner and Brenner, 2016; Aguado and López-Moliner, 2021). In those cases, where and when the observer averts the gaze from the target may provide valuable clues about the most relevant pieces of information according to task demands. Therefore, future studies might investigate when people prefer to divert the gaze from the ball while moving towards the interception location.
In some cases, it has been suggested the existence of privileged portions of the trajectory available for an observer to judge TTC. For example, in juggling or catching a ball, looking at the apex would provide the most relevant information (Whiting, 1968; Todd, 1981; Watson et al., 1992). However, a closer look at experimental data indicates that an observer does not actively search for a particular position in the parabola. Instead, prefers to use fixed temporal viewing windows generating priors during the task. These priors could then be used to weight visual information or estimate TTC when sensory information is unavailable (Amazeen et al., 1999; López-Moliner and Keil, 2012; Aguado and López-Moliner, 2021).
For example, most studies show acceptable catching performance in manual interception tasks for short flight durations no matter the section of the trajectory viewed. However, visual information had to be captured at least 200 ms before the catch to avoid sensorimotor delays (Sharp and Whiting, 1974; López-Moliner et al., 2010; López-Moliner and Brenner, 2016). For longer flight durations (up to 2 s), catching performance describes an inverted U shape with respect to flight duration (Sharp and Whiting, 1974; Amazeen et al., 1999). If the observer can only see the ball well in advance, performance would be low because the predictions decay rapidly (Binsted et al., 2006; Aguado and López-Moliner, 2021).
Zhao and Warren (2015) reasoned that in the case of short flight durations, part of an observer’s performance could be explained by the observer having learnt some of the regularities of a predictable trajectory mapping optic variables with a temporal correlate. Still, this would indicate the usefulness of developing priors during the task, which would be exploited when online visual information is not available. However, the fact that an observer exploits visual information when optic mappings are available indicates that they prefer to update their predictions based on the latest available visual information and combine it with evidence from previous knowledge (Mazyn et al., 2007; Binaee and Diaz, 2019). In the end, having a rough prediction is better than none (Brenner and Smeets, 2018).
In this line, de la Malla and López-Moliner (2015) proved that general rules of integration apply to the estimation of the TTC, which means: the observer integrates past and concurrent information to optimize the precision of temporal responses in a continuous fashion (Todorov, 2004; Liu and Todorov, 2007; Dimitriou et al., 2013). Assuming this is true, we can use Kalman filters to predict an observer’s estimation of TTC and response variability. A Kalman filter (Kalman, 1960) is a Bayesian tool that estimates the state of a system combining new noisy estimates, a prediction from prior measurements and a prior knowledge of how the system behaves. Using this technique, we could estimate both the accuracy and precision of online measurements for temporal judgments, manual interceptive tasks, and more general interceptive tasks such as the locomotion within the outfielder problem.
Future Research
One of the main objectives of this work is to highlight the potential role of prior knowledge in calibrating visual information in terms of actionable predictions such as TTC. In our view, drawing predictions based on prior knowledge is not just a reliable and accurate way to predict future states of the environment but also helps us override the need to use unreliable optic cues (Cutting and Vishton, 1995). For instance: expansion rate () and thus Tau (τ) might not be available at large distances, optic acceleration () is hardly discriminable by humans and would not allow for the continuous control of action (Werkhoven et al., 1992) even though some studies claim that the values of optical acceleration available are large enough to be detected (Babler and Dannemiller, 1993; Zaal et al., 2012); lastly, the literature is mixed with regards to the benefits of providing binocular disparity. Because of these reasons, here we advocate using the rate of change in the elevation angle (), which is very precise, in combination with an internalized knowledge of gravity and physical size for the estimation of TTC in the GS model.
Here, the GS model provides different contexts to test the information included. For example, within the GS model, each contextual piece of information, gravity or size, is either in the denominator or the numerator. Hence, introducing proportional changes in the parameters governing the trajectory would result in proportional errors in the estimates of the remaining TTC. In a similar line, Jörges and López-Moliner (2017) showed that an observer might be able to extract information about the approaching speed of a ball through estimations of the rate of change of the elevation angle prior knowledge of gravity. Therefore, different values of gravity governing a trajectory should influence the prediction of the interception location.
Moreover, using TTC discrimination tasks, it could be possible to study if an observer can detect differences between trajectories under gravity-only conditions and gravity + air drag conditions. Our simulations indicated a WF of about 7% for the GS output (see Figure 7). Therefore, the difference in TTC should be above the discrimination threshold in some cases, as depicted in Figure 9A. Furthermore, decision tasks based on an observer’s ability to decide if there is enough time to perform alternate tasks (e.g., looking for teammates or running towards the interception area) from early visual information might be essential to test the availability of temporal estimates as a parameter to plan action for a broader range of interceptive actions.
On another note, it might be interesting to investigate the use of the GS model as a navigational mechanism. Since the GS model does not specify the interception location to plan interception in advance, we discovered a continuous coupling to keep a constant rate of change of the predicted TTC. To test if an observer would adapt locomotion to a constant rate of change, we should introduce players into contexts in which the value of gravitational acceleration or ball size do not correspond with the parameters assumed a priori. As introduced above, changes in the parameters would result in estimation errors of the remaining TTC. Thus, these manipulations would lead to predictions of the path followed by the observer. Nevertheless, to be able to use such a strategy, an observer might be able to detect deviations from different rates of change in TTC. To our knowledge, there is no previous work providing figures about how well people detect changes in TTC. Thus, our ability to detect differences in the rate of change and the time required to do so will need to be studied in future works.
To generate the suggested experiments, we need immersive and realistic spaces. Virtual scenarios will provide ecologically valid contexts to evaluate to what extent predictions influence interception. To do so, the use of wireless head-mounted displays (HMD) and portable eye-trackers will be essential. HMD insert the participants into rich and controlled environments already being used to train professional sports players (Zaal and Bootsma, 2011; Gray, 2017; Harris et al., 2020). Combining this technique with built-in eye-tracking systems provides access to how players interact with the environment to gather relevant visual information (Binaee et al., 2016; Moran et al., 2018). Those findings would still need to be replicated in real life under full-cue conditions. However, augmented reality devices are becoming more and more accessible and are likely to become more widespread. Those results may not be fully transferable to real life. However, it still would provide us information about human performance interacting with increasingly in-demand devices with potential applicability in a growing industry, eSports.
Code Availability
The code to reproduce all the simulations included in this work can be found in the following link: https://osf.io/sa3cm/.
Data Availability Statement
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article.
Author Contributions
JL-M initially conceptualized the manuscript. BA performed the literature review based on JL-M’s suggestions and initial guidance. BA and JL-M programmed the simulations. BA wrote the manuscript with input from JL-M. All authors contributed to the article and approved the submitted version.
Funding
The research group was funded by the Catalan government (2017SGR-48) and grant ref. PSI2017-83493-R (AEI/FEDER, UE). BA was supported by the fellowship FPU17/01248 from Ministerio de Educación y Formación Profesional of the Spanish government.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
We thank Cristina de la Malla and Björn Jörges for their comments and proof reading. We also thank the reviewers for their insightful and constructive criticism during peer review.
References
Adams, W. J., Graf, E. W., and Ernst, M. O. (2004). Experience can change the’light-from-above’prior. Nat. Neurosci. 7, 1057–1058. doi: 10.1038/nn1312
Aguado, B., and López-Moliner, J. (2021). Flexible viewing time when estimating time-to-contact in 3D parabolic trajectories. J. Vis. 21:9. doi: 10.1167/jov.21.4.9
Aguilar-Lleyda, D., Tubau, E., and López-Moliner, J. (2018). An object-tracking model that combines position and speed explains spatial and temporal responses in a timing task. J. Vis. 18:12. doi: 10.1167/18.12.12
Alam, F., Chowdhury, H., Moria, H., and Fuss, F. K. (2010). A comparative study of football aerodynamics. Procedia Eng. 2, 2443–2448. doi: 10.1016/j.proeng.2010.04.013
Amazeen, E. L., Amazeen, P. G., Post, A. A., and Beek, P. J. (1999). Timing the selection of information during rhythmic catching. J. Mot. Behav. 31, 279–289. doi: 10.1080/00222899909600994
Babler, T. G., and Dannemiller, J. L. (1993). Role of image acceleration in judging landing location of free-falling projectiles. J. Exp. Psychol. Hum. Percept. Perform. 19, 15–31. doi: 10.1037//0096-1523.19.1.15
Battaglia, P. W., Kersten, D., and Schrater, P. R. (2011). How haptic size sensations improve distance perception. PLoS Comput. Biol. 7:e1002080. doi: 10.1371/journal.pcbi.1002080
Battaglia, P. W., Schrater, P. R., and Kersten, D. J. (2005). “Auxiliary object knowledge influences visually-guided interception behavior,” in Proceedings of the 2nd Symposium on Applied Perception in Graphics and Visualization (A Coruña, Spain), 145–152.
Baurès, R., Benguigui, N., Amorim, M.-A., and Siegler, I. A. (2007). Intercepting free falling objects: better use occam’s razor than internalize newton’s law. Vis. Res. 47, 2982–2991. doi: 10.1016/j.visres.2007.07.024
Belousov, B., Neumann, G., Rothkopf, C. A., and Peters, J. R. (2016). “Catching heuristics are optimal control policies,” in Advances in Neural Information Processing Systems, eds D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and R. Garnett (Curran Associates, Inc.), 1426–1434.
Binaee, K., and Diaz, G. (2019). Movements of the eyes and hands are coordinated by a common predictive strategy. J. Vis. 19, 3–3. doi: 10.1167/19.12.3
Binaee, K., Diaz, G., Pelz, J., and Phillips, F. (2016). “Binocular eye tracking calibration during a virtual ball catching task using head mounted display,” in Proceedings of the ACM Symposium on Applied Perception, (Anaheim: CA), 15–18.
Binsted, G., Rolheiser, T. M., and Chua, R. (2006). Decay in visuomotor representations during manual aiming. J. Mot. Behav. 38, 82–87. doi: 10.3200/JMBR.38.2.82-87
Bootsma, R. J., and van Wieringen, P. C. (1990). Timing an attacking forehand drive in table tennis. J. Exp. Psychol. Hum. Percept. Perform. 16, 21–29. doi: 10.1037/0096-1523.16.1.21
Bosco, G., Delle Monache, S., and Lacquaniti, F. (2012). Catching what we can’t see: manual interception of occluded fly-ball trajectories. PLoS One 7:e49381. doi: 10.1371/journal.pone.0049381
Brancazio, P. J. (1985). Looking into chapman’s homer: the physics of judging a fly ball. Am. J. Phys. 53, 849–855. doi: 10.1119/1.14350
Bratzke, D., and Ulrich, R. (2021). Mental imagery of free fall: does a falling apple accelerate in our minds? Timing Time Percept. 9, 150–160. doi: 10.1163/22134468-bja10022
Brenner, E., Driesen, B., and Smeets, J. B. J. (2014). Precise timing when hitting falling balls. Front. Hum. Neurosci. 8:342. doi: 10.3389/fnhum.2014.00342
Brenner, E., and Smeets, J. B. J. (2007). Flexibility in intercepting moving objects. J. Vis. 7, 14.1–17. doi: 10.1167/7.5.14
Brenner, E., and Smeets, J. B. J. (2011). Continuous visual control of interception. Hum. Mov. Sci. 30, 475–494. doi: 10.1016/j.humov.2010.12.007
Brenner, E., and Smeets, J. B. J. (2015a). How moving backgrounds influence interception. PLoS One 10:e0119903. doi: 10.1371/journal.pone.0119903
Brenner, E., and Smeets, J. B. J. (2015b). How people achieve their amazing temporal precision in interception. J. Vis. 15:8. doi: 10.1167/15.3.8
Brenner, E., and Smeets, J. B. J. (2018). Continuously updating one’s predictions underlies successful interception. J. Neurophysiol. 120, 3257–3274. doi: 10.1152/jn.00517.2018
Brouwer, A.-M., López-Moliner, J., Brenner, E., and Smeets, J. B. (2006). Determining whether a ball will land behind or in front of you: not just a combination of expansion and angular velocity. Vis. Res. 46, 382–391. doi: 10.1016/j.visres.2005.09.002
Bruggeman, H., Zosh, W., and Warren, W. H. (2007). Optic flow drives human visuo-locomotor adaptation. Curr. Biol. 17, 2035–2040. doi: 10.1016/j.cub.2007.10.059
Calderone, J. B., and Kaiser, M. K. (1989). Visual acceleration detection: effect of sign and motion orientation. Percept. Psychophys. 45, 391–394. doi: 10.3758/bf03210711
Cámara, C., de la Malla, C., López-Moliner, J., and Brenner, E. (2018). Eye movements in interception with delayed visual feedback. Exp. Brain Res. 236, 1837–1847. doi: 10.1007/s00221-018-5257-8
Chodosh, L. A., Lifson, L. E., and Tabin, C. (1995). Play ball!. Science 268, 1682–1683. doi: 10.1126/science.7792585
Craig, C. M., Berton, E., Rao, G., Fernandez, L., and Bootsma, R. J. (2006). Judging where a ball will go: the case of curved free kicks in football. Naturwissenschaften 93, 97–101. doi: 10.1007/s00114-005-0071-0
Crowell, J. A., and Banks, M. S. (1996). Ideal observer for heading judgments. Vis. Res. 36, 471–490. doi: 10.1016/0042-6989(95)00121-2
Crowell, J. A., Banks, M. S., Shenoy, K. V., and Andersen, R. A. (1998). Visual self-motion perception during head turns. Nat. Neurosci. 1, 732–737. doi: 10.1038/3732
Cutting, J. E., and Vishton, P. M. (1995). “Perceiving layout and knowing distances: the integration, relative potency, and contextual use of different information about depth,” in Perception of Space and Motion, eds W. Epstein and S. Rogers (San Diego, CA: Academic Press), 69–117.
Dayan, P., and Abbott, L. (2001). Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems. Cambridge, MA: MIT Press.
de Bruyn, B., and Orban, G. A. (1988). Human velocity and direction discrimination measured with random dot patterns. Vis. Res. 28, 1323–1335. doi: 10.1016/0042-6989(88)90064-8
de la Malla, C., and López-Moliner, J. (2015). Predictive plus online visual information optimizes temporal precision in interception. J. Exp. Psychol. Hum. Percept. Perform. 41, 1271–1280. doi: 10.1037/xhp0000075
de la Malla, C., Smeets, J. B., and Brenner, E. (2017). Potential systematic interception errors are avoided when tracking the target with one’s eyes. Sci. Rep. 7:10793. doi: 10.1038/s41598-017-11200-5
DeLucia, P. R. (2004). “Multiple sources of information influence time-to-contact judgments: do heuristics accommodate limits in sensory and cognitive processes?,” in Advances in Psychology, Vol. 135. Time-to-Contact, (Amsterdam: Elsevier/North-Holland), 243–285.
DeLucia, P. R., Meza-Arroyo, M., Baurès, R., Ranjit, M., Hsiang, S., and Gorman, J. C. (2016). Continuous response monitoring of relative time-to-contact judgments: does effective information change during an approach event? Ecol. Psychol. 28, 1–22. doi: 10.1080/10407413.2016.1121735
Dessing, J. C., Wijdenes, L. O., Peper, C. E., and Beek, P. J. (2009). Visuomotor transformation for interception: catching while fixating. Exp. Brain Res. 196, 511–527. doi: 10.1007/s00221-009-1882-6
Diaz, G., Cooper, J., Rothkopf, C., and Hayhoe, M. (2013). Saccades to future ball location reveal memory-based prediction in a virtual-reality interception task. J. Vis. 13:20. doi: 10.1167/13.1.20
Dimitriou, M., Wolpert, D. M., and Franklin, D. W. (2013). The temporal evolution of feedback gains rapidly update to task demands. J. Neurosci. 33, 10898–10909. doi: 10.1523/JNEUROSCI.5669-12.2013
Ernst, M. O., and Banks, M. S. (2002). Humans integrate visual and haptic information in a statistically optimal fashion. Nature 415, 429–433. doi: 10.1038/415429a
Faisal, A. A., and Wolpert, D. M. (2009). Near optimal combination of sensory and motor uncertainty in time during a naturalistic perception-action task. J. Neurophysiol. 101, 1901–1912. doi: 10.1152/jn.90974.2008
Fajen, B. R. (2005a). Calibration, information, and control strategies for braking to avoid a collision. J. Exp. Psychol. Hum. Percept. Perform. 31, 480–501. doi: 10.1037/0096-1523.31.3.480
Fajen, B. R. (2005b). Perceiving possibilities for action: on the necessity of calibration and perceptual learning for the visual guidance of action. Perception 34, 717–740. doi: 10.1068/p5405
Fajen, B. R. (2007). Affordance-based control of visually guided action. Ecol. Psychol. 19, 383–410. doi: 10.1080/10407410701557877
Fink, P. W., Foo, P. S., and Warren, W. H. (2009). Catching fly balls in virtual reality: a critical test of the outfielder problem. J. Vis. 9, 14.1–8. doi: 10.1167/9.13.14
Flavell, J. C. (2014). An Investigation into the Directional and Amplitude Aspects of An Internal Model of Gravity. Thesis. Manchester Metropolitan University
Friston, K. (2010). The free-energy principle: a unified brain theory? Nat. Rev. Neurosci. 11, 127–138. doi: 10.1038/nrn2787
Gibson, J. J. (1979). The Ecological Approach to Visual Perception. Boston, MA: Houghton Mifflin Company.
Glennerster, A., Tcheang, L., Gilson, S. J., Fitzgibbon, A. W., and Parker, A. J. (2006). Humans ignore motion and stereo cues in favor of a fictional stable world. Curr. Biol. 16, 428–432. doi: 10.1016/j.cub.2006.01.019
Gómez, J., and López-Moliner, J. (2013). Synergies between optical and physical variables in intercepting parabolic targets. Front. Behav. Neurosci. 7:46. doi: 10.3389/fnbeh.2013.00046
Gottsdanker, R., Frick, J. W., and Lockard, R. B. (1961). Identifying the acceleration of visual targets. Br. J. Psychol. 52, 31–42. doi: 10.1111/j.2044-8295.1961.tb00765.x
Gravano, S., Zago, M., and Lacquaniti, F. (2017). Mental imagery of gravitational motion. Cortex 95, 172–191. doi: 10.1016/j.cortex.2017.08.005
Gray, R. (2017). Transfer of training from virtual to real baseball batting. Front. Psychol. 8:2183. doi: 10.3389/fpsyg.2017.02183
Gray, R., and Regan, D. (1998). Accuracy of estimating time to collision using binocular and monocular information. Vis. Res. 38, 499–512. doi: 10.1016/s0042-6989(97)00230-7
Harris, D. J., Bird, J. M., Smart, P. A., Wilson, M. R., and Vine, S. J. (2020). A framework for the testing and validation of simulated environments in experimentation and training. Front. Psychol. 11:605. doi: 10.3389/fpsyg.2020.00605
Harris, J. M., and Watamaniuk, S. N. J. (1995). Speed discrimination of motion-in-depth using binocular cues. Vis. Res. 35, 885–896. doi: 10.1016/0042-6989(94)00194-q
Hayhoe, M., Mennie, N., Sullivan, B., and Gorgos, K. (2005). “The role of internal models and prediction in catching balls,” in From Reactive to Anticipatory Cognitive Embodied Systems, ed C. Castelfranchi (Menlo Park, CA: American Association for Artificial Intelligence), 78–82.
Hecht, H., Kaiser, M. K., and Banks, M. S. (1996). Gravitational acceleration as a cue for absolute size and distance? Percept. Psychophys. 58, 1066–1075. doi: 10.3758/bf03206833
Hosking, S. G., and Crassini, B. (2010). The effects of familiar size and object trajectories on time-to-contact judgements. Exp. Brain Res. 203, 541–552. doi: 10.1007/s00221-010-2258-7
Indovina, I., Maffei, V., Bosco, G., Zago, M., Macaluso, E., and Lacquaniti, F. (2005). Representation of visual gravitational motion in the human vestibular cortex. Science 308, 416–419. doi: 10.1126/science.1107961
Ittelson, W. H. (1951). Size as a cue to distance: static localization. Am. J. Psychol. 64, 54–67. doi: 10.2307/1418595
Jacobs, D. M., and Michaels, C. F. (2006). Lateral interception I: operative optical variables, attunement, and calibration. J. Exp. Psychol. Hum. Percept. Perform. 32, 443–458. doi: 10.1037/0096-1523.32.2.443
Jacobs, D. M., and Michaels, C. F. (2007). Direct learning. Ecol. Psychol. 19, 321–349. doi: 10.1080/10407410701432337
Jörges, B., and López-Moliner, J. (2017). Gravity as a strong prior: implications for perception and action. Front. Hum. Neurosci. 11:203. doi: 10.3389/fnhum.2017.00203
Jörges, B., and López-Moliner, J. (2019). Earth-gravity congruent motion facilitates ocular control for pursuit of parabolic trajectories. Sci. Rep. 9:14094. doi: 10.1038/s41598-019-50512-6
Jörges, B., and López-Moliner, J. (2020). Determining mean and standard deviation of the strong prior through simulations. PLoS One 15:e0236732. doi: 10.1371/journal.pone.0236732
Judge, S., and Rind, F. (1997). The locust DCMD, a movement-detecting neurone tightly tuned to collision trajectories. J. Exp. Biol. 200, 2209–2216. doi: 10.1242/jeb.200.16.2209
Kagan, D., and Nathan, A. M. (2014). Simplified models for the drag coefficient of a pitched baseball. Phys. Teacher 52, 278–280. doi: 10.1119/1.4872406
Kalman, R. E. (1960). A new approach to linear filtering and prediction problems. J. Basic Eng. 82, 35–45. doi: 10.1115/1.3662552
Katsumata, H., and Russell, D. M. (2012). Prospective versus predictive control in timing of hitting a falling ball. Exp. Brain Res. 216, 499–514. doi: 10.1007/s00221-011-2954-y
Keil, M. S., and López-Moliner, J. (2012). Unifying time to contact estimation and collision avoidance across species. PLoS Comput. Biol. 8:e1002625. doi: 10.1371/journal.pcbi.1002625
Kersten, D., Mamassian, P., and Yuille, A. (2004). Object perception as bayesian inference. Annu. Rev. Psychol. 55, 271–304. doi: 10.1146/annurev.psych.55.090902.142005
Kistemaker, D. A., Faber, H., and Beek, P. J. (2009). Catching fly balls: a simulation study of the Chapman strategy. Hum. Mov. Sci. 28, 236–249. doi: 10.1016/j.humov.2008.11.001
Kistemaker, D. A., Van Soest, A. J., and Bobbert, M. F. (2006). Is equilibrium point control feasible for fast goal-directed single-joint movements? J. Neurophysiol. 95, 2898–2912. doi: 10.1152/jn.00983.2005
Klein, S. A., and Levi, D. M. (1987). Position sense of the peripheral retina. J. Opt. Soc. Am. A 4, 1543–1553. doi: 10.1364/josaa.4.001543
Knill, D. C., and Pouget, A. (2004). The Bayesian brain: the role of uncertainty in neural coding and computation. Trends Neurosci. 27, 712–719. doi: 10.1016/j.tins.2004.10.007
Kwon, O.-S., Tadin, D., and Knill, D. C. (2015). Unifying account of visual motion and position perception. Proc. Natl. Acad. Sci. U S A 112, 8142–8147. doi: 10.1073/pnas.1500361112
Lacquaniti, F., and Maioli, C. (1989). The role of preparation in tuning anticipatory and reflex responses during catching. J. Neurosci. 9, 134–148. doi: 10.1523/JNEUROSCI.09-01-00134.1989
Lee, D. N. (1976). A theory of visual control of braking based on information about time-to-collision. Perception 5, 437–459. doi: 10.1068/p050437
Lee, D., Young, D., Reddish, P., Lough, S., and Clayton, T. (1983). Visual timing in hitting an accelerating ball. Q. J. Exp. Psychol. 35, 333–346. doi: 10.1080/14640748308402138
Lewis, R. F., Gaymard, B. M., and Tamargo, R. J. (1998). Efference copy provides the eye position information required for visually guided reaching. J. Neurophysiol. 80, 1605–1608. doi: 10.1152/jn.1998.80.3.1605
Liu, D., and Todorov, E. (2007). Evidence for the flexible sensorimotor strategies predicted by optimal feedback control. J. Neurosci. 27, 9354–9368. doi: 10.1523/JNEUROSCI.1110-06.2007
López-Moliner, J., and Brenner, E. (2016). Flexible timing of eye movements when catching a ball. J. Vis. 16:13. doi: 10.1167/16.5.13
López-Moliner, J., Brenner, E., Louw, S., and Smeets, J. B. J. (2010). Catching a gently thrown ball. Exp. Brain Res. 206, 409–417. doi: 10.1007/s00221-010-2421-1
López-Moliner, J., Field, D. T., and Wann, J. P. (2007). Interceptive timing: prior knowledge matters. J. Vis. 7, 11.1–8. doi: 10.1167/7.13.11
López-Moliner, J., and Keil, M. S. (2012). People favour imperfect catching by assuming a stable world. PLoS One 7:e35705. doi: 10.1371/journal.pone.0035705
López-Moliner, J., Supèr, H., and Keil, M. S. (2013). The time course of estimating time-to-contact: switching between sources of information. Vis. Res. 92, 53–58. doi: 10.1016/j.visres.2013.09.007
Mann, D. L., Spratford, W., and Abernethy, B. (2013). The head tracks and gaze predicts: how the world’s best batters hit a ball. PLoS One 8:e58289. doi: 10.1371/journal.pone.0058289
Mazyn, L. I. N., Savelsbergh, G. J. P., Montagne, G., and Lenoir, M. (2007). Planning and on-line control of catching as a function of perceptual-motor constraints. Acta Psychol. 126, 59–78. doi: 10.1016/j.actpsy.2006.10.001
McBeath, M. K., Nathan, A. M., Bahill, A. T., and Baldwin, D. G. (2008). Paradoxical pop-ups: why are they difficult to catch? Am. J. Phys. 76, 723–729. doi: 10.1119/1.2937899
McBeath, M. K., Tang, T. Y., and Shaffer, D. M. (2018). The geometry of consciousness. Conscious. Cogn. 64, 207–215. doi: 10.1016/j.concog.2018.04.015
McBeath, M., Shaffer, D., and Kaiser, M. (1995). How baseball outfielders determine where to run to catch fly balls. Science 268, 569–573. doi: 10.1126/science.7725104
McIntyre, J., Zago, M., Berthoz, A., and Lacquaniti, F. (2001). Does the brain model newton’s laws? Nat. Neurosci. 4, 693–694. doi: 10.1038/89477
McKee, S. P. (1981). A local mechanism for differential velocity detection. Vis. Res. 21, 491–500. doi: 10.1016/0042-6989(81)90095-x
McKee, S. P., and Welch, L. (1992). The precision of size constancy. Vis. Res. 32, 1447–1460. doi: 10.1016/0042-6989(92)90201-s
McLeod, P., and Dienes, Z. (1993). Running to catch the ball. Nature 362, 23–23. doi: 10.1038/362023a0
McLeod, P., and Dienes, Z. (1996). Do fielders know where to go to catch the ball or only how to get there? J. Exp. Psychol. Hum. Percept. Perform. 22, 531–543. doi: 10.1037/0096-1523.22.3.531
McLeod, P., Reed, N., and Dienes, Z. (2006). The generalized optic acceleration cancellation theory of catching. J. Exp. Psychol. Hum. Percept. Perform. 32, 139–148. doi: 10.1037/0096-1523.32.1.139
Michaels, C. F., and Oudejans, R. R. D. (1992). The optics and actions of catching fly balls: zeroing out optical acceleration. Ecol. Psychol. 4, 199–222. doi: 10.1207/s15326969eco0404_1
Miller, W. L., Maffei, V., Bosco, G., Iosa, M., Zago, M., Macaluso, E., et al. (2008). Vestibular nuclei and cerebellum put visual gravitational motion in context. J. Neurophysiol. 99, 1969–1982. doi: 10.1152/jn.00889.2007
Moran, A., Campbell, M., and Ranieri, D. (2018). Implications of eye tracking technology for applied sport psychology. J. Sport Psychol. Action 9, 249–259. doi: 10.1080/21520704.2018.1511660
Moscatelli, A., and Lacquaniti, F. (2011). The weight of time: gravitational force enhances discrimination of visual motion duration. J. Vis. 11:5. doi: 10.1167/11.4.5
Oberle, C. D., McBeath, M. K., Madigan, S. C., and Sugar, T. G. (2005). The Galileo bias: a naive conceptual belief that influences people’s perceptions and performance in a ball-dropping task. J. Exp. Psychol. Learn. Mem. Cogn. 31, 643–653. doi: 10.1037/0278-7393.31.4.643
Orban, G. A., de Wolf, J., and Maes, H. (1984). Factors influencing velocity coding in the human visual system. Vis. Res. 24, 33–39. doi: 10.1016/0042-6989(84)90141-x
Oudejans, R. R. D., Michaels, C. F., Bakker, F. C., and Davids, K. (1999). Shedding some light on catching in the dark: perceptual mechanisms for catching fly balls. J. Exp. Psychol. Hum. Percept. Perform. 25, 531–542. doi: 10.1037//0096-1523.25.2.531
Peper, L., Bootsma, R. J., Mestre, D. R., and Bakker, F. C. (1994). Catching balls: how to get the hand to the right place at the right time. J. Exp. Psychol. Hum. Percept. Perform. 20, 591–612. doi: 10.1037//0096-1523.20.3.591
Peters, M. A. K., Ma, W. J., and Shams, L. (2016). The Size-Weight Illusion is not anti-Bayesian after all: a unifying Bayesian account. PeerJ 4:e2124. doi: 10.7717/peerj.2124
Pizlo, Z. (2001). Perception viewed as an inverse problem. Vis. Res. 41, 3145–3161. doi: 10.1016/s0042-6989(01)00173-0
Portfors-Yeomans, C. V., and Regan, D. (1996). Cyclopean discrimination thresholds for the direction and speed of motion in depth. Vis. Res. 36, 3265–3279. doi: 10.1016/0042-6989(96)00065-x
Postma, D. B. W., Den Otter, A. R., and Zaal, F. T. J. M. (2014). Keeping your eyes continuously on the ball while running for catchable and uncatchable fly balls. PLoS One 9:e92392. doi: 10.1371/journal.pone.0092392
Postma, D. B., Lemmink, K. A., and Zaal, F. T. (2018). The affordance of catchability in running to intercept fly balls. J. Exp. Psychol. Hum. Percept. Perform. 44, 1336–1347. doi: 10.1037/xhp0000531
Postma, D. B. W., Smith, J., Pepping, G.-J., van Andel, S., and Zaal, F. T. J. M. (2017). When a fly ball is out of reach: catchability judgments are not based on optical acceleration cancelation. Front. Psychol. 8:535. doi: 10.3389/fpsyg.2017.00535
Reed, N., McLeod, P., and Dienes, Z. (2010). Implicit knowledge and motor skill: what people who know how to catch don’t know. Conscious. Cogn. 19, 63–76. doi: 10.1016/j.concog.2009.07.006
Regan, D., and Beverley, K. (1978). Looming detectors in the human visual pathway. Vis. Res. 18, 415–421. doi: 10.1016/0042-6989(78)90051-2
Regan, D., and Hamstra, S. J. (1993). Dissociation of discrimination thresholds for time to contact and for rate of angular expansion. Vis. Res. 33, 447–462. doi: 10.1016/0042-6989(93)90252-r
Regan, D., and Kaushal, S. (1994). Monocular discrimination of the direction of motion in depth. Vis. Res. 34, 163–177. doi: 10.1016/0042-6989(94)90329-8
Rind, F. C., and Simmons, P. J. (1999). Seeing what is coming: building collision-sensitive neurones. Trends Neurosci. 22, 215–220. doi: 10.1016/s0166-2236(98)01332-0
Roy, J. E., and Cullen, K. E. (2003). Brain stem pursuit pathways: dissociating visual, vestibular, and proprioceptive inputs during combined eye-head gaze tracking. J. Neurophysiol. 90, 271–290. doi: 10.1152/jn.01074.2002
Rushton, S. K., and Wann, J. P. (1999). Weighted combination of size and disparity: a computational model for timing a ball catch. Nat. Neurosci. 2, 186–190. doi: 10.1038/5750
Savelsbergh, G. J. P., and Whiting, H. T. A. (1992). The acquisition of catching under monocular and binocular conditions. J. Mot. Behav. 24, 320–328. doi: 10.1080/00222895.1992.9941628
Savelsbergh, G., Whiting, H., and Bootsma, R. J. (1991). Grasping tau. J. Exp. Psychol. Hum. Percept. Perform. 17, 315–322. doi: 10.1037/0096-1523.17.2.315
Saxberg, B. (1987a). Projected free fall trajectories. I. Theory and simulation. Biol. Cybern. 56, 159–176. doi: 10.1007/BF00317991
Saxberg, B. (1987b). Projected free fall trajectories. II. Human experiments. Biol. Cybern. 56, 177–184. doi: 10.1007/BF00317992
Shaffer, D. M., Marken, R. S., Dolgov, I., and Maynor, A. B. (2013). Chasin’choppers: using unpredictable trajectories to test theories of object interception. Atten. Percept. Psychophys. 75, 1496–1506. doi: 10.3758/s13414-013-0500-7
Shaffer, D. M., and McBeath, M. K. (2005). Naive beliefs in baseball: systematic distortion in perceived time of apex for fly balls. J. Exp. Psychol. Learn. Mem. Cogn. 31, 1492–1501. doi: 10.1037/0278-7393.31.6.1492
Sharp, R. H., and Whiting, H. T. A. (1974). Exposure and occluded duration effects in a ball-catching skill. J. Mot. Behav. 6, 139–147. doi: 10.1080/00222895.1974.10734990
Smith, M. R. H., Flach, J. M., Dittman, S. M., and Stanard, T. (2001). Monocular optical constraints on collision control. J. Exp. Psychol. Hum. Percept. Perform. 27, 395–410. doi: 10.1037//0096-1523.27.2.395
Soechting, J. F., Juveli, J. Z., and Rao, H. M. (2009). Models for the extrapolation of target motion for manual interception. J. Neurophysiol. 102, 1491–1502. doi: 10.1152/jn.00398.2009
Sousa, R., Brenner, E., and Smeets, J. B. J. (2011). Judging an unfamiliar object’s distance from its retinal image size. J. Vis. 11:10. doi: 10.1167/11.9.10
Spering, M., Schütz, A. C., Braun, D. I., and Gegenfurtner, K. R. (2011). Keep your eyes on the ball: smooth pursuit eye movements enhance prediction of visual motion. J. Neurophysiol. 105, 1756–1767. doi: 10.1152/jn.00344.2010
Stocker, A. A., and Simoncelli, E. P. (2006). Noise characteristics and prior expectations in human visual speed perception. Nat. Neurosci. 9, 578–585. doi: 10.1038/nn1669
Sun, H., and Frost, B. J. (1998). Computation of different optical variables of looming objects in pigeon nucleus rotundus neurons. Nat. Neurosci. 1, 296–303. doi: 10.1038/1110
Tcheang, L., Gilson, S. J., and Glennerster, A. (2005). Systematic distortions of perceptual stability investigated using immersive virtual reality. Vis. Res. 45, 2177–2189. doi: 10.1016/j.visres.2005.02.006
Timmerman, P., and van der Weele, J. P. (1999). On the rise and fall of a ball with linear or quadratic drag. Am. J. Phys. 67, 538–546. doi: 10.1119/1.19320
Todd, J. T. (1981). Visual information about moving objects. J. Exp. Psychol. Hum. Percept. Perform. 7, 795–810. doi: 10.1037/0096-1523.7.4.795
Todorov, E. (2004). Optimality principles in sensorimotor control. Nat. Neurosci. 7, 907–915. doi: 10.1038/nn1309
Turvey, M. T., Shaw, R. E., Reed, E. S., and Mace, W. M. (1981). Ecological laws of perceiving and acting: in reply to Fodor and Pylyshyn (1981). Cognition 9, 237–304. doi: 10.1016/0010-0277(81)90002-0
Von Helmholtz, H. (1867). Treatise on Physiological Optics Vol. III. New York, NY: Dover Publications.
Wann, J. P. (1996). Anticipating arrival: is the tau margin a specious theory? J. Exp. Psychol. Hum. Percept. Perform. 22, 1031–1048. doi: 10.1037//0096-1523.22.4.1031
Warren, W. H. Jr., Kay, B. A., Zosh, W. D., Duchon, A. P., and Sahuc, S. (2001). Optic flow is used to control human walking. Nat. Neurosci. 4, 213–216. doi: 10.1038/84054
Watson, J. S., Banks, M. S., von Hofsten, C., and Royden, C. S. (1992). Gravity as a monocular cue for perception of absolute distance and/or absolute size. Perception 21, 69–76. doi: 10.1068/p210069
Wei, X.-X., and Stocker, A. A. (2015). A Bayesian observer model constrained by efficient coding can explain ‘anti-Bayesian’ percepts. Nat. Neurosci. 18, 1509–1517. doi: 10.1038/nn.4105
Werkhoven, P., Snippe, H. P., and Alexander, T. (1992). Visual processing of optic acceleration. Vis. Res. 32, 2313–2329. doi: 10.1016/0042-6989(92)90095-z
Westheimer, G., and McKee, S. P. (1977). Spatial configurations for visual hyperacuity. Vis. Res. 17, 941–947. doi: 10.1016/0042-6989(77)90069-4
Whiting, H. T. A. (1968). Training in a continuous ball throwing and catching task. Ergonomics 11, 375–382. doi: 10.1080/00140136808930985
Whiting, H. T. A., and Sharp, R. H. (1974). Visual occlusion factors in a discrete ball-catching task. J. Mot. Behav. 6, 11–16. doi: 10.1080/00222895.1974.10734974
Wilkie, R., and Wann, J. (2003). Controlling steering and judging heading: retinal flow, visual direction, and extraretinal information. J. Exp. Psychol. Hum. Percept. Perform. 29, 363–378. doi: 10.1037/0096-1523.29.2.363
Wolpert, D. M., Ghahramani, Z., and Jordan, M. I. (1995). An internal model for sensorimotor integration. Science 269, 1880–1882. doi: 10.1126/science.7569931
Yonas, A., Bechtold, A. G., Frankel, D., Gordon, F. R., McRoberts, G., Norcia, A., et al. (1977). Development of sensitivity to information for impending collision. Percept. Psychophys. 21, 97–104. doi: 10.3758/bf03198713
Zaal, F. T. J. M., Bongers, R. M., Pepping, G.-J., and Bootsma, R. J. (2012). Base on balls for the Chapman strategy: reassessing Brouwer, Brenner, and Smeets (2002). Atten. Percept. Psychophys. 74, 1488–1498. doi: 10.3758/s13414-012-0328-6
Zaal, F. T. J. M., and Bootsma, R. J. (2011). Virtual reality as a tool for the study of perception-action: the case of running to catch fly balls. Presence: Teleoperators and Virtual Environ. 20, 93–103. doi: 10.1162/pres_a_00037
Zago, M., Bosco, G., Maffei, V., Iosa, M., Ivanenko, Y. P., and Lacquaniti, F. (2004). Internal models of target motion: expected dynamics overrides measured kinematics in timing manual interceptions. J. Neurophysiol. 91, 1620–1634. doi: 10.1152/jn.00862.2003
Zago, M., Bosco, G., Maffei, V., Iosa, M., Ivanenko, Y. P., and Lacquaniti, F. (2005). Fast adaptation of the internal model of gravity for manual interceptions: evidence for event-dependent learning. J. Neurophysiol. 93, 1055–1068. doi: 10.1152/jn.00833.2004
Zago, M., and Lacquaniti, F. (2005). Internal model of gravity for hand interception: parametric adaptation to zero-gravity visual targets on earth. J. Neurophysiol. 94, 1346–1357. doi: 10.1152/jn.00215.2005
Zago, M., McIntyre, J., Senot, P., and Lacquaniti, F. (2008). Internal models and prediction of visual gravitational motion. Vis. Res. 48, 1532–1538. doi: 10.1016/j.visres.2008.04.005
Zhao, H., Straub, D., and Rothkopf, C. A. (2019). The visual control of interceptive steering: how do people steer a car to intercept a moving target? J. Vis. 19:11. doi: 10.1167/19.14.11
Keywords: 3D perception, calibration, internal models, optic flow, prior knowledge, TTC
Citation: Aguado B and López-Moliner J (2021) Gravity and Known Size Calibrate Visual Information to Time Parabolic Trajectories. Front. Hum. Neurosci. 15:642025. doi: 10.3389/fnhum.2021.642025
Received: 15 December 2020; Accepted: 28 July 2021;
Published: 23 August 2021.
Edited by:
Paul Michael Corballis, The University of Auckland, New ZealandReviewed by:
Frank T. J. M. Zaal, University of Groningen, NetherlandsDennis Shaffer, Ohio State University at Mansfield, United States
Copyright © 2021 Aguado and López-Moliner. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Joan López-Moliner, j.lopezmoliner@ub.edu