Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 21 May 2020
Sec. Perception Science
This article is part of the Research Topic Spatial and Temporal Perception in Sensory Deprivation View all 18 articles

Brain Activation for Audiovisual Information in People With One Eye Compared to Binocular and Eye-Patched Viewing Controls

  • 1Department of Psychology, York University, Toronto, ON, Canada
  • 2Centre for Vision Research, York University, Toronto, ON, Canada
  • 3The Hospital for Sick Children, Toronto, ON, Canada
  • 4School of Kinesiology and Health Science, York University, Toronto, ON, Canada

Blindness caused by early vision loss results in complete visual deprivation and subsequent changes in the use of the remaining intact senses. We have also observed adaptive plasticity in the case of partial visual deprivation. The removal of one eye, through unilateral eye enucleation, results in partial visual deprivation and is a unique model for examining the consequences of the loss of binocularity. Partial deprivation of the visual system from the loss of one eye early in life results in behavioral and structural changes in the remaining senses, namely auditory and audiovisual systems. In the current study we use functional neuroimaging data to relate function and behavior of the audiovisual system in this rare patient group compared to controls viewing binocularly or with one eye patched. In Experiment 1, a whole brain analysis compared common regions of cortical activation between groups, for auditory, visual and audiovisual stimuli. People with one eye demonstrated a trend for increased activation for low-level audiovisual stimuli compared to patched viewing controls but did not differ from binocular viewing controls. In Experiment 2, a region of interest (ROI) analysis for auditory, visual, audiovisual and illusory McGurk stimuli revealed that people with one eye had an increased trend for left hemisphere audiovisual activation for McGurk stimuli compared to binocular viewing controls. This aligns with current behavioral analysis and previous research showing reduced McGurk Effect in people with one eye. Furthermore, there is no evidence of a correlation between behavioral performance on the McGurk Effect task and functional activation. Together with previous behavioral work, these functional data contribute to the broader understanding of cross-sensory effects of early sensory deprivation from eye enucleation. Overall, these results contribute to a better understanding of the sensory deficits experienced by people with one eye, as well as, the relationship between behavior, structure and function in order to better predict the outcome of early partial visual deafferentation.

Introduction

Complete visual deprivation from blindness leads to adaptive changes in other sensory systems. For example, congenitally blind individuals have shorter response times for auditory discrimination tasks (Röder et al., 1999), faster processing of language (Röder et al., 2002), enhanced sound localization (Lessard et al., 1998) and enhanced tactile perception (Sathian, 2000; Goldreich and Kanics, 2003) compared to sighted individuals. These adaptations suggest that underlying physiological changes have occurred within sensory systems to support such behavioral enhancements. It is possible that visual cortex is recruited or reorganized by other sensory systems in the congenitally blind. Neuroimaging studies have shown activation of visual cortex for sensory stimuli normally processed elsewhere in the brain such as audition (Collignon et al., 2009; Merabet et al., 2009), sound localization (Weeks et al., 2000), and tactile perception and Braille reading (Sadato et al., 1996; Cohen et al., 1997; Buchel et al., 1998; Kupers et al., 2007). Not all recruitment or reorganization results in adaptive change as evidence for the disruption of complementary senses when the visual system is compromised also exists. For example, some have shown congenitally blind individuals have decreased sound localization accuracy in the vertical plane (Lewald, 2002), and horizontal plane (Gori et al., 2014) or decreased distance judgment of auditory stimuli (Wanet and Veraart, 1985). Overall, it appears that in the case of complete sensory deprivation, specifically blindness, it is possible for other intact sensory systems to be altered and in some cases, to adaptively compensate for the loss of vision. One might ask whether such neuroplasticity also holds true in cases of partial sensory deprivation, such as the loss of one eye early in life?

Partial visual deprivation from unilateral eye enucleation, the surgical removal of one eye, is a unique model for examining the consequences of the loss of binocularity (see Steeves et al., 2008, for a review). It is unlike other forms of monocular visual deprivation such as cataract or strabismus which leave abnormal binocular input and contributes to competitive binocular interactions. Surgically removing the eye completely eliminates all forms of visual input to the brain from that eye leaving a single stream of information to the visual system and a complete lack of competitive binocular interactions (Steeves et al., 2008). Early monocular enucleation is a useful model of study since the visual system may not have been exposed to abnormal visual input from the removed eye.

Monocular enucleation during postnatal visual system maturation leads to both enhancements and reductions in visual function. These differences in outcome appear to align with whether one is measuring visual spatial ability or visual motion processing and oculomotor systems (reviewed in Steeves et al., 2008; Kelly et al., 2013). Visual spatial ability is largely intact while visual motion processing and oculomotor systems show small deficits. More recently, a number of behavioral studies of people who have only one eye have assessed abilities outside of the visual system, specifically, audiovisual abilities. These studies aimed to investigate what types of accommodations the brain might make across the senses after losing half of its visual input (see Steeves et al., 2008; Kelly et al., 2013).

Within the auditory domain, people with one eye have enhanced sound localization in all locations (within 78 degrees to the left or right of straight ahead) except for the extreme periphery compared to control participants who were binocular viewing, eye-patched or had both eyes closed (Hoover et al., 2012). In terms of audiovisual processing, people with one eye do not show the typical pattern of visual dominance when asked to categorize rapidly presented audiovisual targets, but rather, they show equivalent auditory and visual processing suggesting an enhanced relative weighting to the auditory component of bimodal stimuli (Moro and Steeves, 2012). These results persist even when the temporal load is increased in the same task by asking participants to detect and discriminate auditory, visual, or bimodal repetitions in a one-back task (Moro and Steeves, 2013). People with one eye do not differ in the width of their temporal binding window compared to binocular and eye-patched viewing controls, however, they have longer response latencies relative to controls indicating a longer processing time required for this task. Eye-patched controls’ response latencies were intermediate to the two other groups (Moro and Steeves, 2018c). Despite no difference in width of temporal binding window, people with one eye are also less susceptible to the double flash illusion compared to both binocular and eye-patched viewing controls. Furthermore, in that task, people with one eye responded as quickly as binocular and eye-patched viewing controls (Moro and Steeves, 2018c).

People with one eye show no difference in variance of audiovisual localization along the horizontal plane compared to binocular and patched viewing control groups (Moro et al., 2014). However, unlike binocular and eye-patched controls, they take longer to localize unimodal visual stimuli compared to unimodal auditory stimuli (Moro et al., 2014). In terms of audiovisual motion in depth, people with one eye demonstrate the same rate of dynamic visual capture (perception of the direction of an auditory signal to be moving in the direction of the incongruent visual signal despite being asked to respond to the auditory signal alone) (Moro and Steeves, 2018b). Unlike static audiovisual localization, people with one eye have no difference in reaction time or accuracy compared to both control groups for this task (Moro and Steeves, 2018b). Together these audiovisual behavioral studies indicate that perhaps task requirements affect behavioral outcomes for this patient group. Both localization studies used low-level flash and beep stimuli, and people with one eye did not differ in overall performance compared to control groups but did perform slower on tasks with less ecological validity (sounds and images moving along the horizontal plane) (Moro et al., 2014) compared to those moving in depth (Moro and Steeves, 2018b).

To increase ecological validity of audiovisual stimuli, face and voice processing has been studied. People with one eye have increased sensitivity to voices on their own (but not non-human sounds, specifically car horns) in a face-voice and car-horn recognition task (Moro and Steeves, 2019). Perhaps this increased sensitivity to voices compensates for the mild face processing deficits in discriminating feature spacing, the face composite effect and face processing time that have previously been found (Kelly et al., 2012). Finally, people with one eye perceive the illusory McGurk effect less often than binocular viewing controls (Moro and Steeves, 2018a). Additionally, they have no difference in reaction time compared to both control groups (Moro and Steeves, 2018a). Overall, these behavioral results might suggest forms of behavioral adaptation following the reduction of visual input from one eye early in life.

No other lab has investigated changes in brain structure in people who have had one eye removed early in life. Not surprisingly, significant degeneration of the anterior visual system, including decreased optic chiasm volume and width is found in people with one eye compared to binocular controls (Kelly et al., 2014). People with one eye also have an overall decrease in lateral geniculate nucleus (LGN) volume compared to binocular controls as expected but surprisingly, the LGN volume contralateral to the remaining eye is less reduced likely from recruitment of deafferented LGN cells by the intact eye (Kelly et al., 2014). These findings indicate that subcortical level reorganization of the visual system occurs after losing one eye early in life (Kelly et al., 2014). At a cortical level, a subsequent study revealed that, compared to binocular viewing controls, people with one eye have increased surface area and gyrification in visual, auditory and multisensory cortices (Kelly et al., 2015). White matter tracts in the visual and auditory systems of people with one eye were examined using Diffusion tensor imaging (DTI) (Wong et al., 2018, 2019). White matter tracts are greater contralateral to the surgically removed eye in the optic radiations, V1-LGN projections and interhemispheric V1 projections of people with one eye compared to binocular viewing controls, likely a reflection of the differences observed in the LGN volume and optic tract contralateral to the removed eye (Kelly et al., 2014; Wong et al., 2018). Auditory wiring appears more substantial than in controls and, further, the connections between the visual and auditory systems are more intact than expected (Wong et al., 2019). Unlike controls, people with one eye have an asymmetric medial geniculate body (MGB) volume with a larger left than right MGB, regardless of which eye was enucleated perhaps reflecting dominance of left hemisphere in auditory processing (Moro et al., 2015). Taken together, there is moderate cortical, subcortical and wiring alterations of auditory and visual processing following early eye enucleation.

In terms of brain function more recently, our lab found reduced functional activation in people with one eye compared to binocular viewing controls in face-preferential brain regions [left fusiform face area (FFA) and bilateral occipital face area (OFA)] (Kelly et al., 2019). These results complement the mild behavioral face deficits in people with one eye (Kelly et al., 2012). The current study examines audiovisual functional activation in people with one eye compared to binocular and patched viewing controls in two separate experiments. Experiment 1 investigates differences in activation intensity in regions of interest localized by conjunction analysis between groups for low-level audiovisual stimuli and high-level face and voice stimuli. Experiment 2 probes audiovisual regions of interest (ROIs) during the presentation of auditory, visual, audiovisual and illusory McGurk stimuli. We predict that functional activation will reflect our previous behavioral findings, specifically relevant is the absence of a McGurk effect in this group (Moro and Steeves, 2018a). Results from this study will provide a better understanding of how people with one eye process auditory and visual information contributing to better clinical outcomes through cross-sensory accommodations and the promotion of long-term visual health in the remaining eye.

Materials and Methods

Participants

Monocular Enucleation Group

Seven adult participants who had undergone monocular enucleation (ME) at The Hospital for Sick Children participated in this study (mean age = 34 years, SD = 13 years). All ME participants with one eye had been unilaterally eye enucleated (four right eye removed) due to retinoblastoma, a rare childhood cancer of the retina. Age at enucleation ranged from 4 to 60 months (mean age at enucleation = 23 months, SD = 18 months).

Binocular Viewing Control Group (BV)

Ten binocularly intact controls with a mean age of 32 years (SD = 13 years) were tested viewing stimuli out of both eyes.

Monocular Viewing Control Group (MV)

Ten binocularly intact participants, with a mean age of 31 years (SD = 16 years), completed the experiments with one eye patched. Participants’ non-preferred eye was patched with a semi-opaque eye covering and translucent tape (five right-eye covered).

All participants (ME, BV, MV) reported normal hearing and normal or corrected-to-normal acuity as assessed by an EDTRS eye chart (Precision VisionTM, La Salle, IL, United States) and wore optical correction if needed. All participants gave informed consent prior to their inclusion in the study, which was approved by York University’s Office of Research Ethics.

Stimuli

Audiovisual Blocked Design

Low-level stimuli

Visual stimuli consisted of a white ring on a black background with a visual fixation cross in the center. Each visual stimulus was 17 ms in length. The auditory stimulus consisted of a 3500 Hz pure tone with a duration of 13 ms. Stimuli were repeated eight times per trial for a duration of 2 s. Trials were presented in a block design. Each block consisted of eight trials (16 s/block). Blocks consisting of visual only, auditory only and audiovisual and asynchronous audiovisual (sound was displaced by 500 ms) stimuli were presented (Figure 1A). A 16 s interstimulus interval of silence and a blank screen was presented between each block.

FIGURE 1
www.frontiersin.org

Figure 1. (A) A schematic illustration of the stimuli used in the audiovisual blocked design study. (B) A schematic illustration of the presentation of stimuli used in the rapid event related design study. Stimuli used in the rapid event related design study have been previously published (see Quinto et al., 2010; Stevenson et al., 2012; Moro and Steeves, 2018a). All visual stimuli were presented to participants in color.

High-level stimuli

Visual stimuli consisted of a 16 s video of a female speaker reading an excerpt from a children’s story. Auditory stimuli consisted of a 16 s audio clip of the female speaker from the video saying the story aloud in a clear and articulate voice. Audiovisual stimuli consisted of two 16 s videos of the female speaker reading the excerpt with the auditory component of the video either synchronously or asynchronously (auditory presented 500 ms after the visual stimulus) paired with the corresponding video. Participants viewed one repetition per trial. Trials were presented in a block design. Each block consisted of one trial (16 s; Figure 1A). A 16 s interstimulus interval consisting of silence and a blank screen was presented after each block.

Rapid-Event Related Design

Visual stimuli consisted of two 2 s videos of a female speaker mouthing the syllables “ba” and “ga,” with each presentation containing the entire articulation of the syllable (Quinto et al., 2010; Stevenson et al., 2012; Moro and Steeves, 2018a). Auditory stimuli consisted of 2 s audio clips of the female speaker from the videos saying the syllables “ba” and “ga”. Audiovisual stimuli consisted of two 2 s videos of the female speaker saying the syllables “ba” and “ga,” paired with the corresponding video, respectively. McGurk illusory stimuli consisted of video footage of the female speaker mouthing the “ga” syllable but paired with the auditory sound clip of the female speaker saying “ba” (Figure 1B). Stimuli were presented using a rapid-event related design with jittered interstimulus intervals of variable lengths up to 15 s in order to improve the sampling of the hemodynamic response function (HRF). All stimuli were counterbalanced using OptSeq2 (Greve, 2002). Each stimulus condition was presented 15 times per run.

Procedure

All participants were scanned at York University’s Sherman Health Science Research Center with a Siemens MAGNETOM Trio 3T MRI scanner (Siemens, Erlangen, Germany) using a 32-channel high-resolution brain array coil. BOLD (blood-oxygen-level dependent) fMRI imaging was utilized to acquire functional images. An echoplanar imaging sequence with the following specifications was used to obtain functional volumes: 35 contiguous axial slices; in-plane resolution 3 × 3 mm; slice thickness 3.5 mm; TR = 2000 ms; TE = 30 ms; imaging matrix 96 x 96; flip angle 90°; FoV = 192 mm. Following the functional scans, a high-resolution whole brain structural image was obtained with a T1 magnetization-prepared rapid gradient echo imaging sequence. The anatomical imaging had the following parameters: 192 slices; in-plane resolution 1 × 1 mm; slice thickness 1 mm; TR 1900 ms; TE 2500 ms; imaging matrix 256 × 256; flip angle 9°; FoV = 256 mm.

Experimental stimuli were presented using VPixx visual testing software (VPixx Technologies Inc., Montréal, QC, United States) via a 33 × 19.5 cm screen inside of the scanner and noise-canceling headphones (MR Confon GmbH, Magdeburg, Germany). Prior to scanning, sound samples were presented to participants to ensure that the sound pressure level was audible and comfortable. Participants were instructed to press a button on a Current Designs 8-Button Bimanual Curved Lines button box (Current Designs, Philadelphia, PA, United States) whenever they perceived the stimuli to be asynchronous to ensure that they were alert and paying attention throughout the task. Each participant performed in seven experimental runs. Five runs consisted of the audiovisual blocked design stimuli and two runs consisted of the rapid-event related McGurk stimuli. All runs were presented to each participant in counterbalanced order.

Data Analysis

Experiment 1: Audiovisual Block Design Whole Brain and ROI Analyses

During MRI data collection, estimates of head motion (translation and rotation) were viewed in real time in the MRI control room to verify that head movement did not exceed 1 mm in any direction. Participants who exceeded this threshold for head motion repeated imaging runs when necessary. For Experiment 1, analysis of MRI data was performed using BrainVoyager v20.6 (Brain Innovation, Maastricht, Netherlands). Preprocessing of the audiovisual block design functional data included slice time correction followed by motion correction, and linear trend removal. Motion correction used a trilinear/sinc interpolation method with the first functional volume used as the reference. Plots of head motion estimates and movies of head motion over the course of each functional run were generated and visually inspected to confirm that all experimental runs were free from head movements over 1 mm in any direction and free from obvious hardware-related artifacts. Spatial smoothing was applied to each functional run using a 6 mm full width half maximum isotropic kernel. Functional runs were coregistered with corresponding high resolution T1-weighted anatomical images. Images were transformed from subject-space to Talairach template space.

Design matrices for use in general linear model (GLM) analyses were constructed for each participant using a boxcar design convolved with a hemodynamic response function. Within the design matrices, stimulation timing protocols were used to define separate predictors for each of the eight experimental conditions (i.e., low-level and high-level auditory, visual, synchronous audiovisual, and asynchronous audiovisual stimuli). Non-parametric permutation testing was performed for whole brain group comparisons of each condition using the randomize plug-in for BrainVoyager. Four experimental runs were included in this analysis for each participant (one experimental run was excluded for use as a functional localizer, as described in the following section). Since the randomize plug-in does not support ANOVA testing, comparisons between groups were performed using two-sample unpaired t-tests with 1000 permutations, threshold free cluster enhancement (TFCE), and FDR thresholding of q < 0.01.

To perform region of interest (ROI) analyses, one of the audiovisual block design imaging runs collected from each participant was used as a functional localizer. Group analyses were performed on these localizer runs using multi-subject GLMs computed separately for each of the three experimental groups (i.e., ME, BV, MV). This approach resulted in a group mean map of activity associated with each experimental contrast of interest for each group of participants. Maps to localize the auditory cortex were produced by using activity associated with the unimodal auditory stimuli condition relative to baseline. The visual cortex was localized using activity associated with the unimodal visual stimuli condition relative to baseline. “Audiovisual” regions were isolated by contrasting activity in the synchronous audiovisual condition versus activity in the auditory and visual conditions (i.e., audiovisual > auditory + visual). These group maps were imported into Neuroelf v1.11 where conjunction maps were computed for (1) the BV and ME participant groups, (2) the MV and ME groups, and (3) the BV and MV groups to produce ROIs associated with each of the contrasts described above. All voxels included in ROIs had a p-value equal to or lower than 0.05 in included groups of participants. This approach was used to help mitigate the possibility of biasing ROI masks toward any one participant group (i.e., to prevent the inclusion of voxels that were only active in one participant group). These ROI masks were then imported back into BrainVoyager where they were applied to analyses of the remaining four functional imaging runs from each participant. Group ROI GLM analyses were performed using these imaging runs, and ANOVAs were used to test for group-related differences in mean beta weight values associated with each region.

Experiment 2: McGurk Region of Interest Analysis

Data associated with Experiment 2 were analyzed using the Analysis of Functional Neuroimages (AFNI) software package (Cox, 1996). All five audiovisual block design functional runs were used to create region of interest masks for ROI analyses of the McGurk functional imaging runs. Preprocessing of the audiovisual block design runs included slice time correction, coregistration of functional and anatomical images, transformation of images from subject space to Talairach space, motion correction using coregistration of each functional volume to the volume with the minimum outlier fraction, masking of functional data, within mask spatial smoothing using a 4 mm full width half maximum Gaussian kernel, and scaling of intensity values to a mean value of 100. Pairs of volumes where the Euclidean Norm of the motion derivative exceeded 0.3 mm were removed (in practice, this resulted in very few volumes being removed), and motion estimate parameters were included as regressors of no interest to account for motion-correlated variance in the data. In addition, volumes containing greater than 10% outlier voxel fraction were removed (again resulting in few volumes being removed in practice). Stimulation timing for each condition was provided for GLM analyses as described in Experiment 1. Masks of left and right hemisphere superior temporal regions associated with audiovisual processing were created using a conjunction analysis to isolate voxels equally active (p < 0.05) in both the Auditory stimuli and Visual stimuli conditions for each individual study participant. These ROIs were applied in analyses of the McGurk stimuli functional runs (see Figure 2).

FIGURE 2
www.frontiersin.org

Figure 2. Relative locations of superior temporal gyrus “audiovisual” regions of interest for each participant group. Regions are plotted as spheres (radius = 6 mm) centerd on the mean Talairach coordinates for each group. Left hemisphere regions are shown on an axial slice at Talairach z = –2 and right hemisphere regions are shown on an axial slice at z = –5. Mean coordinates for the Monocular Enucleation Group (ME) are shown in yellow, Binocular Viewing Controls (BV) in red, and eye-patched Monocular Viewing Controls (MV) in blue. Overlap of the right hemisphere ME and MV regions is shown in green.

Preprocessing of the McGurk runs was performed as described above for the audiovisual block design runs. For these data, a predictor was defined for each condition (i.e., auditory only, visual only, synchronous audiovisual, and McGurk stimuli) and the timing of each stimulus was used in deconvolution of the rapid event-related structure of these imaging runs. Tent functions for deconvolution analysis were centered at TR times. For each participant, ROI masks (localized using the audiovisual block design runs) were applied to the preprocessed McGurk runs. For each condition, mean beta weights associated with each ROI were extracted for each participant. SPSS was used to perform ANOVA comparisons of mean beta weights between participant groups.

Results

None of the whole brain or ROI analyses performed resulted in any statistically significant differences between any of the participant groups tested. Due to the rare patient group involved in this study and the subsequent small sample size, only descriptive statistics are reported. The data reported are the median and interquartile range (IR) mean beta weight signal in each region. The data reports only participants with an active ROI. See Table 1 (Experiment 1) and Table 2 (Experiment 2) for each group including mean cluster size and TAL coordinates.

TABLE 1
www.frontiersin.org

Table 1. Conjunction ROIs containing voxels significantly activated in both participant groups with mean cluster size (SD) and mean Talairach coordinates (STG = superior temporal gyrus, IFG = inferior frontal gyrus).

TABLE 2
www.frontiersin.org

Table 2. Total number of participants with active ROI, mean (SD) cluster size (mm3) and mean (SD) Talairach coordinates for experiment 2: rapid event related design.

Experiment 1: Whole Brain Analysis

A conjunction analysis on functional localizer runs was conducted to isolate overlapping areas of functional activation for auditory, visual and audiovisual stimuli between groups. The overlapping ROIs localized for each group comparison are depicted in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3. Relative locations of areas isolated using conjunction analyses of functional activation for auditory, visual and audiovisual stimuli between groups. Regions are plotted as spheres (radius = 6 mm) centerd on the mean Talairach coordinates for each pair of groups. Overlapping areas of functional activation for the conjunction analysis between BV and ME groups are shown on an axial slice at Talairach z = 6 in shades of orange. Overlapping areas of functional activation for the conjunction analysis between MV and ME groups are shown on an axial slice at Talairach z = 7 in shades of green. Overlapping areas of functional activation for the conjunction analysis between BV and MV groups are shown on an axial slice at Talairach z = 6 and z = 33) in shades of purple.

For the BV group (n = 10) compared to the ME group (n = 7) five ROIs were identified with significant activation in common between groups. These ROIs were applied to the remaining four experimental runs that were not used for ROI localization. People with one eye did not differ in intensity of cortical activation in these common regions of interest compared to binocular viewing controls (see Table 3). Figure 4A shows that the majority of people with one eye were within the 95% confidence interval (CI) of the binocular viewing control group for all identified ROIs.

TABLE 3
www.frontiersin.org

Table 3. Median and interquartile range of the mean beta weight signal in each overlapping area of functional activation for auditory, visual and audiovisual stimuli isolated using conjunction analyses in localizer runs.

FIGURE 4
www.frontiersin.org

Figure 4. Box and whisker plots of the distribution of peak beta weights for the control groups for the identified auditory, visual and audiovisual ROIs. The horizontal line within each box represents the median of the BV group (A,C) and the MV group (B). The boxes correspond to the 25th and 75th percentiles, the whiskers correspond to the 5th and 95th percentiles. Individual data points for the ME group are plotted (A,B) and MV group (C).

For the MV group (n = 10) compared to the ME group (n = 7) five ROIs were identified with significant activation in common between groups. People with one eye had increased intensity of activation for low-level audiovisual stimuli compared to patched viewing controls (see Table 3). Figure 4B shows that 5/7 and 6/7 ME participants were above the 95% CI of the MV group for the low-level audiovisual ROIs 1 and 2, respectively.

For the BV group (n = 10) compared to the MV group (n = 10) seven ROIs were identified with significant activation in common between groups. Patched viewing controls did not differ in the intensity of cortical activation in these common regions of interest compared to binocular viewing controls (see Table 3). Figure 4C shows that 5/10 MV participants were below the 95% CI of the BV group for the high-level audiovisual ROI.

Experiment 2: McGurk Region of Interest Analysis

A region of interest analysis was conducted for individual functionally localized audiovisual regions in the left and right hemisphere. Intensity of activation was compared between groups for auditory, visual, audiovisual and McGurk illusory stimuli. Median and interquartile range (IR) mean beta weight signal in each region is listed for each stimulus type in Table 4.

TABLE 4
www.frontiersin.org

Table 4. Median and interquartile range of the mean beta weight signal in each ROI for experiment 2: region of interest analysis.

Left and right audiovisual ROIs were identified for the ME (left, n = 6; right, n = 6) group, BV (left, n = 9; right, n = 8) group and MV (left, n = 9; right, n = 8) group. People with one eye have increased intensity of activation in the left audiovisual ROI for McGurk stimuli compared to binocular viewing controls. Patched viewing controls did not differ in intensity of activation compared to people with one eye. Patched viewing controls have increased intensity of activation in the left audiovisual ROI for McGurk stimuli compared to binocular viewing controls. Figure 5 (A, B) show that 5/6 people with one eye were outside the 95% confidence interval (CI) of the binocular viewing control group for the McGurk stimuli for the left audiovisual ROI. Panels (C, D) indicate that the majority of people with one eye were within the 95% confidence interval (CI) of the patched viewing control group. Panels (D, E) indicate that the majority of patched viewing controls were within the 95% confidence interval (CI) of the binocular viewing control group.

FIGURE 5
www.frontiersin.org

Figure 5. Box and whisker plots of the distribution of peak beta weights for the auditory, visual, audiovisual and McGurk stimulus conditions for the left audiovisual ROI [BV group = (A,E); MV group = (C)] and right audiovisual ROI [BV group = (B,F); MV group = (D)]. The horizontal line within each box represents the median. The boxes correspond to the 25th and 75th percentiles, the whiskers correspond to the 5th and 95th percentiles. Individual data points for ME group are plotted (A–D) and MV group (E,F).

McGurk Effect Behavioral Analysis

Behavioral performance recorded during the scan session was analyzed. Two of the original 10 MV participants were removed from this data analysis due to performance under chance. A Greenhouse-Geisser corrected (X2(2) = 13.383, p = 0.001), 3 × 3 repeated measures analysis of variance (ANOVA) comparing group (ME vs BV vs MV) and McGurk condition (“Ba”, “Ga”, “Da”) revealed a significant interaction, F(2.719, 29.906) = 6.624, p = 0.002, ηp2 = 0.376 and main effect of McGurk condition, F(1.359, 29.906) = 26.313, p < 0.000, ηp2 = 0.545. There was no significant main effect of participant group, F(2, 22) = 1.043, p = 0.369, ηp2 = 0.087. The BV group demonstrated increased perception of the illusory “Da” condition compared “Ba” (p < 0.000) and “Ga” (p < 0.000). The MV group demonstrated increased perception of the illusory “Da” condition compared “Ba” (p = 0.000) and “Ga” (p < 0.000). Furthermore, the ME group did not demonstrate an increase in perception of the “Da” condition compared to the “Ba” condition (p < 1.000) and the “Ga” condition (p < 1.000). Overall these results indicate a replication of the findings of our previous behavioral study (Moro and Steeves, 2018a). Figure 6 plots the behavioral data for each participant group.

FIGURE 6
www.frontiersin.org

Figure 6. Behavioral McGurk effect perceived (number of times a participant perceived “da” during McGurk trials) for each of the BV (black), MV (gray) and ME group (light gray).

Behavioral Performance and Functional Activation Correlations

We investigated the relationship between the current behavioral McGurk data obtained during the fMRI session for the BV, MV, and ME groups and the current peak beta weight data for left and right audiovisual ROI’s with McGurk stimuli. Since both of the control groups (BV and MV) demonstrated a McGurk effect and the patient group (ME) did not, we decided to collapse the data and conduct an omnibus correlation in order to accommodate for the small sample size. Non-parametric Spearman correlations indicate a significant correlation, rs(24) = −0.620, p = 0.002 for left audiovisual activation compared to behavioral performance and non-significant correlation, rs(22) = −0.114, p = 0.615 for right audiovisual activation to behavioral performance. Figure 7 plots the behavioral and peak beta weight correlations for the left and right audiovisual ROI.

FIGURE 7
www.frontiersin.org

Figure 7. McGurk behavioral performance during scan sessions correlated with peak beta weight activation in the left (column (A)) and right (column (B)) audiovisual ROI during McGurk trials. The BV group is represented with a circle marker, the MV group is represented with a triangle marker and the ME group is represented with a square marker.

Discussion

The current study investigated whether people who had one eye surgically removed early in life have altered functional activation for auditory, visual and audiovisual stimuli. In Experiment 1, a region of interest analysis using ROIs localized with group conjunction analyses was conducted to compare overlapping areas of functional activation for auditory, visual and audiovisual stimuli. When comparing people with one eye to binocular viewing controls five common regions of interest were identified. People with one eye did not differ in intensity of cortical activation in these common regions of interest compared to binocular viewing controls. When comparing people with one eye to patched viewing controls five common regions of interest were identified. People with one eye had a trend for increased intensity of activation When comparing patched viewing controls to binocular viewing controls seven common regions of interest were identified. Patched viewing controls did not differ in the intensity of cortical activation in these common regions of interest compared to binocular viewing controls.

In Experiment 2, a region of interest analysis was conducted for individual functionally localized audiovisual regions in the left and right hemisphere. Intensity of activation was compared between groups for auditory, visual, audiovisual and McGurk illusory stimuli. Both people with one eye and patched viewing controls have a trend for increased intensity of activation in the left audiovisual ROI for McGurk stimuli compared to binocular viewing controls. Patched viewing controls did not differ in intensity of activation compared to people with one eye. Additionally, behavioral performance recorded during the scan session indicates a replication of previous findings (Moro and Steeves, 2018a) where people with one eye do not perceive the McGurk Effect. Correlating behavioral performance on the McGurk task with functional activity yielded a significant negative correlation for the left audiovisual ROI and no significant correlation for the right audiovisual ROI.

The current analyses indicate that in common areas of activation people with one eye and binocular viewing controls do not differ in intensity of activation despite previously observed audiovisual behavioral (Moro and Steeves, 2012, 2013, 2018a,b,c, 2019; Moro et al., 2014) and structural (Kelly et al., 2014, 2015; Moro et al., 2015; Wong et al., 2018, 2019) differences. The lack of difference in activation intensity between these two groups provides evidence that behavioral differences can nonetheless exist in the absence of functional differences.

In contrast, intermediate behavioral performance has been observed for eye patched controls where they fall in between that of people with one eye and binocular viewing controls (Moro et al., 2014; Moro and Steeves, 2018a,b,c, 2019). Consistent with this, the present neuroimaging study indicates functional differences between people with one eye and patched viewing controls. Specifically, people with one eye have a trend toward increased activation intensity in common regions for low-level audiovisual stimuli compared to patched viewing controls. This indicates that short term partial visual deprivation from wearing an eye patch may have a more negative effect on functional activation than long term partial visual deprivation from unilateral eye enucleation. Previous research on monocular deprivation in binocular viewing participants indicates the presence of neuroplasticity in the visual cortex even after short term monocular deprivation through eye patching (see Castaldi et al., 2020 for review). For example, evidence of strengthened cortical excitability after short term monocular deprivation (Lunghi et al., 2015) and enhanced BOLD V1 activation for high spatial frequency stimuli (Binda et al., 2018). These results are restricted to V1, V2, V3, and V4 while not present in V3a and hMT+ (Binda et al., 2018). Results of the current study may indicate that outside of the visual cortex, a decrease in activation intensity may be present. We have speculated that the previously observed intermediate behavioral performance may be the result of binocular inhibitory interactions from wearing an opaque eye patch (Steeves et al., 2004) and may be reflected by reduced functional activation when temporarily wearing an eye patch. Further studies investigating the relationship between behavior and function, within both visual and audiovisual processing regions in long and short term partial visual deprivation should be considered.

A popular tool for studying the mechanisms underlying multisensory integration is the McGurk effect. Susceptibility to the illusion is often inconsistent and shows inter-subject variability possibly due to different cognitive processes that are being used (Beauchamp et al., 2010; Alsius et al., 2017). Its neural substrates have been examined and increased activation of the left STS has been correlated with greater perception of the McGurk effect (Nath and Beauchamp, 2012). Further, clinical populations show differences in their perception of the McGurk effect. People with amblyopia have a reduced susceptibility to the McGurk effect that persists with both binocular and fellow eye viewing (Narinesingh et al., 2014). People with one eye perceive the McGurk effect less often than binocular viewing controls (Moro and Steeves, 2018a). These results have been replicated with the behavioral data obtained during scan sessions in our current study. The present neuroimaging findings indicate people with one eye have a trend toward increased activation intensity for McGurk stimuli along the STS compared to binocular viewing controls. These results are unexpected since behaviorally this group has a much weaker McGurk effect. Since increased activation in the left STS has been shown to be associated with increased perception of the McGurk effect (Nath and Beauchamp, 2012) we expected that the decreased perception of the McGurk effect would be associated with a trend toward decreased activation. The present findings instead show an inverse relationship, as illustrated with a significant negative correlation between behavior and level of brain activity in the left audiovisual ROI. This negative correlation is driven by participants with one eye and indicates that the activation in the left audiovisual ROI for these participants was higher than that reported by the control participants. This finding suggesting perhaps that other cortical regions contribute more heavily to the perception of the McGurk effect in this patient group. It is also possible that the trend toward increased activation may be associated with reorganization of neurons typically dedicated to binocular vision or the remaining eye activating for removed eye, resulting in overall increased activation. These results should be interpreted with caution, however, since the small sample size investigated in this study is not ideal to conduct correlational analyses.

Additional studies examining individual differences that relate brain structure, function and behavioral performance, specifically in sensory deprived individuals should be investigated. As is typical in studying patients with rare diseases, our study was limited due to the rare patient group of people who had one eye surgically removed early in life due to childhood retinoblastoma. It is challenging to obtain a normalized and sufficiently large sample size to conduct inferential statistics and as such to lessen these limitations each patient was sex- and approximately age- matched with participants in both control groups.

In conclusion, the growing body of evidence demonstrates that a number of perceptual accommodations, as well as, structural and functional brain changes occur across the senses in people who have lost one eye early in life. These adaptations likely serve to mitigate the loss of binocularity during early brain development through altered sensory processing compared to binocular and patched viewing controls.

Data Availability Statement

The datasets generated for this study are available on request to the corresponding author.

Ethics Statement

The studies involving human participants were reviewed and approved by York University Research Ethics Board. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

Author contributions by SM include conceptualization, formal analysis, investigation, methodology, visualization, writing (original draft). Author contributions by DG include analysis methodology, visualization, writing. Author contributions by JS include conceptualization, methodology, funding acquisition, project administration, resources, supervision, writing (review and editing).

Funding

This research was supported by grants from the Natural Sciences and Engineering Research Council of Canada, as well as the Canada Foundation for Innovation to JS. SM was supported by the Natural Sciences and Engineering Research Council of Canada, Toronto Rehab’s TD Graduate Scholarship for People with Disabilities and York University.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We sincerely thank all of our participants for their participation in this study. Special thanks to Matthew Pachai for his support with data analysis and Ben Sachar for his support with data collection.

Footnotes

  1. ^ www.neuroelf.net, RRID:SCR_014147

References

Alsius, A., Paré, M., and Munhall, K. (2017). Forty years after hearing lips and seeing voices: the McGurk effect revisited. Multisens. Res. 31, 111–144. doi: 10.1163/2213480800002565

CrossRef Full Text | Google Scholar

Beauchamp, M. S., Nath, A. R., and Pasalar, S. (2010). fMRI-guided transcranial magnetic stimulation reveals that the superior temporal sulcus is a cortical locus of the McGurk effect. J. Neurosci. 30, 2414–2417. doi: 10.1523/JNEUROSCI.4865-09.2010

CrossRef Full Text | Google Scholar

Binda, P., Kurzawski, J. W., Lunghi, C., Biagi, L., Tosetti, M., and Morrone, M. C. (2018). Response to short-term deprivation of the human adult visual cortex measured with tT BOLD. eLife 7:e40014. doi: 10.7554/eLife.40014

PubMed Abstract | CrossRef Full Text | Google Scholar

Buchel, C., Price, C., and Friston, K. (1998). A multimodal language region in the ventral visual pathway. Nature 394, 274–277. doi: 10.1038/28389

PubMed Abstract | CrossRef Full Text | Google Scholar

Castaldi, E., Lunghi, C., and Morrone, M. C. (2020). Neuroplasticity in adult human visual cortex. Neurosci. Behav. Rev. 112, 542–552.

Google Scholar

Cohen, L. G., Celnik, P., Pascual-Leone, A., Corwell, B., Falz, L., and Dambrosia, J. (1997). Functional relevance of cross-modal plasticity in blind humans. Nature 389, 180–183. doi: 10.1038/38278

PubMed Abstract | CrossRef Full Text | Google Scholar

Collignon, O., Voss, P., Lassonde, M., and Lepore, F. (2009). Cross-modal plasticity for the spatial processing of sounds in visually deprived subjects. Exp. Brain Res. 192, 343–358. doi: 10.1007/s00221-008-1553-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Cox, R. W. (1996). AFNI: software for analysis of visualization of functional magnetic resonance neuroimages. Comput. Biomed. Res. 29, 162–173. doi: 10.1006/cbmr.1996.0014

PubMed Abstract | CrossRef Full Text | Google Scholar

Goldreich, D., and Kanics, I. M. (2003). Tactile acuity is enhanced in blindness. J. Neurosci. 23, 3439–3445.

Google Scholar

Gori, M., Sandini, G., Martinoli, C., and Burr, D. C. (2014). Impairment of auditory spatial localization in congenitally blind human subjects. Brain 137, 288–293. doi: 10.1093/brain/awt311

PubMed Abstract | CrossRef Full Text | Google Scholar

Greve, D. N. (2002). Optseq Home Page. Available at: http://surfer.nmr.mgh.harvard.edu/optseq (accessed July 9, 2014).

Google Scholar

Hoover, A. E. N., Harris, L. R., and Steeves, J. K. E. (2012). Sensory compensation in sound localization in people with one eye. Exp. Brain Res. 216, 565–574. doi: 10.1007/s00221-011-2960-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Kelly, K. R., DeSimone, K. D., Gallie, B. L., and Steeves, J. K. E. (2015). Increased cortical surface area and gyrification following long-term survival from early monocular enucleation. Neuroimage Clin. 7, 297–305. doi: 10.1016/j.nicl.2014.11.020

PubMed Abstract | CrossRef Full Text | Google Scholar

Kelly, K. R., Gallie, B. L., and Steeves, J. K. E. (2012). Impaired face processing in early monocular deprivation from enucleation. Invest. Optomet. Vis. Sci. 89, 137–147. doi: 10.1097/OPX.0b013e318240488e

PubMed Abstract | CrossRef Full Text | Google Scholar

Kelly, K. R., Gallie, B. L., and Steeves, J. K. E. (2019). Early monocular enucleation selectively disrupts neural development of face perception in the occipital face area. Exp. Eye Res. 183, 57–61. doi: 10.1016/j.exer.2018.09.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Kelly, K. R., McKetton, L., Schneider, K. A., Gallie, B. L., and Steeves, J. K. E. (2014). Altered anterior visual system development following early monocular enucleation. Neuroimage Clin. 4, 72–81. doi: 10.1016/j.nicl.2013.10.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Kelly, K. R., Moro, S. S., and Steeves, J. K. E. (2013). “Living with one eye: plasticity in visual and auditory systems,” in Plasticity in Sensory Systems, eds J. K. E. Steeves, and L. R. Harris, (Cambridge: Cambridge University Press), 94–108.

Google Scholar

Kupers, R., Pappens, M., Maertens de Noordhout, A., Schoenen, J., Ptito, M., and Fumal, A. (2007). rTMS of the occipital cortex abolishes braille reading and repetition priming in blind subjects. Neurology 68, 691–693. doi: 10.1212/01.wnl.0000255958.60530.11

PubMed Abstract | CrossRef Full Text | Google Scholar

Lessard, N., Paré, M., Lepore, F., and Lassonde, M. (1998). Early blind human subjects localize sound sources better than sighted subjects. Nature 395, 278–280. doi: 10.1038/26228

PubMed Abstract | CrossRef Full Text | Google Scholar

Lewald, J. (2002). Vertical sound localization in blind humans. Neuropsychologia 40, 1868–1872. doi: 10.1016/s0028-3932(02)00071-4

CrossRef Full Text | Google Scholar

Lunghi, C., Berchicci, M., Morrone, M. C., and Di Russo, F. (2015). Short-term monocular deprivation alters early components of visual evoked potentials. J. Physiol. 593, 4361–4372. doi: 10.1113/JP270950

PubMed Abstract | CrossRef Full Text | Google Scholar

Merabet, L. B., Battelli, L., Obretenova, S., Maguire, S., Meijer, P., and Pascual-Leone, A. (2009). Functional recruitment of visual cortex for sound encoded object identification in the blind. Neuroreport 20, 132–138. doi: 10.1097/WNR.0b013e32832104dc

PubMed Abstract | CrossRef Full Text | Google Scholar

Moro, S. S., Harris, L. R., and Steeves, J. K. E. (2014). Optimal audiovisual processing in people with one eye. Multisens. Res. 27, 173–188.

Google Scholar

Moro, S. S., Kelly, K. R., McKetton, L., Gallie, B. L., and Steeves, J. K. E. (2015). Evidence of multisensory plasticity: asymmetrical medial geniculate body in people with one eye. Neuroimage Clin. 9, 513–518. doi: 10.1016/j.nicl.2015.09.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Moro, S. S., and Steeves, J. K. E. (2012). No Colavita effect: equal auditory and visual processing in people with one eye. Exp. Brain Res. 216, 367–373.

Google Scholar

Moro, S. S., and Steeves, J. K. E. (2013). No Colavita effect: increasing temporal load maintains equal auditory and visual processing in people with one eye. Neurosci. Lett. 556, 186–190. doi: 10.1016/j.neulet.2013.09.064

PubMed Abstract | CrossRef Full Text | Google Scholar

Moro, S. S., and Steeves, J. K. E. (2018a). Audiovisual plasticity following early abnormal visual experience: reduced McGurk effect in people with one eye. Neurosci. Lett. 672, 103–107. doi: 10.1016/j.neulet.2018.02.031

PubMed Abstract | CrossRef Full Text | Google Scholar

Moro, S. S., and Steeves, J. K. E. (2018b). Intact dynamic visual capture in people with one eye. Multisens. Res. 31, 675–688. doi: 10.1163/22134808-20181311

PubMed Abstract | CrossRef Full Text | Google Scholar

Moro, S. S., and Steeves, J. K. E. (2018c). Normal temporal binding window but no sound-induced flash illusion in people with one eye. Exp. Brain Res. 236, 1825–1834. doi: 10.1007/s00221-018-5263-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Moro, S. S., and Steeves, J. K. E. (2019). Short and long-term visual deprivation leads to adapted use of audiovisual information for face-voice recognition. Vis. Res. 157, 274–281. doi: 10.1016/j.visres.2018.01.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Narinesingh, C., Wan, M., Goltz, H. C., Chandrakumar, M., and Wong, A. M. F. (2014). Audiovisual perception in adults with amblyopia: a study using the McGurk effect. Invest. Ophthalmol. Vis. Sci. 55, 3158–3164. doi: 10.1167/iovs.14-14140

PubMed Abstract | CrossRef Full Text | Google Scholar

Nath, A. R., and Beauchamp, M. S. (2012). A Neural basis for interindividual differences in the McGurk effect, a multisensory speech illusion. Neuroimage 59, 781–787. doi: 10.1016/j.neuroimage.2011.07.024

PubMed Abstract | CrossRef Full Text | Google Scholar

Quinto, L., Thompson, W. F., Russo, F. A., and Trehub, S. E. (2010). A comparison of the McGurk effect for spoken and sung syllables. Attent. Percept. Psychophys. 72, 1450–1454. doi: 10.3758/APP.72.6.1450

PubMed Abstract | CrossRef Full Text | Google Scholar

Röder, B., Stock, O., Bien, S., Neville, H., and Rösler, F. (2002). Speech processing activates visual cortex in congenitally blind humans. Eur. J. Neurosci. 16, 930–936. doi: 10.1046/j.1460-9568.2002.02147.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Röder, B., Teder-Sälejärvi, W., Sterr, A., Rösler, F., Hillyard, S. A., and Neville, H. J. (1999). Improved auditory spatial tuning in blind humans. Nature 400, 162–166. doi: 10.1038/22106

PubMed Abstract | CrossRef Full Text | Google Scholar

Sadato, N., Pascual-Leone, A., Grafman, J., Ibañez, V., Deiber, M. P., Dold, G., et al. (1996). Activation of the primary visual cortex by braille reading in blind subjects. Nature 380, 526–528. doi: 10.1038/380526a0

PubMed Abstract | CrossRef Full Text | Google Scholar

Sathian, K. (2000). Practice makes perfect: sharper tactile perception in the blind. Am. Acad. Neurol. 54:2203. doi: 10.1212/wnl.54.12.2203

PubMed Abstract | CrossRef Full Text | Google Scholar

Steeves, J. K. E., González, E. G., and Steinbach, M. J. (2008). Vision with one eye: a review of visual function following monocular enucleation. Spat. Vis. 21, 509–529.

Google Scholar

Steeves, J. K. E., Wilkinson, F., González, E. G., Wilson, H. R., and Steinbach, M. J. (2004). Global shape discrimination at reduced contrast in enucleated observers. Vis. Res. 44, 943–949. doi: 10.1016/j.visres.2003.11.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Stevenson, R. A., Zemstov, R. K., and Wallace, M. T. (2012). Individual differences in the multisensory temporal binding window predict susceptibility to audiovisual illusions. J. Exper. Psychol. 38, 1517–1529. doi: 10.1037/a0027339

PubMed Abstract | CrossRef Full Text | Google Scholar

Wanet, M. C., and Veraart, C. (1985). Processing of auditory information by the blind in spatial location tasks. Percept. Psychophys. 38, 91–96. doi: 10.3758/bf03202929

PubMed Abstract | CrossRef Full Text | Google Scholar

Weeks, R., Horwitz, B., Aziz-Sultan, A., Tian, B., Wessinger, C. M., Cohen, L. G., et al. (2000). A positron emission tomographic study of auditory localization in the congenitally blind. J. Neurosci. 20, 2664–2672. doi: 10.1523/JNEUROSCI.20-07-02664.2000

CrossRef Full Text | Google Scholar

Wong, N. A., Kelly, K. R., Gallie, B. L., Moro, S. S., Rafique, S. A., and Steeves, J. K. E. (2018). Altered white matter structure in the visual system following earlier monocular enucleation. Hum. Brain Mapp. 39, 133–144. doi: 10.1002/hbm.23831

PubMed Abstract | CrossRef Full Text | Google Scholar

Wong, N. A., Rafique, S. A., Moro, S. S., Kelly, K. R., and Steeves, J. K. E. (2019). Altered white matter structure in auditory tracts following early monocular enucleation. Neuroimage 24:102006. doi: 10.1016/j.nicl.2019.102006

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: monocular enucleation, audiovisual processing, multisensory, sensory deprivation, fMRI

Citation: Moro SS, Gorbet DJ and Steeves JKE (2020) Brain Activation for Audiovisual Information in People With One Eye Compared to Binocular and Eye-Patched Viewing Controls. Front. Neurosci. 14:529. doi: 10.3389/fnins.2020.00529

Received: 13 February 2020; Accepted: 29 April 2020;
Published: 21 May 2020.

Edited by:

Giulia Cappagli, Neurological Institute Foundation Casimiro Mondino (IRCCS), Italy

Reviewed by:

Elisa Castaldi, University of Florence, Italy
Claudia Lunghi, UMR8248 Laboratoire des Systèmes Perceptifs, France

Copyright © 2020 Moro, Gorbet and Steeves. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jennifer K. E. Steeves, c3RlZXZlc0B5b3JrdS5jYQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.