- 1Department of Otolaryngology, Sun Yat-sen Memorial Hospital, Guangzhou, China
- 2South China Normal University, Guangzhou, China
- 3School of Foreign Languages, Shenzhen University, Shenzhen, China
The mechanism underlying visual-induced auditory interaction is still under discussion. Here, we provide evidence that the mirror mechanism underlies visual–auditory interactions. In this study, visual stimuli were divided into two major groups—mirror stimuli that were able to activate mirror neurons and non-mirror stimuli that were not able to activate mirror neurons. The two groups were further divided into six subgroups as follows: visual speech-related mirror stimuli, visual speech-irrelevant mirror stimuli, and non-mirror stimuli with four different luminance levels. Participants were 25 children with cochlear implants (CIs) who underwent an event-related potential (ERP) and speech recognition task. The main results were as follows: (1) there were significant differences in P1, N1, and P2 ERPs between mirror stimuli and non-mirror stimuli; (2) these ERP differences between mirror and non-mirror stimuli were partly driven by Brodmann areas 41 and 42 in the superior temporal gyrus; (3) ERP component differences between visual speech-related mirror and non-mirror stimuli were partly driven by Brodmann area 39 (visual speech area), which was not observed when comparing the visual speech-irrelevant stimulus and non-mirror groups; and (4) ERPs evoked by visual speech-related mirror stimuli had more components correlated with speech recognition than ERPs evoked by non-mirror stimuli, while ERPs evoked by speech-irrelevant mirror stimuli were not significantly different to those induced by the non-mirror stimuli. These results indicate the following: (1) mirror and non-mirror stimuli differ in their associated neural activation; (2) the visual–auditory interaction possibly led to ERP differences, as Brodmann areas 41 and 42 constitute the primary auditory cortex; (3) mirror neurons could be responsible for the ERP differences, considering that Brodmann area 39 is associated with processing information about speech-related mirror stimuli; and (4) ERPs evoked by visual speech-related mirror stimuli could better reflect speech recognition ability. These results support the hypothesis that a mirror mechanism underlies visual–auditory interactions.
Introduction
Visual–auditory interactions represent the interference between the visual system and the auditory system (Bulkin and Groh, 2006). A classic example of a visual–auditory interaction is the McGurk effect, in which video lip movements for [ga], dubbed by syllable sound [ba], lead to the auditory illusion of the fused syllable [da] (McGurk and MacDonald, 1976). This indicates that the visual system interferes with the auditory system. Another example is that pictures containing auditory information, such as “playing the violin” or “welding,” can activate the primary auditory cortex Brodmann area 41, while other pictures without auditory information do not (Proverbio et al., 2011). Furthermore, people with hearing loss show greater activation in the auditory cortex than normal-hearing controls during these kinds of stimuli; this suggests that the visual cortex compensates for hearing loss and could explain the basis of visual–auditory interactions (Finney et al., 2001; Liang et al., 2017).
Neuroimaging evidence has suggested that cross-modal plasticity might underlie visual–auditory interactions in patients with hearing loss (Finney et al., 2001; Mao and Pallas, 2013; Stropahl and Debener, 2017). This could indicate that a loss of auditory input causes the visual cortex to “take up” the auditory cortex via cross-modal plasticity and causes an interference of the visual system on the auditory system.
However, there are still some questions under discussion. For example, the McGurk effect is seen not only in patients with hearing loss but also in normal-hearing people, whose sensory tendency (a preference for visual or auditory interference) is no different to patients with hearing loss (Rosemann et al., 2020). Considering that cross-modal plasticity differs between these two groups while they showed no difference in visual–auditory interaction, there might be other mechanisms underlying visual–auditory interactions. We hypothesized that the mirror mechanism is one such mechanism underlying this visual–auditory interaction.
First, neurons that discharge in response to both observation and execution are called mirror neurons, and the theory to explain their functions is called mirror mechanism. For example, it has been reported that a large proportion of a monkey’s premotor cortical neurons discharge not only while performing specific actions but also when hearing associated sounds or observing the same actions (Keysers et al., 2002); thus, in observational processes, mirror neurons imitate what they do in executional processes. Incidentally, visual–auditory interactions are associated with a similar phenomenon, whereby the auditory cortex imitates what it does in response to auditory stimuli during the presentation of visual stimuli.
Second, mirror mechanism allows “hearing” to be an executional process. One ongoing issue is that hearing is a feeling rather an executable action and thus cannot be classified as a “motor action,” which refers to the executional process of the mirror mechanism. However, some researchers have argued that “motor actions” encoded by the mirror mechanism are actually “action goals” rather than muscle contractions or joint displacement (Rizzolatti et al., 2001; Rizzolatti and Sinigaglia, 2016). For example, mirror mechanism has been related with empathy (Gazzola et al., 2006) and animal sounds that were mistakenly recognized as tool sounds, thus activating similar cortical areas as the tool sounds (Brefczynski et al., 2005); empathy and sound recognition are not muscle contractions or joint displacement. Currently, mirror mechanism has been defined as “a basic brain mechanism that transforms sensory representations of others’ behavior into one’s own motor or visceromotor representations concerning that behavior and, depending on the location, can fulfill a range of cognitive functions, including action and emotion understanding”; furthermore, “motor actions” usually refer to the outcomes induced by actions rather than the actions themselves (Rizzolatti and Sinigaglia, 2016). Thus, mirror mechanism might allow “hearing” to be an executional process.
Third, we have consistency of activated nerve localization in visual–auditory interaction and mirror neurons. One experiment reported that visual–auditory stimuli can activate mirror neurons (Proverbio et al., 2011), but has not been further discussed. This might be due to limitations that are inherent to the study design, whereby pictures with no auditory information are complex and may therefore cause activation of mirror neurons.
To conclude, it is reasonable to suspect that the mirror mechanism might underlie visual–auditory interactions. In this study, we measured event-related potentials (ERPs) and speech recognition ability in children with cochlear implants (CIs) during mirror neuron-activated and mirror neuron-silent visual–auditory interaction tasks. We aimed to determine the role of this mirror mechanism in visual–auditory interactions.
Materials and Methods
Participants
Twenty-five prelingually deaf children fitted with a CI for at least 4 years were recruited. These included 12 boys and 13 girls aged between 5 and 18 years (mean: 9.86; SD: 3.80). Table 1 presents the detailed demographic information. All participants were right-handed and had normal or corrected-to-normal vision. Ethical approval was obtained from the Institutional Review Board at Sun Yat-sen Memorial Hospital at Sun Yat-sen University. Detailed information was provided to the parents, and written consent was obtained before proceeding with the study.
Speech-Related Behavioral Tests
Of the 25 participants, 11 took part in the behavioral experiment. The behavioral experiment comprised six tasks as follows: easy tone, difficult tone, easy vowel, difficult vowel, easy consonant, and difficult consonant. For the experimental materials and classification standards, we referred to Research on the Spectrogram Similarity Standardization of Phonetic Stimulus for Children (Xibin, 2006). Vowels were classified according to the opening characteristics and internal structural characteristics of the first sound. As a result, vowel recognition can be divided into four groups according to these two dimensions, namely, the same structure with different openings, the same opening with different structures, the same opening with the same structure, and the compound vowels of the front nose and the rear nose. Consonants can be classified into six groups according to the two dimensions of pronunciation position and pronunciation mode, namely, fricative/non-fricative recognition, voiced consonant/clear consonant recognition, aspirated/non-aspirated consonant recognition, same position/different consonant recognition, rolled tongue/non-rolled tongue sound recognition, and same position/different consonant recognition (Note: The recognition of/z/zh/,/c/ch/, and/s/sh/fall under the recognition groups of flat tongue sound and crooked tongue sound. Given that the stimulation is very close to the pronunciation part, it is also very difficult for the normal-hearing population, so it is classified as a separate group as the most difficult content.).
Using the syllable as the task unit, recognition difficulty was defined to control for its different effects between consonants and vowels, in the tone discrimination task, and consonants within syllables of the same category, and pure in the plosives. Vowels only included the monophonies of/a/,/e/,/i/,/o/, and/u/. In the easy tone task, subjects were asked to distinguish between the first tone and the fourth tone, which are evidently different; in the difficult tone task, subjects were asked to distinguish between the third tone and the fourth tone, which are more similar to each other. There were 80 syllable pairs presented in the easy and difficult tone discrimination conditions, and the task contained 32 “same” and 48 “different” trials.
In the consonant recognition task, as mentioned above, we divided consonants into six groups according to the two dimensions. To reduce the effect of the syllable and vowel, the syllables in each group contained four tones, and the vowels in each group were monosyllabic. The difficulty of a consonant discrimination task was determined by the similarity of the two consonants. Those with a large difference in consonant type were classified as easy, while those with a small difference were classified as difficult. For example,/h/and/m/are different in pronunciation position and pronunciation mode, while/h/and/k/are only different in pronunciation mode; thus,/h/and/m/are classified as belonging to the easy recognition group, while/h/and/k/are classified as belonging to the difficult recognition group. There were a total of 100 syllable pairs in the consonant recognition task, with 40 “same” and 60 “different” trials.
In the vowel recognition task, vowels were divided into four groups according to the two dimensions. The syllables of each group contained four tones. Easy vowel recognition trials contained vowels from different vowel groups, and difficult vowel recognition trials included compound vowels with more similarity, thus making it harder to distinguish syllables. The vowel discrimination task included 60 easy trials, namely, 36 different stimulus pairs and 24 same stimulus pairs. There were 100 trials, which included 40 “same” stimulus pairs and 60 “different” stimulus pairs.
ERP Measurement
Non-mirror Stimuli
All children underwent ERP measurements. We adopted visual stimuli consisting of reversing displays of circular checkerboard patterns created by Sandmann et al. (2012), which have been used to examine cross-modal reorganization in the auditory cortex of CI users. There were four different pairs of checkerboard patterns, which systematically varied by the luminance ratio. The proportions of white pixels in the stimulus patterns were 12.5% (Level 1), 25% (Level 2), 37.5% (Level 3), and 50% (Level 4) (Supplementary Figure 1). The contrast between white and black pixels was identical in all images used.
Subjects were seated comfortably in front of a high-resolution 19-inch VGA computer monitor at a viewing distance of approximately 1 m, in a soundproof and electromagnetically shielded room. All stimuli were presented via E-prime 2.0, and stimulus software was compatible with Net Station 4 (Electrical Geodesics, Inc.). The checkerboard stimulus remained on the monitor for 500 ms and was immediately followed by blank-screen with an inter-stimulus interval of 500 ms. Each presented blank stimulus image included a fixation point (a white cross) at the center of the screen. Participants performed four experimental blocks (i.e., conditions), in which they were presented with one of the four image pairs. The block order was counterbalanced across participants. During the experimental session, each checkerboard image was repeated 60 times, resulting in a total of 480 stimuli (4 conditions × 2 images × 60 repetitions). Participants were instructed to keep their eyes on the pictures before each condition and could rest for over 1 min between blocks.
Mirror Stimuli
In this study, the mirror stimulus is a visual stimulus with certain behavioral information, thus they are able to active mirror neurons. Based on the relationship of behavioral information to speech, mirror stimulus could be further divided into speech-related mirror stimulus (i.e., a photograph of the action of reading) and speech-irrelevant stimulus (i.e., a photograph of the action of singing). Considering that current CIs cannot faithfully relay musical rhythm (Kong et al., 2004; Galvin et al., 2007; Limb and Roy, 2014), the picture of singing might just mean the movement of mouth to CI children. However, we will still use the term “singing” to describe the action in this paper. Visual stimuli were presented in a similar way to those in the study by Proverbio (Proverbio et al., 2011). These stimuli have been found to induce visual–auditory interactions in a previous study (Liang et al., 2017). Photographs that most of the children were familiar with and the content of which was understood were chosen. Supplementary Figure 2 shows the experimental block design, which consisted of an intermittent stimulus mode using speech-related mirror stimuli and speech-irrelevant mirror stimuli. To measure ERPs elicited by speech-related mirror stimuli, the experiment consisted of 85 trials of the “reading” photo stimuli and 15 trials of the “singing” photo stimuli as deviant stimuli. In contrast, to measure ERPs elicited by speech-irrelevant mirror stimuli, the experiment consisted of 85 trials of the “singing” photo stimuli and 15 trials of the “reading” photo stimuli as deviant stimuli. As shown in Supplementary Figure 2, each stimulus was presented for 1 s, followed by a blank screen (1.7–1.9 s in duration) as the inter-stimulus. To ensure that participants remained focused on the stimuli, one novel trial of 15 photographs was presented after 5–10 trials, in response to which children were asked to press a button when they saw the deviant photograph.
EEG Recording
Electroencephalography (EEG) data were continuously recorded using a 128-channel EEG electrode recording system (Electrical Geodesics, Inc.). The sampling rate for the EEG recording was 1 kHz, and electrode impedances were kept below 50 kΩ. For ERP analyses, the data of individual participants were band-pass filtered offline at 0.3–30 Hz and segmented into epochs of 100 ms pre-stimulus and 600 ms post-stimulus. Artifact rejection set at 200 mV was applied to visual EEG, and epochs were rejected if they contained any eye blinking (eye channel exceeded 140 mV) or eye movement (eye channel exceeded 55 mV). Bad channels were removed from the recording. Data were then re-referenced using a common average reference. The data were baseline-corrected to the pre-stimulus time of −100 to 0 ms.
Amplitudes and latencies of the P1–N1–P2 complex on the 75(Oz) electrode for individual participants were analyzed. The time window was set to 90–180 ms for the P1, 110–320 ms for the N1, and 220–400 ms for the P2. The amplitudes of the P1, N1, and P2 were measured as the baseline to peak value. Individual subject latencies were defined using the highest peak amplitude for each visually evoked potential. Individual waveform averages were averaged together for each group to compute a grand-average waveform.
Statistical Analysis
For statistical analysis, groups were divided by the various types of stimuli as follows: speech-related mirror stimuli, speech-irrelevant mirror stimuli, non-mirror stimuli level 1, non-mirror stimuli level 2, non-mirror stimuli level 3, and non-mirror stimuli level 4. An ANOVA was applied to examine between-group differences in ERP components. The Least Significant Difference test was then used to determine from which group–group comparison the significant difference originated.
Applying standardized low-resolution brain electromagnetic tomography (sLORETA) (Fuchs et al., 2002; Pascual-Marqui, 2002; Jurcak et al., 2007), a source location analysis was performed for the ERP components (P1, N1, and P2). The point in ERP (such as group A vs. group B) we chose to measure the difference in latency was calculated from the mean of group A and group B {[Latency (A) + Latency (B)]/2}. The top five brain regions in which there was a difference were revealed.
The assumption of a normal distribution was not satisfied, and so Spearman’s correlation coefficient was used to test the correlations between ERP components (amplitude and latency included) and speech behavior. For all tests, a p-value < 0.05 was taken to indicate significance.
Results
Event-Related Potentials
Between-Group Differences
Event-Related Potentials results are summarized in Figure 1. Significant differences were observed in the P1, N1, and P2 components between ERPs evoked by mirror stimuli and those evoked by non-mirror stimuli (Figure 1C). The data from normal-hearing controls are shown in Supplementary Figure 1. In a word, in the two major stimulus groups, ERPs elicited by the mirror stimuli differed significantly to the non-mirror stimuli. However, in subgroups, no significant difference in ERPs was found between the speech-related mirror stimuli and speech-irrelevant mirror stimuli, nor between the four types of non-mirror stimuli. Data for the normal-hearing control group are shown in Supplementary Data.
Figure 1. ERP results. (A) ERP traces for the different groups. Stimuli were delivered at 0 ms, and a P1–N1–P2 complex was observed after 100 ms. MVSR: mirror and visual speech-related, MVSI: mirror and visual speech-irrelevant, NM: no-mirror. Delta: mean value. (B) Values used to describe ERP components. (C) An ANOVA showed that the difference between groups was mainly caused by differences in ERP amplitudes. ∗p < 0.05. (D) Multiple comparisons post-ANOVA. “Boldface number” means p < 0.05, indicating that significant differences in the ANOVA were mainly caused by the difference between the mirror stimulus group (including speech-related mirror stimuli and speech-irrelevant stimuli) and the non-mirror stimulus group (including the four types of non-mirror stimuli).
Source Location of ERP Differences
An sLORETA analysis was applied to the ERP data to assess between-group differences, and the results are shown in Figure 2.
Figure 2. Source location of ERP differences. (A) sLORETA for ERP differences between groups. Colorful regions indicate the source of ERP differences. (B) Red: BA39 (left), which is the visual speech cortex. Green: BA41 and BA42, which make up the primary auditory cortex. Yellow: BA4 and BA6, which make up the precentral gyrus. Only speech-related stimuli activated BA39 differently, compared with non-mirror stimuli.
On comparing ERPs between the non-mirror and mirror stimuli groups, the primary auditory cortex in the superior temporal gyrus, including Brodmann area 41 (B41) and BA42, was strongly activated for the mirror stimuli, which indicates the presence of a visual–auditory interaction. Furthermore, the precentral gyrus, including BA4 and BA6, was more strongly activated in the mirror stimuli vs. the non-mirror stimuli groups. Moreover, compared with ERPs elicited by the non-mirror stimuli, BA39, the visual speech area in the angular gyrus was strongly activated only for the speech-related mirror group. This was not observed when comparing ERPs between the non-mirror stimuli and speech-irrelevant mirror groups.
Correlation Between ERP Component and Speech Recognition
Correlation analysis showed that the combination of stimulus ERP peak–task RT performance enriches most significant correlations (Figures 3A,B, left). Further crosstab analysis revealed that the combination of ERP peak in mirror and visual speech-related (MVSR)–task reaction time performance includes most significant correlations within the combination of ERP peak–task RT performance. This indicates that, compared with the “mirror: visual speech-irrelevant” and “no-mirror” stimuli, the “mirror: visual speech-irrelevant” stimuli induced ERPs that better reflected speech task performance.
Figure 3. Correlation between ERP component and speech recognition. (A) Heatmaps of Spearman’s correlation coefficient (–1, +1). ACC, accuracy; RT, reaction time; MVSR, mirror and visual speech-related; MVSI, mirror and visual speech-irrelevant; S, stimulus peak; L, latency; ∗p < 0.05. (B) As the heatmap is a combination of ERP components and speech measures, we tried to find which type of combination (for example: ERP latency–task RT performance vs. ERP peak–task RT performance) resulted in stronger correlations by calculating the percentage of cells with “∗” in all cells of specific combination. Left: four quadrants as four combinations; right: in II quadrant, MSR/MSI/NM peak–RT performance as three combinations. (I, II, III, IV: quadrant of the heatmap, I: latency–task RT performance, II: peak–task RT performance, III: peak–task ACC performance, IV: latency–task ACC performance. a, b, c, d, e, f: in analysis of crosstab, the same character indicates no significant difference, while a different character indicates a significant difference).
Discussion
Explanation for the Results
We hypothesized that the mirror mechanism underlies visual–auditory interactions. Concerning the visual–auditory interaction element, mirror stimuli denoting reading and singing actions were able to induce visual–auditory interactions, while the non-mirror stimuli did not. People with hearing loss tend to receive speech-related information indirectly, such as via lip reading (Geers et al., 2003; Burden and Campbell, 2011). Also, rehabilitation training in those with hearing loss encourages reading loudly; while the sounds may not be accurate, this may activate the auditory cortex. Singing is obviously sound-related, while a black–white checkerboard is not related to sound. This could explain why the two major stimulus groups (mirror vs. non-mirror groups) induced different ERPs, which was driven by the primary auditory cortex (BA41and BA42).
Concerning the element of the mirror mechanism, mirror stimuli showing reading and singing might activate mirror neurons, unlike non-mirror stimuli. Since mirror stimuli that evoke this mirror mechanism should have the same action goal while both observing and executing the process, stimuli with an imitable action are likely to activate mirror neurons; this interpretation is consistent with previous studies (Kohler et al., 2002; Vogt et al., 2004; Michael et al., 2014). In contrast, a checkerboard image is much less likely to activate mirror neurons, as the function of mirror mechanism is to benefit the individual by imitating actions, comprehending actions, and understanding feelings; a meaningless geometric figure fits none of these functions. This could explain why the two major stimulus groups evoked different neural activities, reflected in ERP differences.
Spatial source of ERP difference is located in mirror neurons. According to the mirror mechanism theory, a mirror neuron can be activated during both execution and observation of an action. In this context, both the action and perception of reading will activate the visual speech area (BA39). To be classified as a mirror neuron, the neuron needs to be activated in both the observation and execution of reading. For singing, our region of interest was the motor cortex, including BA4 and BA6.
Previous works have noted that visual–auditory interactions differ between normal-hearing people and those with hearing loss (Stropahl and Debener, 2017; Rosemann et al., 2020), and that visual–auditory interactions might be related to speech performance in children with a CI (Liang et al., 2017). This indicates that visual–auditory interactions are related to speech comprehension in people with hearing loss. Considering the ability of mirror neurons to encode motor actions (di Pellegrino et al., 1992; Gallese et al., 1996; Rizzolatti et al., 1996), the relationship between visual–auditory interactions and speech recognition ability might be related to this ability. This could explain why the speech-related stimuli evoked stronger visual–auditory interactions that are reflected by speech recognition ability.
ERP Results
First, ERP results revealed that visual–auditory interactions were induced in the mirror stimuli groups, whether the stimuli were speech-related or irrelevant, with the strongest activation in the primary auditory cortex (Figure 2). Our finding of the primary auditory cortex activation is consistent with those of a previous study, in which the primary auditory cortex (BA41) was activated by sound-related photos at around 200 ms after stimulus presentation (Proverbio et al., 2011).
Second, on comparing the ERPs elicited by the mirror stimuli and non-mirror stimuli, the results indicated that the mirror neurons were more strongly activated in the “motor action” condition than in the functional cortex. More precisely, an image of reading evoked a stronger activation in the visual speech area (left BA39; Figure 2), while the image of singing induced a stronger activation in the precentral gyrus (BA4 and BA6), where the mouth and hand areas are located (Figure 2).
Third, we found that that difference in the degree of checkerboard lightness did not significantly affect ERPs. However, a trend was found in the effect of luminance on ERPs, which is consistent with a previous study (Sandmann et al., 2012). Given that we only used four luminance (lightness-value of the ERP component), a correlation analysis was not possible. Even if differences in ERPs between mirror and non-mirror stimuli partly originated from other differences between stimuli, such as lightness or color, these differences do not convincingly explain the activation in response to a “motor action” corresponding to activity of the functional cortex.
Finally, the activated BA39, BA41, and BA42 regions are spatially close to each other and may be involved in the function of the left inferior parietal lobe in social cognition, language, and other comprehension tasks (Hauk et al., 2005; Jefferies and Lambon, 2006; Binder and Desai, 2011; Ishibashi et al., 2011; Bzdok et al., 2016). This finding supports the idea that the mirror mechanism is one of the mechanisms underlying visual–auditory interactions.
Behavioral Results and Correlation Analysis
The significant correlation found between ERP features and speech recognition ability indicates that visual speech-related mirror stimuli better reflect the ability of speech recognition than non-mirror stimuli and visual speech-irrelevant mirror stimuli. Considering that the function of the mirror mechanism is to comprehend motor actions, this indicates that the function of visual–auditory interactions to reflect speech behavior is one function of the mirror mechanism.
Above all, the behavioral results and correlation analysis indicate that the function of visual–auditory interactions to reflect speech behavior is one function of the mirror mechanism, but that the primary cortex also participates in this function of visual–auditory interaction.
Cross-Modal Plasticity and the Mirror Mechanism
Cross-modal plasticity is a double-edged sword in the auditory cortex regeneration of post-CI deaf people who regain auditory input. On the one hand, in the early period, cross-modal plasticity allows the visual cortex to take over the auditory cortex for functional compensation, which weakens auditory function (Finney et al., 2001; Hauk et al., 2005). On the other hand, cross-modal plasticity allows the occupied auditory cortex to regain its auditory function under auditory input via a CI (Anderson et al., 2017; Liang et al., 2017). Thus, it is controversial to include speech-related visual stimuli into rehabilitation training of post-CI children.
In the context of the mirror mechanism, we can also give an explanation to the dual effect to auditory function of cross-modal plasticity. In normal-hearing people, there is a balance between the executing and observing processes of a visual–auditory interaction, since they can verify whether executing (such as reading/speaking) and observing processes (such as seeing/hearing) share the same outcome of action goal (Rizzolatti et al., 2001), and auditory information benefits this process.
However, in people with hearing loss, especially those with prelingual deafness, the observing process (hearing) is absent, which causes an imbalance between executing and observing processes. This means that the patient uses the visual input (such as reading lips) as the observing process to compensate for the imbalance, which changes the outcome of the action goal into certain movements of a muscle or joint (such as lip movement). Some mirror neurons in the auditory cortex change the input of the observing process from auditory input into visual input to compensate for the absence of the observing process, which leads to a “takeover” or “inhibition” phenomenon, according to the theory of cross-modal plasticity (Finney et al., 2001; Rizzolatti et al., 2001).
Considering that the mirror mechanism is also sensitive to mastery, whereby skilled individuals have greater mirror neuron activation (Glaser et al., 2005; Calvo-Merino et al., 2006; Cross et al., 2006), the departure of auditory information and body movement (such as lip language) will be strengthened in the hearing loss period of prelingually deaf patients through repeated “training” of “comprehending” the meaning of some speech-related movements, such as lip movement (Geers et al., 2003; Burden and Campbell, 2011), for survival. Even when they regain auditory input after cochlear implantation (Giraud et al., 2001), the misleading departure is still strong. Considering the mirror mechanism sensitivity to mastery, proper rehabilitation training may change input mirror neurons to the auditory cortex from those sensitive to the observing process into an auditory input. Learning allows the auditory cortex to regain its auditory function, which leads to the phenomenon of auditory cortex recovery (weakened activation of cross-modal regions is related to improved speech performance) (Sandmann et al., 2012; Liang et al., 2017), according to cross-modal plasticity theory.
Implications for Rehabilitation Training
For post-CI children, in the observing process, body movements to produce speech that are related to sound, such as lip reading, should be limited, since they will strengthen the misleading departure of action goal as mentioned above (Cooper and Craddock, 2006; Oba et al., 2013). However, abstract visual input to produce speech-related sound, such as words and sentences, should be promoted, since associating words or text with speech for speech comprehension is an ability of normal-hearing people and mirror neurons contribute to the comprehension of motor actions.
In the executing process, children should be trained to verify whether what they say is consistent with what they hear, which calls for the cochlear to promote the quality of auditory input and the correct of speech from others.
Conclusion
Compared with non-mirror stimuli, mirror stimuli activated the primary auditory cortex, including BA41 and BA42, which prompted a visual–auditory interaction. Speech-related mirror stimuli (reading) activated BA39, the visual speech area, which implies the activation of mirror neurons. ERPs of the speech-related mirror stimuli group could best reflect the speech recognition ability of participants. Cross-modal plasticity is considered to underlie the correlation between visual–auditory interactions and speech recognition performance, and we hypothesized that the mirror mechanism is related to cross-modal plasticity and underlies visual–auditory interactions.
Data Availability Statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics Statement
The studies involving human participants were reviewed and approved by Ethics Committee of Sun Yat-sen Memorial Hospital of Sun Yat-sen University. Written informed consent to participate in this study was provided by the participants’ legal guardian/next of kin.
Author Contributions
All authors listed have made a substantial, direct and intellectual contribution to the work, and approved it for publication.
Funding
This work is supported by publication fee, personnel, and materials provided by the Youth Science Foundation Project of the National Natural Science Foundation of China (Code: 81900954, person in charge: JL). The project of the National Natural Science Foundation of China Youth Science Foundation (Code: 81800922, principal: ML) provides the support of instruments and sites.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary Material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnins.2021.692520/full#supplementary-material
References
Anderson, C. A., Wiggins, I. M., Kitterick, P. T., and Hartley, D. E. H. (2017). Adaptive benefit of cross-modal plasticity following cochlear implantation in deaf adults. Proc. Natl. Acad. Sci. U S A 114, 10256–10261. doi: 10.1073/pnas.1704785114
Binder, J. R., and Desai, R. H. (2011). The neurobiology of semantic memory. Trends Cogn. Sci. 15, 527–536. doi: 10.1016/j.tics.2011.10.001
Brefczynski, J. A., Phinney, R. E., Janik, J. J., and DeYoe, E. A. (2005). Distinct cortical pathways for processing tool versus animal sounds. J. Neurosci. 25, 5148–5158. doi: 10.1523/jneurosci.0419-05.2005
Bulkin, D. A., and Groh, J. M. (2006). Seeing sounds: visual and auditory interactions in the brain. Curr. Opin. Neurobiol. 16, 415–419. doi: 10.1016/j.conb.2006.06.008
Burden, V., and Campbell, R. (2011). The development of word-coding skills in the born deaf: An experimental study of deaf school-leavers. Br. J. Dev. Psychol. 12, 331–349. doi: 10.1111/j.2044-835x.1994.tb00638.x
Bzdok, D., Hartwigsen, G., Reid, A., Laird, A. R., Fox, P. T., and Eickhoff, S. B. (2016). Left inferior parietal lobe engagement in social cognition and language. Neurosci. Biobehav. Rev. 68, 319–334. doi: 10.1016/j.neubiorev.2016.02.024
Calvo-Merino, B., Grèzes, J., Glaser, D. E., Passingham, R. E., and Haggard, P. (2006). Seeing or doing? Influence of visual and motor familiarity in action observation. Curr. Biol. 16, 1905–1910. doi: 10.1016/j.cub.2006.07.065
Cooper, H., and Craddock, L. (2006). Cochlear Implants: A Practical Guide, 2nd Edition. Practical Aspects of Audiology. Hoboken, NJ: Wiley.
Cross, E. S., Hamilton, A. F., and Grafton, S. T. (2006). Building a motor simulation de novo: observation of dance by dancers. Neuroimage 31, 1257–1267. doi: 10.1016/j.neuroimage.2006.01.033
di Pellegrino, G., Fadiga, L., Fogassi, L., Gallese, V., and Rizzolatti, G. (1992). Understanding motor events: a neurophysiological study. Exp. Brain Res. 91, 176–180. doi: 10.1007/bf00230027
Finney, E. M., Fine, I., and Dobkins, K. R. (2001). Visual stimuli activate auditory cortex in the deaf. Nat. Neurosci. 4, 1171–1173. doi: 10.1038/nn763
Fuchs, M., Kastner, J., Wagner, M., Hawes, S., and Ebersole, J. S. (2002). A standardized boundary element method volume conductor model. Clin. Neurophysiol. 113, 702–712. doi: 10.1016/s1388-2457(02)00030-5
Gallese, V., Fadiga, L., Fogassi, L., and Rizzolatti, G. (1996). Action recognition in the premotor cortex. Brain 119, (Pt 2), 593–609. doi: 10.1093/brain/119.2.593
Galvin, J. R., Fu, Q. J., and Nogaki, G. (2007). Melodic contour identification by cochlear implant listeners. Ear. Hear 28, 302–319. doi: 10.1097/01.aud.0000261689.35445.20
Gazzola, V., Aziz-Zadeh, L., and Keysers, C. (2006). Empathy and the somatotopic auditory mirror system in humans. Curr. Biol. 16, 1824–1829. doi: 10.1016/j.cub.2006.07.072
Geers, A., Brenner, C., and Davidson, L. (2003). Factors associated with development of speech perception skills in children implanted by age five. Ear. Hear 24, (1 Suppl.), 24S–35S.
Giraud, A. L., Price, C. J., Graham, J. M., Truy, E., and Frackowiak, R. S. (2001). Cross-modal plasticity underpins language recovery after cochlear implantation. Neuron 30, 657–663. doi: 10.1016/s0896-6273(01)00318-x
Glaser, D. E., Grèzes, J., Passingham, R. E., and Haggard, P. (2005). Action observation and acquired motor skills: an FMRI study with expert dancers. Cereb Cortex 15, 1243–1249. doi: 10.1093/cercor/bhi007
Hauk, O., Nikulin, V. V., and Ilmoniemi, R. J. (2005). Functional links between motor and language systems. Eur. J. Neurosci. 21, 793–797. doi: 10.1111/j.1460-9568.2005.03900.x
Ishibashi, R., Lambon Ralph, M. A., Saito, S., and Pobric, G. (2011). Different roles of lateral anterior temporal lobe and inferior parietal lobule in coding function and manipulation tool knowledge: evidence from an rTMS study. Neuropsychologia 49, 1128–1135. doi: 10.1016/j.neuropsychologia.2011.01.004
Jefferies, E., and Lambon, R. M. (2006). Semantic impairment in stroke aphasia versus semantic dementia: a case-series comparison. Brain 129, (Pt 8), 2132–2147. doi: 10.1093/brain/awl153
Jurcak, V., Tsuzuki, D., and Dan, I. (2007). 10/20, 10/10, and 10/5 systems revisited: their validity as relative head-surface-based positioning systems. Neuroimage 34, 1600–1611. doi: 10.1016/j.neuroimage.2006.09.024
Keysers, C., Umiltà, M. A., Fogassi, L., Gallese, V., and Rizzolatti, G. (2002). Hearing sounds, understanding actions: action representation in mirror neurons. Science 297, 846–848.
Kohler, E., Keysers, C., Umiltà, M. A., Fogassi, L., Gallese, V., and Rizzolatti, G. (2002). Hearing sounds, understanding actions: action representation in mirror neurons. Science 297, 846–848. doi: 10.1126/science.1070311
Kong, Y. Y., Cruz, R., Jones, J. A., and Zeng, F. G. (2004). Music perception with temporal cues in acoustic and electric hearing. Ear Hear 25, 173–185. doi: 10.1097/01.aud.0000120365.97792.2f
Liang, M., Zhang, J., Liu, J., Chen, Y., Cai, Y., Wang, X., et al. (2017). Visually Evoked Visual-Auditory Changes Associated with Auditory Performance in Children with Cochlear Implants. Front. Hum. Neurosci. 11:510.
Limb, C. J., and Roy, A. T. (2014). Technological, biological, and acoustical constraints to music perception in cochlear implant users. Hear Res. 308, 13–26. doi: 10.1016/j.heares.2013.04.009
Mao, Y. T., and Pallas, S. L. (2013). Cross-modal plasticity results in increased inhibition in primary auditory cortical areas. Neural. Plast 2013:530651.
McGurk, H., and MacDonald, J. (1976). Hearing lips and seeing voices. Nature 264, 746–748. doi: 10.1038/264746a0
Michael, J., Sandberg, K., Skewes, J., Wolf, T., Blicher, J., Overgaard, M., et al. (2014). Continuous theta-burst stimulation demonstrates a causal role of premotor homunculus in action understanding. Psychol. Sci. 25, 963–972. doi: 10.1177/0956797613520608
Oba, S. I., Galvin, J. R., and Fu, Q. J. (2013). Minimal effects of visual memory training on auditory performance of adult cochlear implant users. J. Rehabil. Res. Dev. 50, 99–110. doi: 10.1682/jrrd.2011.12.0229
Pascual-Marqui, R. D. (2002). Standardized low-resolution brain electromagnetic tomography (sLORETA): technical details. Methods Find Exp. Clin. Pharmacol. 2002, (24 Suppl. D), 5–12.
Proverbio, A. M., D’Aniello, G. E., Adorni, R., and Zani, A. (2011). When a photograph can be heard: vision activates the auditory cortex within 110 ms. Sci. Rep. 1:54.
Rizzolatti, G., and Sinigaglia, C. (2016). The mirror mechanism: a basic principle of brain function. Nat. Rev. Neurosci. 17, 757–765. doi: 10.1038/nrn.2016.135
Rizzolatti, G., Fadiga, L., Gallese, V., and Fogassi, L. (1996). Premotor cortex and the recognition of motor actions. Brain Res. Cogn. Brain Res. 3, 131–141. doi: 10.1016/0926-6410(95)00038-0
Rizzolatti, G., Fogassi, L., and Gallese, V. (2001). Neurophysiological mechanisms underlying the understanding and imitation of action. Nat. Rev. Neurosci. 2, 661–670. doi: 10.1038/35090060
Rosemann, S., Smith, D., and Dewenter, M. (2020). Thiel, Age-related hearing loss influences functional connectivity of auditory cortex for the McGurk illusion. Cortex 129, 266–280. doi: 10.1016/j.cortex.2020.04.022
Sandmann, P., Dillier, N., Eichele, T., Meyer, M., Kegel, A., Pascual-Marqui, R. D., et al. (2012). Visual activation of auditory cortex reflects maladaptive plasticity in cochlear implant users. Brain 135, (Pt 2), 555–568. doi: 10.1093/brain/awr329
Stropahl, M., and Debener, S. (2017). Auditory cross-modal reorganization in cochlear implant users indicates audio-visual integration. Neuroimage Clin. 16, 514–523. doi: 10.1016/j.nicl.2017.09.001
Vogt, S., Ritzl, A., Fink, G. R., Zilles, K., Freund, H. J., and Rizzolatti, G. (2004). Neural circuits underlying imitation learning of hand actions: an event-related fMRI study. Neuron 42, 323–334. doi: 10.1016/s0896-6273(04)00181-3
Keywords: mirror mechanism, event-related potential, cochlear implant, hearing loss, speech performance, visual auditory
Citation: Wang J, Liu J, Lai K, Zhang Q, Zheng Y, Wang S and Liang M (2021) Mirror Mechanism Behind Visual–Auditory Interaction: Evidence From Event-Related Potentials in Children With Cochlear Implants. Front. Neurosci. 15:692520. doi: 10.3389/fnins.2021.692520
Received: 08 April 2021; Accepted: 16 July 2021;
Published: 24 August 2021.
Edited by:
Ke Liu, Capital Medical University, ChinaCopyright © 2021 Wang, Liu, Lai, Zhang, Zheng, Wang and Liang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Suiping Wang, wangsuiping@m.scnu.edu.cn; Maojin Liang, liangmj3@mail.sysu.edu.cn
†These authors have contributed equally to this work