- 1Department of Psychology and Logopedics, Faculty of Medicine, University of Helsinki, Helsinki, Finland
- 2Cognitive Brain Research Unit, Department of Psychology and Logopedics, Faculty of Medicine, University of Helsinki, Helsinki, Finland
- 3Centre of Excellence in Music, Mind, Body and Brain, Faculty of Medicine, University of Helsinki, Helsinki, Finland
- 4Department of Digital Humanities, Faculty of Arts, University of Helsinki, Helsinki, Finland
Objective: The aim of the present study was to investigate speech processing development in children with normal hearing (NH) and cochlear implants (CI) groups using a multifeature event-related potential (ERP) paradigm. Singing is associated to enhanced attention and speech perception. Therefore, its connection to ERPs was investigated in the CI group.
Methods: The paradigm included five change types in a pseudoword: two easy- (duration, gap) and three difficult-to-detect (vowel, pitch, intensity) with CIs. The positive mismatch responses (pMMR), mismatch negativity (MMN), P3a and late differentiating negativity (LDN) responses of preschoolers (below 6 years 9 months) and schoolchildren (above 6 years 9 months) with NH or CIs at two time points (T1, T2) were investigated with Linear Mixed Modeling (LMM). For the CI group, the association of singing at home and ERP development was modeled with LMM.
Results: Overall, responses elicited by the easy- and difficult to detect changes differed between the CI and NH groups. Compared to the NH group, the CI group had smaller MMNs to vowel duration changes and gaps, larger P3a responses to gaps, and larger pMMRs and smaller LDNs to vowel identity changes. Preschoolers had smaller P3a responses and larger LDNs to gaps, and larger pMMRs to vowel identity changes than schoolchildren. In addition, the pMMRs to gaps increased from T1 to T2 in preschoolers. More parental singing in the CI group was associated with increasing pMMR and less parental singing with decreasing P3a amplitudes from T1 to T2.
Conclusion: The multifeature paradigm is suitable for assessing cortical speech processing development in children. In children with CIs, cortical discrimination is often reflected in pMMR and P3a responses, and in MMN and LDN responses in children with NH. Moreover, the cortical speech discrimination of children with CIs develops late, and over time and age, their speech sound change processing changes as does the processing of children with NH. Importantly, multisensory activities such as parental singing can lead to improvement in the discrimination and attention shifting toward speech changes in children with CIs. These novel results should be taken into account in future research and rehabilitation.
Introduction
Cochlear implants (CIs) provide hearing for congenitally deaf children. The CI captures external sound with a microphone, processes and changes it to electric signal, and delivers the signal with the electrodes implanted in the inner ear to the auditory nerve of the user. For most children hearing with CIs, the ability to perceive speech is the most important change post-implantation, as it allows oral communication with others. The CIs, however, provide some speech cues well, whereas other speech cues are underrepresented by the device, when compared to the auditory system of normally hearing (NH) individuals. In general, gross temporal changes, such as changes in duration or insertion of gaps in the speech stream, are well delivered by CIs (see Limb and Roy, 2014; Torppa et al., 2014a,b). Changes in vowel duration signal in most languages prosodic word and sentence stress, while gaps are sometimes inserted between words and sentences, and thus these temporal changes can aid in speech segmentation (Vavatzanidis et al., 2015). However, in quantity languages such as Finnish, they can also indicate a change in word meaning (see Suomi et al., 2008). For example, inserting a gap before the /t/ in the Finnish word “muta” (mud) changes it to “mutta” (but) and elongating the /u/ changes it to “muuta” (other). Thus, Finnish CI users will have to use temporal information to capture both semantic and prosodic changes in the speech stream.
In contrast to the relatively well-preserved gross temporal aspects, the signal delivered by CIs lacks detail in spectral and fine temporal information compared to that of typical human hearing (Moore, 2003). This hampers the accurate perception of pitch (fundamental frequency; F0) and timbre (Chatterjee and Peng, 2008; Limb and Roy, 2014; Oxenham, 2018), and the perception of changes in the vowels and consonants in children with CIs (Geers et al., 2003; Donaldson and Kreft, 2006; for a review, see Rødvik et al., 2018). Therefore, the changes from “muta” (mud) to “muna” (an egg) or from “muuta” (other) to “maata” (to lie down) are typically less easily perceived by CI users than the semantic changes based on temporal changes. Moreover, the dynamic range (the range between the detection threshold of sound or electric current and the point where the sensation becomes uncomfortable, Moore, 2003) of CIs is limited compared to NH, harming the perception of changes in intensity (Moore, 2003; Limb and Roy, 2014). Poor pitch and intensity resolution further leads to difficulties in the perception of speech prosody, which is important to children’s speech segmentation and language learning (word and sentence stress: O’Halpin, 2010; Torppa et al., 2014a,2020; intonation: Chatterjee and Peng, 2008; Moein et al., 2017) and the perception of emotions (Luo et al., 2007; Hopyan-Misakyan et al., 2009; Chatterjee et al., 2015, 2019; Paquette et al., 2018). The perceptual deficits also lead to increases in listening effort compared to NH, often leading to frustration and poor quality of life in CI users (Dimitrijevic et al., 2019).
The evidence on where or how the cortical processing of children with CIs might differ from that of children with NH is based mainly on indirect evidence, using degraded speech with adults or children with NH. Speech of children is processed by the bilateral temporal cortices and inferior frontal gyrus (Lawrence et al., 2021). When listening of children or adults comes more effortful (as for children with CIs compared to children with NH), in adults, the brain activity seems to increase in the prefrontal cortex, and in children, the speech processing lateralizes more to the left hemisphere (Wild et al., 2012; Evans and Davis, 2015; Lawrence et al., 2021). It thus possible that when listening to speech effortfully, children with CIs utilize cognitive auditory attention and working memory mechanisms, as well as motoric representations of speech.
On the other hand, long-term, congenital sensory deficiencies have been shown to lead to cross-modal brain plasticity, that is, to the recruitment of the auditory cortex by visual or tactile (somatosensory) input, harming the development of cortical networks for speech processing and speech perception (for a review, see Campbell and Sharma, 2016; Glick and Sharma, 2017). Atypical development of the cortical speech areas and their connections, together with the unclear signal from CIs, may harm auditory attention and cognitive processing necessary for speech perception and understanding in children with CIs (for reviews, see Price, 2012; Friederici, 2017; King et al., 2021). Fortunately, children’s brains are plastic, and, if implanted earlier than at the age of 3.5 years, children with CIs acquire fairly good behavioral speech perception and language skills (van Wieringen and Wouters, 2015; Campbell and Sharma, 2016). However, the behavioral observations tell very little on whether the processing is similar to that of NH peers, or if it is affected by higher listening effort and/or compensatory processing. Therefore, research has turned to event-related potentials (ERPs) as psychophysiological markers of the underlying cognitive processes in studying typical and atypical speech perception in both children and adults.
In investigating cortical processing, ERPs, extracted from electroencephalograpy (EEG), can illuminate the reactions of the brain to changes in the auditory environment, even without the listener having to pay attention to the sound (Näätänen et al., 2017, among others). EEG is a non-invasive method, which can be used in small children, even in neonates (see, e.g., Partanen et al., 2013a). The cortical processing of an auditory change is typically reflected in a chain of positive and negative ERP responses, each deflection corresponding to a different aspect of change detection, such as detection, attention call, and further processing (for a review, see Hu et al., 2021).
Perhaps the most widely studied of these responses is the mismatch negativity (MMN), elicited in adults 100–200 ms after stimulus onset for changes in speech sound features (for a review, see Näätänen et al., 2017). In addition, increases in MMN amplitudes and decreases in their latencies are associated with better behavioral change detection abilities (Näätänen et al., 2007, 2017). In NH listeners, MMN amplitudes are maximal in the frontotemporal scalp areas, with generators located bilaterally in the supratemporal auditory cortices, and also in the right frontal cortex (Pihko et al., 2005; Kuuluvainen et al., 2014; for a review, see Näätänen et al., 2007). Compared to the adult MMNs, children’s MMNs are elicited at a wider time window, between 100 and 400 ms after a change in the sound stream, the latency and amplitude varying depending on age, magnitude of change, and stimulus type (Shafer et al., 2010; Paquette et al., 2013; for a review, see Näätänen et al., 2007).
In children, MMN can be replaced with positive mismatch response (pMMR), and sometimes both pMMR and MMN are elicited, with the pMMR preceding or following the MMN (Shafer et al., 2010; Lee et al., 2012). The pMMR was first reported in babies (Dehaene-Lambertz and Dehaene, 1994), but later studies reported pMMRs also in older children, up to the age of 12 years, for changes that are small or difficult to detect (Maurer et al., 2003; Shafer et al., 2010; Lee et al., 2012; Partanen et al., 2013b). Since pMMRs have been found to diminish with children’s age (Morr et al., 2002), and the same stimulus feature change can elicit positive pMMRs in younger NH children but negative MMN responses in older NH children (Shafer et al., 2010), the pMMR may also reflect neural immaturity. As further evidence of this, in NH children aged 4–12 years, individual level incidence of pMMR for vowel changes was associated with poorer performance intelligence quotient (PIQ) scores 14–17 months after the recording (Partanen et al., 2013b).
The MMN has been elicited in both children and adults with CIs (for a review, see Näätänen et al., 2017), and with amplitudes becoming larger with better behavioral perception (for pitch: Cai et al., 2020; for speech: Turgeon et al., 2014; for a review, see Näätänen et al., 2017). Furthermore, the pMMR has also been elicited in children with CIs (for a review, see Ziatabar Ahmadi et al., 2022). For instance, using a multifeatured paradigm (see Näätänen et al., 2017) with harmonic or musical tones, pMMRs were observed in the time window of the MMN (Engström et al., 2020; for “CI singers” in Torppa et al., 2014b,2018). Notably, pMMRs in these studies were associated with good rather than with poor perceptual abilities. Further, positive, pMMR-like deflections before MMN can be seen (but were not analyzed) for changes in speech in 7–19 year-old children with CIs who perform well in speech discrimination tasks (Figure 2, CFz electrode, approximately 150 ms, in Ortmann et al., 2013a).
In NH adults and children, the MMN is often followed by the P3a response. The P3a is elicited if the change is clearly detected (see Torppa et al., 2014b; for a review, see Wetzel and Schröger, 2014) and, in the NH population, P3a reflects an attention switch toward a sound change (Friedman et al., 2001; Horváth et al., 2008; Wetzel and Schröger, 2014). Similarly to the MMN, P3a amplitude increases together with increasing difference between the deviant and standard, suggesting that the more salient the change is for the perceiver, the more likely it is to result in an involuntary attention shift (Winkler et al., 1998; Wetzel et al., 2006; Wetzel and Schröger, 2014). In NH, the neural network of the P3a is distributed across frontal, parietal and temporal cortical regions (Takahashi et al., 2013). However, while the auditory MMN generators are typically stronger in the auditory temporal areas, the emphasis in P3a elicitation is in frontal brain areas devoted to attention-related processing (Takahashi et al., 2013). Thus, the P3a is also usually easily detectable from the frontocentral electrode locations, if elicited.
Similarly to pMMR and MMN, the P3a has also been elicited in children with CIs (for a review, see Näätänen et al., 2017). To the best of our knowledge, there is only one study directly assessing P3a to changes in a speech stimulus in children with CIs. Kileny et al. (1997) found that the P3a latency was later but its amplitude larger for speech contrasts (heed vs. who’d) than for frequency changes (1500 vs. 3000 Hz tone bursts) in children with CIs aged between 4 and 12 years. Furthermore, the larger and earlier the P3a for the frequency changes was, the better was the speech recognition (Kileny et al., 1997). The results suggest that in children with CIs, automatic attention shifts to auditory stimuli are important for their speech perception abilities.
The late differentiating negativity (LDN), sometimes also named late MMN, appears often after the pMMR or MMN and P3a responses in NH children (Korpilahti et al., 2001; Bishop et al., 2011; Liu et al., 2014; Kuuluvainen et al., 2016a). The LDN amplitude becomes smaller when the physical change in the stimulus becomes larger, and the LDN appears relatively often in children but is usually absent in adults (Bishop et al., 2011; Kuuluvainen et al., 2014, 2016a; Liu et al., 2014). The functional role of LDN is unclear. It may reflect cognitive processing or additional cognitive processing (“a second look”) of auditory stimuli that are complex or hard to discriminate (Shafer et al., 2005; Bishop et al., 2011), or the ongoing establishment of internal phonological representations in children (Liu et al., 2014). Kuuluvainen et al. (2016a) investigated the relationship of the LDN to several cognitive tasks in 6-year-old children. In this study, better verbal and non-verbal reasoning were associated with declining amplitudes of the LDN to vowel-like changes in complex non-speech sounds, but not to vowel changes in syllables. The results suggest that the appearance of the LDN to irrelevant sound changes indexes poorer cognitive maturity. Thus, it is, to date, unclear whether the elicitation or size of the LDN is a marker of more mature or more immature processing, as it might depend on the stimulus and the change type.
According to the best of our knowledge, the LDN responses for changes in speech have only rarely been found in children with CIs. Singh (2005) studied in his doctoral dissertation 7–17 year-old children with CIs using computer-generated /ba/ and /da/ stimuli. The LDN was found in children with CIs, but only in those children whose auditory performance was poor. Singh also found that the longer was the duration of LDN response, the poorer was the auditory performance and speech intelligibility (Singh, 2005). Hu et al. (2021) studied in Mandarin-Chinese speaking children with NH and CIs, aged 7–13 years, the LDN responses to changes in lexical tones in monosyllable or four-syllable idiom (phrase) conditions. While the MMN was elicited in both child groups in both conditions (being smaller in the idiom context in children with CIs), the LDN was found only in the children with NH (Hu et al., 2021; see also similar results by Uhlén et al., 2017, for harmonical sine tones).
There are only few studies directly addressing the development of cortical processing of speech reflected in the chain of pMMR and/or MMN, and P3a responses in children with CIs, with literature on LDN development lacking completely. Regarding early development, Vavatzanidis et al. (2015) found that already 4 months after the first auditory input (age at CI activation from 11 to 45 months), the MMN amplitudes to changes in vowel duration, representing syllabic stress, became similar in CI children compared to those of the NH group. Vavatzanidis et al. (2016) also followed the development of the MMN of children with CIs for changes in syllable stress patterns, again cued with vowel duration, in trochaic (similar to the native language of the participants) and iambic (not typical for the native language) stimuli. In this study, the age at first activation of the CI was 9–50 months. They found that the ERP waveforms of the children with CIs were very similar to those of their NH peers. For both CI and NH children, a MMN-like mismatch response was present for the “non-native” iambic deviant, but not for the trochaic stimulus, suggesting that familiarity and multisensory learning of native language affects the development of MMN for speech stimuli similarly for CI and NH children (Vavatzanidis et al., 2016).
In addition to the signal for CIs and age at implantation and onset of hearing with CIs (explained above), there are several factors that have been extensively studied and known to improve the speech perception of children using unilateral or bilateral CIs. These factors include early age at diagnosis of the hearing impairment, extensive oral communication, the ability of the electrodes to activate the auditory nerve, and supportive involvement of the parents (for a review, see van Wieringen and Wouters, 2015). One of the most understudied factors that could improve speech perception of children with CIs is singing. The children in the current study have participated in previous studies in which the relationship of the children’s own and their parents’ singing to the children’s perceptual processing was investigated. Those children with CIs who sang regularly (“CI singers”), and whose parents sang regularly for them, were better in perception of speech in noise than other children with CIs (”CI non-singers”; Torppa et al., 2018). Moreover, the more the children with CIs participated in musical activities including singing, the better they perceived word and sentence stress and pitch (F0) changes in synthesized speech syllables (Torppa et al., 2014a). Furthermore, the more parents sang for them, the better they were in word finding and verbal IQ (Torppa et al., 2020). There are several possible reasons for the improvement of speech perception and language skills with singing. For instance, the slower tempo of lyrics in songs compared to the tempo of speech makes it easy to improve speech perception through listening to singing. Moreover, singing by children themselves is a repetitive, multisensory activity, both aspects being important for perceptual learning and the development of neural networks related to speech and language processing (for reviews, see Torppa et al., 2018; King et al., 2021). Above all, singing is known to arise the attention of young children (Corbeil et al., 2013) and keeping it for longer time than speech (Politimou et al., 2019), which is probably highly beneficial for speech processing. Since also children with CIs pay special attention to singing, it is used in their speech and language therapy (Ronkainen, 2011; McConkey Robbins, 2020). Attention toward sounds increases activation in auditory cortex (Fritz et al., 2007; Woods and Alain, 2009; Woods et al., 2009), and it has been emphasized that speech has to be brought directly to the attention of children with hearing impairments, to make them aware of sounds and speech (for instance Cole and Flexer, 2019, p. 189). Thus, there are several reasons why parental singing could improve attention and speech perception of children with CIs.
There is also previous evidence that in the participants of the present study, singing is connected to enhanced development of cortical processing of changes in a musical stimulus in a multifeature paradigm. In the paradigm, two changes which are, based on functioning of CIs (see above), easy to detect with CIs (duration changes and gap insertions), and three which are (also based on the functioning of the CIs) difficult to detect with CIs (changes in pitch, musical instrument timbre, and intensity) were presented alternating with a piano tone standard stimulus (Torppa et al., 2014b). The stimulus rate was fast, mimicking the tempo in real music. During the follow-up of 14–17 months, the ERP-responses of “CI singers,” who sang regularly and whose parents sang for them extensively, and whose speech perception in noise was good (Torppa et al., 2018), had enhanced or rapidly developing P3a or P3a-like responses over all change types. In contrast, in “CI non-singers,” the P3a to changes in timbre became smaller and later over time (Torppa et al., 2014b). At the second time point of measurements, the P3a to two difficult-to-detect change types (F0 and timbre) was larger and earlier in the “CI singers” than in the “CI non-singers” (Torppa et al., 2014b,2018). In summary, in the previous studies with these same children, the amount of singing was associated with enhanced speech-in-noise perception, verbal skills, as well as increases in the cortical indices of automatic attention shifts (reflected in P3a responses) to those changes in musical stimuli which are difficult to detect with CIs. Therefore, investigating the possible impact of singing at home to the development of cortical processing of different changes in speech is the logical next step in unraveling the importance of home activities in the development of CI children’s auditory skills.
The present study aims to investigate the cortical processing and its development of speech changes in Finnish-speaking children with CIs and NH (from here on: CI and NH groups) who have participated in the studies of Torppa et al., 2014a,b, 2018, 2020. More specifically, two changes that are easy-to detect (gap insertions and duration changes) and three that are difficult to detect with CIs (vowel identity, pitch = F0, intensity) will be presented in the middle syllable of the pseudoword /tatata/, using a fast-rate multi-feature MMN paradigm (Partanen et al., 2011, 2013b), in which the standard and the deviants alternate in a pseudorandom fashion. The paradigm mimics the challenges posed by everyday speech perception and is also comparable to the previous paradigm with musical stimuli (Torppa et al., 2014b,2018). To get a more precise picture of differences across age groups, the children are further divided into preschoolers (4–6 year-olds) and schoolchildren (7–13-year-olds), as children go to school and learn to read at the age of seven in Finland. Acquiring reading skills is expected to affect the processing of speech changes in the present participants, as reading instruction is mainly focused on letter-to-sound correspondence due to the consistent orthography in Finnish (see, e.g., Korkeamäki and Dreher, 1993), leading to rapid acquisition of basic reading skills (see, e.g., Aro and Wimmer, 2003) and to increasing phonemic awareness (see, e.g., Lerkkanen et al., 2004). Thus, learning to read and write provides the children an opportunity to visually map the sounds and semantically relevant sound lengths of Finnish, and practice them intensively both by listening and producing speech. The effect of letter-sound mapping on cortical responses has been shown in the study of Lovio et al. (2012) in which active practice of letter-sound mapping with GraphoGame resulted in enhanced MMN responses. Naturally, overall cognitive development is also another possible factor affecting response sizes and polarities in preschoolers and schoolchildren, as discussed earlier in reference to the different ERPs. The decision to divide the age groups from age 6 years 9 months is, however, based on the likelihood of reading instruction enhancing cortical change detection in the children. In the current study, a longitudinal design is used, repeating the measurements after a 14–17 months follow-up period (similarly to Torppa et al., 2014a,b, 2018), and allowing the investigation of the development of cortical speech processing over time. Finally, to study the connections of ERP responses and their development to singing, the present study also investigates the links of ERP responses to children’s own and their parent’s singing for their children in the CI group.
Cortical processing is assessed with pMMR, MMN, P3a and LDN responses. Based on earlier findings, we expect that
(1) Overall, the children with CIs will show less mature cortical responses than their NH peers, and this will be pronounced for the difficult-to-detect changes compared to the easy-to-detect changes. Therefore, we expect that
(1a) For all change types, there will be more significant pMMRs (at group level) elicited at T1 and in the preschoolers than in the schoolchildren, and in the CI than the NH group, as it has been observed to diminish with maturation (Morr et al., 2002) and because the development of auditory cortex of the CI group is assumed to be delayed in maturation compared to the NH children due to the period of deafness before implantation (Campbell and Sharma, 2016);
(1b) Significant MMNs at group level will be elicited for all change types in both CI and NH groups at both time points (Vavatzanidis et al., 2015, 2016). However, they may be smaller in the CI group than in the NH group due to difficulties in auditory discrimination, and particularly for the difficult-to detect changes. In addition, MMN amplitudes will be smaller in preschoolers than in schoolchildren, and at T1 than at T2, for those changes that are coded in the Finnish writing system (vowels, vowel duration, and gap insertion), and for which discrimination is actively practiced when learning to read and write (Korkeamäki and Dreher, 1993; Suomi et al., 2008; Hämäläinen et al., 2015);
(1c) Significant P3a responses at group level are expected to be elicited for the easy-to-detect changes in both CI and NH groups at both time points. P3a may be smaller in the CI group than in the NH group due to difficulties in the discrimination of changes and attention shifting (Kileny et al., 1997; Limb and Roy, 2014);
(1d) Significant LDNs at group level will be elicited in the NH group, and especially for those changes that change word meaning in Finnish (vowel identity and duration changes, and gap insertions; Bishop et al., 2011; Kuuluvainen et al., 2016a). The LDNs will not be elicited in the CI group, as observed by Hu et al. (2021);
(2) In the CI group, singing at home (by the children themselves and/or by their parents) will enhance the children’s speech perception and change detection abilities (Torppa et al., 2014a,b, 2018), and will thus be reflected in larger pMMR and/or MMN responses, and in improved attention shift reflected by the enhancement of P3a (Torppa et al., 2014b).
Materials and methods
Participants
The participants were 21 congenitally deaf, unilaterally implanted (CI; 9 boys) and 22 NH (11 boys) children, aged 4–13 years at the time of the first recording (Table 1). All children were native, monolingual Finnish-speakers and attended mainstream daycare or school. The children with CIs included in the study (the CI group) had their CI switch-on prior to age three years one month, had full insertion of the electrode, and more than six CI channels in use. Note that at the time the data was collected, it was assumed that the optimal time for cochlear implantation would be before the age of 3.5–4.0 years (see e.g., Kral and Sharma, 2012) and thus implantation after the age of 3 years 6 months was set as an exclusion criterion. The children with CIs had no diagnosed additional developmental or linguistic problems, and no re-implantation between the two measurements. Before the first measurement (T1), all children with CIs had been using their implants continuously for at least 30 months, and before the second (T2), at least for 46 months. Seventeen children with CIs used Cochlear, and four used Med-EL devices. Four CI children used an acoustic hearing aid (HA) in the non-implanted ear. They listened the stimulus without the HA. Their thresholds for hearing in the non-implanted ear exceeded 50 dB at 250 Hz, 60 dB at 500 Hz, and 70 dB at 1000 Hz, and evidently none of them heard the stimuli with the non-implanted ear during the experiment (see Torppa et al., 2012). Moreover, based on auditory brainstem responses (ABR), the other participants did not have usable residual hearing in the non-implanted ear.
The 22 NH children (the NH group) were siblings of the participating children with CIs or were recruited from local musical play schools, other ongoing studies at the University of Helsinki, or from the neighborhood of one of the authors. The CI and NH groups were matched as accurately as possible by age, gender, handedness and social and musical background. None of the NH children had any diagnosed developmental or linguistic problems, and their hearing was normal, as assessed in regular check-ups at child welfare clinics. Parents of the participants gave a written informed consent and the children gave their consent verbally.
The study was carried out in accordance with the Declaration of Helsinki and all procedures were approved by the ethical committees of the participating hospitals. All 43 children participated also in the Torppa et al., 2012, 2014a,b,2018, 2020 studies.
The children were divided into separate groups according to age. The age groups consisted of 11 CI and 11 NH preschoolers (under the age of 6 years 9 months) and 10 CI and 11 NH schoolchildren (over the age of 6 years 9 months; see Table 1). The CI vs. NH age groups (CI vs. NH preschoolers and CI vs. NH schoolchildren) did not differ statistically for age at T1 [t(20/19) = 0.068/–0.38, p > 0.05] or T2 [t(20/19 = 0.43/0.078, p > 0.05], time between measurements [t(20/19) = 0.093/–0.33, p > 0.05], gender [χ2(1) = 0.18/1.53, p > 0.05], handedness [χ2(1) = 0.000/2.43, p > 0.05], or mother’s education [χ2(1) = 1.49/0.40, p > 0.05]. At T1 and T2, the forward digit span task (Illinois test of Psycholinguistic Abilities; Kirk et al., 1974) was administered to assess the children’s verbal short term memory, and at T2, their non-verbal visuoconstructive performance was assessed with the block design task (Wechsler, 2010). For the CI vs. NH preschoolers, no group differences were observed for the digit span task at T1 [t(20) = –1.87, p > 0.05] or for the block design [t(20) = –1.19, p > 0.05] task, but by T2, the CI preschoolers performed worse than NH preschoolers in the digit span task [t(20) = –2.46, p = 0.023; See Table 1]. The CI schoolchildren performed worse than the NH schoolchildren in all tasks: at the digit span task at both T1 [t(19) = –2.55, p = 0.020] and T2 [t(19) = –2.23, p = 0.033], as well as in the block design [t(19) = –2.25, p = –036] task administered at T2. These age group differences in non-verbal intelligence (PIQ) and verbal short-term memory have been observed in the earlier studies of children with CIs (PIQ: Cejas et al., 2018; verbal short-term memory: Pisoni et al., 2011; Talli et al., 2018). It is also noteworthy that despite the group difference in the scores for block design, they were within the age-typical window of 8–12 standard points in all studied groups (Table 1), as expected by the previous studies on non-verbal intelligence of children with CIs (Cejas et al., 2018).
Comparing the CI preschoolers vs. CI schoolchildren, the CI preschoolers had had their implants switched on earlier than the CI schoolchildren by an average of 6.6 months [t(19) = 3.10, p = 0.006]. This is unsurprising regarding the aim to decrease the age of implantation in the generation of the present participants. However, there were no statistically significant differences for etiology [χ2(1) = 1.29, p > 0.05], processor type [χ2(3) = 5.89, p > 0.05], or pure tone hearing thresholds at 4,000 [t(19 = –0.58, p > 0.05], mean of 500, 1,000, and 1,000 [t(19) = –0.077, p < 0.05], 250 [t(18) = –0.78, p < 0.05], or 125 [t(13) = 1.37, p < 0.05] Hz dB HL, the last two measurements missing from 1 and 6 children, respectively.
The effect of singing on the ERPs was tested for the CI group. Since it was expected that both the singing by the children themselves and parental singing could play a role for brain responses and speech processing (Torppa et al., 2014a,b, 2018, 2020), we used here the data for both aspects of singing at home. Children’s own and their parents’ singing was investigated with questionnaires addressed to parents at T1 and T2. Parents indicated how often they sang with their child, and how often the child sang themselves before T1 and between T1 and T2 using Likert scale 0–5 (see Torppa et al., 2020). The data for both time points was connected by calculating the mean of the responses. To ensure that the parents were able to identify singing of the children with CIs as different from speech, we recorded the children’s singing of “Twinkle twinkle little star” at T2, and the recordings were assessed blindly by a teacher of singing. In conclusion, the children’s singing was recognizable and different from general speech (see Torppa et al., 2014a for further information). Although singing was investigated in CI children only, we would like to note that the parental singing scores for the CI vs. NH children did not differ between the groups [t(38) = –0.82, p > 0.05] (note that there was no information on parental singing for three NH children). There was no information on the NH children’s own singing, and hence that could not be compared between the groups.
Stimuli and procedure
The naturally spoken trisyllabic pseudoword/tatata/ and its variants served as stimuli (see Table 2 and Supplementary material 1). The duration of the standard stimulus was 480 ms and it included two silent gaps of 60 ms between the approximately 120 ms long syllables (for details, see Partanen et al., 2011). In the deviants, the middle syllable varied with either vowel duration, F0, gap insertion, intensity, or vowel identity. The F0, intensity, and gap deviants were made by manipulating the standard stimulus with Praat (Boersma and Weenink, 2001; see also Partanen et al., 2011), and appeared in the beginning of the second syllable (198 ms after the stimulus onset). For the vowel duration deviant, the difference from the standard became apparent approximately 280 ms from stimulus onset. The F0, intensity and gap insertion deviants were derived from the standard stimulus by manipulating it in Praat (Boersma and Weenink, 2001). However, the vowel identity and duration deviants were produced naturally during the recording of the stimuli, and thus their first syllable differed slightly from that of the standard (see Supplementary material 1 for waveforms and spectrograms of the stimuli). Using natural stimuli preserved full information available in natural speech and allowed for comparison of speech processing with CIs vs. in NH in an ecologically valid setting. The F0 deviants had two magnitudes (increases of 15 and 50% in F0), and the intensity deviants were either 6 dB increments or decrements. By dividing these deviant types to two different difficulty levels, we wished to see if processing of these would be different for participating child groups. The stimulus onset asynchrony (SOA) was 900 ms.
The stimuli were presented through two high-quality loudspeakers situated in an acoustically and electronically shielded room, in approximately 45° angles and 1 m distance from the subject’s ears on both sides of the subject. Sounds were presented at most comfortable level of 60 dB SPL for the NH and 70 dB SPL (excluding intensity increment deviants) for the CI group as measured from the ear cantus (for one child with a CI the level had to be lowered to 65 dB SPL because 70 dB SPL was uncomfortable for her). The sound intensities were higher for the CI group to accommodate for their higher thresholds for hearing in free field using their CIs (see Table 1) compared to those of normally hearing children (Rahko and Karma, 1989; Haapaniemi, 1996). During the 30-min EEG experiment, the children watched a silent movie and listened passively to the stimuli. In the one-block stimulus sequence standard and deviant syllables alternated so that the deviants occurred pseudorandomly in the sequences: repetitions of the same deviant type in excess of two were swapped with a deviant in random location of the sequence, after which the sequence was checked again for repetitions until no more such repetitions were present. In total, 2,000 stimuli were presented, with 1000 standard and 5 × 200 equiprobable different deviant stimuli, the F0 and intensity deviants having 100 of each type of change magnitude. The stimuli were presented twice (at T1 and T2) for the participants. The time between T1 and T2 was 14–17 months.
Electroencephalography recording and data analysis
Electroencephalography recordings were conducted using Biosemi ActiveTwo amplifier and 64 active electrodes embedded in a cap (Biosemi B.V., Netherlands), with a sampling rate of 512 Hz and a recording band-pass filtering of DC-102.4 Hz. The CMS/DRL electrodes were used as an online reference. Additional electrodes were placed at the left and right mastoid, the nose, and at the canthi to record eye movements and blinks.
The EEG data were first analyzed using EEGLAB 8 (Delorme and Makeig, 2004). The data were downsampled at 256 Hz, and highpass filtered above 0.5 Hz, and re-referenced to the nose. Because of the location of the CI device, some channels could not be used, and data from these electrodes were interpolated. To remove ocular and muscle artifacts in both CI and NH groups, an independent component analysis (ICA) with the Fastica algorithm was applied (Makeig et al., 2004). In addition, ICA was used in the CI group to reduce the CI related artifact (see Näätänen et al., 2017; and Torppa et al., 2012 for the details of the procedure). Before ICA, data dimensionality was narrowed down by the amount of interpolated channels, and automatic epoch rejection at a threshold between ±300 and ±400 μV was performed. The rejection thresholds were individually adjusted to preserve at least 85% of original epochs for effective statistical analysis.
After ICA, the epoch voltage rejection was done again with a threshold of ±150 μV, using an epoch of 800 ms, starting 100 ms before stimulus presentation. Further, the proportion of remaining epochs after voltage rejection was analyzed for each individual subject. The minimum was set at 73% (73) of remaining epochs for each deviant, in order to preserve as many participants as possible, as this was the number of remaining epochs of one NH child for the intensity increment deviant, with his/her data otherwise being of good quality (78–96% of epochs preserved). On average, 182 (91%) epochs were preserved in the CI group and 188 (94%) in the NH group per condition. To increase the signal to noise ratio, a region of interest (ROI) including F3, Fz, F4, C3, Cz, and C4 electrode locations were averaged for the final ERPs (see Näätänen et al., 2017 for further information). Difference waveforms were calculated by subtracting standard ERP waveform from that of each deviant. This procedure was used in all ERP analyses.
Based on previous findings using the same paradigm (Partanen et al., 2013b) and visual inspection of the ROI ERP difference (deviant-standard) waveforms, peak latencies were determined for each group and time point as follows: First, the MMN was identified as the largest negative deflection at 300–450 ms from stimulus onset. As the second syllable, in which the changes occurred, began at 198 ms from stimulus onset (see Table 2), the MMN was assessed to be elicited earliest by 300 ms from stimulus onset (or 102 ms from the start of the second syllable). Then, the pMMR was identified as a positivity preceding the MMN, at a time window starting from 250 ms and ending at the identified MMN peak (i.e., the pMMR, if present, had to precede the MMN). Two hundred and fifty milliseconds from stimulus onset was kept as the absolute lower bound for pMMR elicitation, since previous studies have shown that deviance detection in adults can occur at the earliest at 30–40 ms after stimulus onset, around the adult P1 response (for reviews, see Escera and Malmierca, 2014; Shtyrov and Lenzen, 2017). Children’s P1 is elicited typically around 80–100 ms (see, e.g., Sharma et al., 2015; Kuuluvainen et al., 2016b), so the pMMR lower bound was set to 52 ms from second syllable onset. The P3a was identified as a positive deflection after the MMN, at a time window starting from the MMN peak and ending at 525 ms, and finally, the LDN was identified from a time window starting from the P3a peak (or MMN peak if the P3a was not observable) and ending at 650 ms. Thus, for each group, timepoint and deviant, the four responses, if elicited, were in the temporal order of pMMR-MMN-P3a-LDN. Out of these 224 responses (four groups, measured twice for 7 deviants, four responses per deviant) 21 could not be identified reliably. They were instead quantified either based on their peak latency at an electrode where the response was clearly peaking, or in three cases of pMMR, the lower bound of the time window at 250 ms was chosen for significance testing, as there was no visible peak within the time window in any electrode (see Table 3). The LDN had to be elicited by 650 ms from stimulus onset (or 452 ms post-change) for it to be included in the analyses.
The data were quantified using EEGLAB 13.6.5b (Delorme and Makeig, 2004) from the ROI ERPs with the baseline set to zero during the 50 ms window before stimulus presentation onset, with an offline low-pass filter of 20 Hz. We used the median method as in Torppa et al. (2014b) in order to avoid including extreme values from possible overlapping noise due to the implants. In this method, the trials of each individual are grouped by stimulus type, and the median value of the signal amplitude values of one sample point is taken as representative of that sample point. Thus, the resulting curve from an individual consists of the samples having the median amplitude over the accepted trials.
Event-related potential amplitudes were quantified from difference waveforms of each deviant and the standard, using a 50 ms time window centered at the peak latency of each response in each of the four age groups, and for CI and NH groups separately (see Table 3).
Statistical analyses
The significances of all ERP response amplitudes (hypothesis 1a) were tested with one sample t-tests comparing to zero. Responses in the CI and NH preschoolers and CI and NH schoolchildren were included in group comparisons if the response was significant at least one of the time points T1/T2 for at least one CI and at least one NH age group, at a p < 0.05 significance level (see Table 4). For the singing analysis, significances of the responses were tested in the entire CI group, combining the preschoolers and schoolchildren in this analysis (see Supplementary material 2). However, the ERP values based on age group quantification were used for significance testing, as the response latencies observed in the age groups were more likely to catch the individual responses better than the response latencies in the combined group. The association of the response with singing was analyzed if the response was significant (p < 0.05) in at least one of the time points T1 or T2.
Linear mixed modeling (Singer and Willett, 2003; West, 2009) was used in statistical testing of (i) group differences of CI vs. NH in the age groups of preschoolers and schoolchildren (hypotheses 1b-d), and (ii) for the analysis of the effect of singing in the combined CI group (hypothesis 2; see also Torppa et al., 2014b,2018). The method was selected because it allows one to combine measurements from the several groups and two time points in a single analysis, and since it allows for missing data (Singer and Willett, 2003; West, 2009). The LMM was first estimated for fixed main effects of clinical group (CI vs. NH), age group (preschoolers vs. schoolchildren) and development over time (the difference between responses at T1 and T2) and all two- and three-way interactions, with participant as the random effect. In the second phase, in all LMM analyses, statistically not significant three-way interactions, and in the third phase, statistically not significant two-way interactions were removed, yielding a final model including the significant fixed effects and interactions, with participant as the random effect. The models were estimated with the aforementioned procedure separately for each response fulfilling the response significance inclusion criteria (see Table 4). Statistically significant interactions were further investigated with post hoc tests, using the Bonferroni correction for multiple comparisons.
Before testing hypothesis 2, the correlations of CI group’s singing scores and child age at T1 were analyzed, in order to determine if age should be included as an additional covariate. This was done as it is feasible to assume that preschoolers might sing more than schoolchildren, and parents might also sing more to preschoolers than schoolchildren. Correlations between the two singing scores, and between each score and age at T1 were not significant. Therefore, to test hypothesis 2, LMMs were estimated separately for both singing scores (parental and by the children themselves), and only the main effects of singing scores and time were included in the model as fixed effects and participant as the random effect. Including the singing scores in both fixed main effects and as a covariate allowed us to investigate the overall impact of singing on response amplitudes, as well as the interaction of measurement time and the singing scores. As in the previous LMM (to test hypotheses 1b–d), statistically not significant two-way interactions were removed from final models. Only the significant (p < 0.05) results related to the hypotheses are reported. All statistical analyses were made with SPSS 25 software (IBM Corp., Armonk, NY, USA).
Results
pMMR, MMN, P3a and LDN to (with CIs) difficult-to-detect changes in vowel duration and gap insertions between syllables
pMMR
At T1, CI preschoolers had significant pMMRs to vowel duration changes, and CI schoolchildren to gap insertions. The responses were no longer significant at T2 (see Table 4 and Figure 1). No statistical group comparisons could be conducted, as the NH group did not have significant pMMRs to these deviants. Instead, the NH children had significant early negative responses at T1 in the pMMR time window: the preschoolers to gap insertions, and the schoolchildren to duration changes. By T2 these responses were no longer significant (see Table 4 and Figure 1).
Figure 1. Difference waveforms (ERPs of deviant minus standard) from the six-electrode ROI (F3, Fz, F4, C3, Cz, C4) for the easy-to-detect change types of gap insertion and vowel duration at measurement time points T1 and T2 for the preschoolers (age below 6 years 9 months) and schoolchildren (age above 6 years and 9 months) with CI and NH. The data is referenced to the nose. Significant responses are marked with an asterisk. ERP, event related potentials; CI, children with cochlear implants; NH, normally hearing children; PR, preschoolers; SC, schoolchildren; pMMR, positive mismatch response; MMN, mismatch negativity; P3a, positive ERP following MMN; LDN, late differentiating negativity.
Figure 2. Barcharts of significant main effects and interactions in Linear Mixed Modeling for the easy-to-detect change types of gap insertion and vowel duration. (A) CI vs. NH group main effects; (B) age (preschoolers vs. schoolchildren) main effects; MMN, mismatch negativity; CI, cochlear implant; NH, normal hearing. *p < 0.05, **p < 0.01, ***p < 0.001. Note the different scales for different responses.
MMN
At the MMN time window the response polarities between groups were negative for all groups at both time points (Table 4 and Figure 1), and also statistically significant except for the MMN to vowel duration changes in both CI age groups at T1. Thus, the MMN responses for changes in vowel duration and gap insertion were both included in the statistical analyses of MMN (Table 4). Both MMNs were smaller for the CI group than for the NH group [gap: F(1,40) = 17.3, p < 0.001; duration: F(1,40) = 17.5, p < 0.001; Table 5 and Figure 2A].
P3a
For the gap insertion, significant P3a responses were found at both time points T1 and T2 for both CI and NH schoolchildren but not for CI nor NH preschoolers (Table 4 and Figure 1). For changes in vowel duration, only CI preschoolers had significant, positive P3a responses at T1 (Table 4). Notably, for the NH groups, responses were significant at the P3a time window, but negative in polarity for both time points T1 and T2. Thus, only the P3a to gap insertion was included in the group comparisons of P3a (Table 4). The P3a was larger for the CI group than for the NH group [F(1,40) = 4.79, p = 0.035; Table 5 and Figure 2A). There was also a main effect of age group [F(1,40) = 6.13, p = 0.018; Table 5 and Figure 2B]. Across CI and NH groups, the P3a was larger for schoolchildren than for preschoolers.
LDA
For gap insertions, LDNs were negative and significant for CI preschoolers and schoolchildren at T1, and for NH preschoolers at T2, but not significant for NH schoolchildren any time point. Instead, for changes in vowel duration, the LDNs were significant and negative for both NH age groups at both time points (Table 4 and Figure 1) while not significant for neither CI group at neither time point. Thus, only the LDN to gap insertions was included in group comparisons (Table 4). There was a main effect of age group [F(1,40) = 7.00, p = 0.012]. Across CI and NH groups, and as opposite to the results for gap P3a, the LDN was larger for preschoolers than for schoolchildren (Table 5 and Figure 2B).
pMMR, MMN, P3a, and LDA to (with CIs) difficult-to-detect changes in vowel identity, pitch (F0) and intensity
pMMR
For changes in vowel identity, significant pMMRs were elicited in CI preschoolers at T2, in CI schoolchildren at T1 and T2, and in NH preschoolers at T2 (Table 4). A significant pMMR to intensity decrements was elicited in the CI preschoolers at T1 (Table 4 and Figure 3). It was possible to include only the pMMR to changes in vowel identity in the statistical group comparisons (see Table 4). The pMMR of the CI group was larger than that of NH group [F(1,40) = 4.81, p = 0.034; Table 5 and Figure 4A]. In addition, across CI and NH groups, the preschoolers had larger pMMRs than schoolchildren at T2 [F(1,40 = 4.15, p = 0.048; Table 5 and Figure 4B]. A further interaction of age group and time [F(1,41) = 8.36, p = 0.006] resulted from (a) pMMRs in the preschoolers being smaller at T1 than at T2 (p = 0.001) whereas there was no statistically significant difference between T1 and T2 for the schoolchildren (p > 0.05) and (b) from preschoolers having larger pMMRs than schoolchildren at T2 (p = 0.048) but not at T1 (Table 5 and Figure 4B).
Figure 3. Difference waveforms (ERPs of deviant minus standard) from the six-electrode ROI (F3, Fz, F4, C3, Cz, C4) for the difficult-to-detect change types of vowel identity, F0 and intensity changes at measurement time points T1 and T2 for the preschoolers (age below 6 years 9 months) and schoolchildren (age above 6 years and 9 months) with CI and NH. The data is referenced to the nose. Significant responses are marked with an asterisk. ERP, event related potentials; CI, children with cochlear implants; NH, normally hearing children; pMMR, positive mismatch response; MMN, mismatch negativity; P3a, positive ERP following MMN; LDN, late differentiating negativity; F0, fundamental frequency.
Figure 4. Barcharts of significant main effects and interactions in Linear Mixed Modeling for the difficult-to-detect change types of vowel identity, F0 and intensity changes. (A) CI vs. NH group main effects; (B) age (preschoolers vs. schoolchildren) main effects and age (preschoolers vs. schoolchildren) and time (measurement 1 and measurement 2) interactions; pMMR, positive mismatch response; LDN, late differentiating negativity; CI, children with cochlear implants; NH, normally hearing children; PR, preschoolers; SC, schoolchildren. *p < 0.05, **p < 0.01. Note the different scales for different responses.
MMN
Statistical group comparisons could not be conducted for MMN responses to difficult-to-detect changes due to the complete lack of significant MMN responses in the CI group. Instead, a significant but positive response was elicited at the MMN time window in the CI preschoolers for the intensity decrement at T1. In contrast, at least one of the NH groups had a significant MMN to all deviant types at either T1 or T2 (Table 4 and Figure 3). More specifically, at T1, the MMNs were significant in the NH preschoolers for vowel identity changes and both F0 changes, and for the NH schoolchildren for vowel identity changes as well as intensity increments. At T2, MMNs to vowel identity changes were no longer significant in the NH preschoolers but were still significant in the NH schoolchildren. Furthermore, at T2, the NH preschoolers had significant MMNs to intensity increments and the NH schoolchildren for intensity decrements (Table 4 and Figure 3).
P3a
No significant P3a responses with correct polarity were elicited in the NH group to the difficult-to-detect changes (Table 4 and Figure 1). In contrast, one or both CI groups had at T1 significant P3a responses to all difficult-to-detect change types except intensity increments (Table 4). Significant P3as at T1 were elicited to the F0 15% and F0 50% changes as well as intensity decrements in the CI preschoolers, and to vowel identity, both F0 changes, and to intensity decrements in the schoolchildren. However, none of these responses were significant at T2 (Table 4 and Figure 3). In the NH children at T1, significant but negative responses were elicited in the P3a time window to F0 50% change in preschoolers, and to vowel identity and intensity increment changes in the schoolchildren. Statistical group comparisons could not be conducted due to lack of significant correct-polarity P3a responses in the NH group.
LDA
In the CI group, the only significant LDN response was elicited by the vowel identity changes in the CI schoolchildren at T2 (Table 4 and Figure 3). Notably, F0 50% change elicited only a significant positive response in the CI preschoolers at T1 in the time window for LDN. In the NH groups, significant LDN responses were elicited to all difficult-to-detect changes except F0 15% at some timepoint. Specifically, vowel identity and F0 50% changes elicited significant LDNs in both NH groups at both timepoints, the intensity decrement in NH schoolchildren at T1 and both NH age groups at T2, and the intensity increment in both NH age groups at T2 (Table 4 and Figure 3). Only the LDNs to vowel identity changes could be compared statistically between groups (Table 4). The LDN to vowel identity changes was smaller in the children with CIs than the NH children [F(1,40) = 13.1, p < 0.001; Figure 4A and Table 5].
Children with cochlear implants and singing
Eighteen responses fulfilled the criteria of having the correct polarity and being significant (p < 0.05) in either T1 or T2 (or both) in the combined CI group (N = 21). These responses were the pMMRs to the gap insertion, duration increment, vowel identity change, and intensity decrement and increment, the MMNs to the gap insertion, duration increment, F0 15% change and intensity decrement, the P3a responses to the gap insertion, duration increment, vowel identity change, F0 15 and 50% changes, and intensity decrement, and LDNs to the gap insertion, and F0 15 and 50% changes (see Supplementary material 2 for amplitudes and p-values). Stable models could be estimated for both singing scores for pMMRs to duration increments, vowel identity changes, and the intensity decrement, for MMNs to vowel duration increment, and for P3as to the gap insertion, vowel identity changes, F0 15% changes, and intensity decrements. A stable model was also estimable for the LDN to F0 50% changes and the child’s singing score, but not the parental singing score. For the remaining responses the data was not sufficient for stable model estimation.
For those models that were estimable, most yielded only significant main effects of time (repeating those observed in the CI vs. NH analysis, and thus not reported here) or no significant results at all. However, a significant interaction of time (T1 vs. T2) and parental singing was found for pMMR [F(1,19) = 5.89, p = 0.025] and P3a [F(1,19) = 4.76, p = 0.042] to vowel identity changes (Table 5 and Figure 5). Importantly, for the pMMR, the response sizes had clearly increased from T1 to T2 in those children whose parents sang for them average or above the mean of the amount of parental singing (see Figure 3A; regression lines meet at parental singing level ≈ 2). However, for the P3a, less parental singing was seen as a diminishing P3a from T1 to T2, whereas in those children whose parents sang more for them, the P3a amplitudes remained at about the same level between T1 and T2 (see Figure 5; regression lines meet at parental singing level ≈ 4).
Figure 5. The interaction of parental singing and pMMR (upper figure) and P3a (lower figure) at time points of measurement T1 and T2 in the CI group as found in the LMM. Parental singing is a combination of the assessments by the parents themselves on a 1–5 Likert scale at T1 and T2. Note the intersection of the illustrative regression lines, which pinpoints the level of parental singing where the responses are approximately of equal size at T1 and T2. Thus, for the pMMR, singing increases the response size from T1 to T2 from an approximate singing level = 2, and for the P3a, singing sustains the size of P3a from T1 to T2 from an approximate singing level = 4.
Discussion
The current study compared, to the best of our knowledge, for the first time the development of all four cognitive ERP responses pMMR, MMN, P3a and LDN to changes in pseudowords between children with CIs and their NH controls, aged 4–13 years. The ERP responses were recorded twice, 14–17 months apart (at T1 and T2). Children were divided into preschoolers (under the age of 6 years 9 months) and schoolchildren (over the age of 6 years 9 months). The differences in the development of responses between the children with CIs and NH were most evident for the change types which are, based on the previous findings and functioning of the CIs, difficult to perceive with CIs (F0, intensity and vowels; see Drennan and Rubinstein, 2008; Limb and Roy, 2014). Overall, the results were rather similar to those from the previous studies using musical stimuli (Torppa et al., 2012, 2018, 2014b). Importantly, the development of the pMMR and the P3a responses to changes in vowel identity were linked to parental singing in children with CIs, consistent with the findings from Torppa et al. (2018).
Event-related potential responses and their development for the easy-to-detect changes
In line with the hypothesis 1a, the pMMR was not elicited to the easy-to-detect change types in the NH group. At T1 (the first time point of EEG recordings), the pMMR was found to changes in vowel duration in the CI preschoolers, and to gap insertions between syllables in the CI schoolchildren, both responses not being significant at T2. Thus, the results implicate that while the auditory system of children with CIs registered the changes in duration and gap insertions between syllables, the processing of these changes was immature at T1, but no longer at T2, as the pMMR is a response typically seen only in young children (Morr et al., 2002; Shafer et al., 2010). Moreover, the pMMR to gap insertions might be a sensitive marker the state of cortical maturation of speech processing in later-implanted children aged between 6 years 9 months and 13 years.
Hypothesis 1b of significant MMN responses for the easy-to-detect changes in both CI and NH groups was supported. For both CI and NH age groups, the MMN responses to gap insertions were elicited at both time points of measurements, and this was the case also for the MMN responses to changes in duration in the NH group. However, the MMN to vowel duration changes was significant for both CI age groups only at T2 (see Table 4). Thus, the appearance of MMN to duration changes at T2 in the children with CIs might be related to the maturation of cortical processing and related brain networks since simultaneously, pMMR to changes in duration disappeared at T2. It is possible that when cortical processing of children with CIs is unmature, discrimination duration changes is reflected in their pMMR responses. However, when cortical processing matures, the discrimination is reflected in their MMN responses. This is in line with the previous findings, showing development of the pMMR toward a MMN in young children with NH (see, for instance, Shafer et al., 2010).
Also supporting the hypothesis 1b, the MMNs to both easy-to-detect change types were smaller in the CI group than in the NH group, suggesting poorer sound discrimination for children with CIs (see Näätänen et al., 2017). Thus, there is poorer cortical discrimination of these changes despite the fact that CIs can deliver these gross temporal changes rather well. It is possible that the presence of pMMR responses in the CI participants decreased their MMN responses to the extent that the MMN was found to be smaller in statistical testing. Unfortunately, the children could not be tested behaviorally for their ability to discriminate the stimuli due to their young age (the youngest were 4 years old), and thus we do not know to what extent the differences in their cortical responses affect their discrimination performance. However, the present results are consistent with the findings in Torppa et al. (2012) with the same participants as in the present study. When the children were presented musical stimuli with a fast tempo mimicking real music, the MMN to gap insertions peaked later, and the MMN to changes in duration was smaller and later for children with CIs than for children with NH, suggesting poorer discrimination of these changes in the CI group. For the gap changes, this can be explained by simultaneous spectral and amplitude cues making gap detection difficult with CIs, as Sagi et al. (2009) proposed. They found poorer gap detection for postlingually deafened English-speaking adults with CIs than for NH controls when they identified gaps in synthesized speech stimuli. Furthermore, the fast tempo of the present multifeature paradigm could also make both duration change and gap insertion detection difficult for children with CIs.
Interestingly, and supporting hypothesis 1c, significant P3a responses to gap insertions were found at both time points for the schoolchildren (but not for the preschoolers) in both CI and NH groups, suggesting that these changes were distractive and captured the attention of schoolchildren only. Moreover, against hypothesis 1c, and despite the smaller MMNs to gap insertions, the gap P3a was larger in the CI than the NH group, and larger in schoolchildren (combining CI and NH groups) than in preschoolers (combining CI and NH groups). The larger P3a for gap insertions in the CI group may be related to the late CI activation of the CI schoolchildren participating the present study. It is known that later implantation leads to higher reliance in visual and tactile stimuli (for a review, see Glick and Sharma, 2017). This might further lead to higher reliance on reading and writing in the development of processing of sound changes affecting word meaning in school-aged, later-implanted children with CIs.
Overall, as assumed in our hypothesis 1b, schoolchildren in both groups have possibly learned to pay more attention to silent gaps between syllables since they have had to learn to distinguish gap insertions with reading and writing [in Finnish, semantically relevant gap insertions are spelled with double letters, distinguishing, e.g., between taka (hind) and takka (fireplace)]. Thus, learning to read and write could lead to stronger attention shifts toward gaps between consonants, especially as it is an audiovisual exercise and thus gives additional cues as to where to expect gaps in Finnish speech. The interpretation on the effect of reading instruction is supported also by the results of Engström et al. (2020). They found that in some children with CIs, multisensory learning with GraphoGame, in which children learn to map speech sounds to letters, led to a change from negative to positive ERP responses. Finally, singing, which is also a multisensory activity, was associated with enhanced P3a responses and their development in the study by Torppa et al. (2014b). Thus, there is a possibility that multisensory learning led to more efficient attention call toward gap insertions reflected in P3a.
In line with hypothesis 1d and previous findings from Uhlén et al. (2017) and Hu et al. (2021), the LDN to changes in vowel duration was elicited for the NH group (for both age groups and time points of measurements) while not for the CI group. Thus, as we expected, it seems that while for children with NH aged 4–13 years, the cortical maturation and accuracy of change detection allows further cognitive processing of duration changes, this is not the case for their peers hearing with CIs. However, again, the processing of changes in duration differed from that of gap insertions. Against our hypothesis 1d, the LDNs to gap insertions were elicited in both CI age groups at T1, and as we expected, in the NH preschoolers at T2, but not in NH schoolchildren at T1 or T2. Statistical group comparisons showed that across CI and NH groups, preschoolers had larger LDN responses than schoolchildren. These results suggest the possibility that while the LDN was still growing in preschool-aged children with NH, it was already disappearing in the school-aged children with NH. In line with the latter, previous findings on LDN have showed that LDN diminishes with age for the listeners who do not have problems with hearing or language processing, and is absent by adulthood (Bishop et al., 2011; Kuuluvainen et al., 2014, 2016a; Liu et al., 2014). The present results may indicate, similarly to the results for P3a for duration changes, that that the development is similar but later for children with CIs. Notably, the present findings are in accordance with the assumption that while the elicitation of the LDN is associated with immaturity of the cortex, possibly indexing a need for further processing, the immaturity disappears with age.
Event-related potential responses and their development for the difficult-to-detect changes
We classified the changes in pitch (F0), intensity and vowel identity as difficult-to detect changes based on the previous findings and functioning of the CIs (Geers et al., 2003; Donaldson and Kreft, 2006; Drennan and Rubinstein, 2008; for a review, see Rødvik et al., 2018). Overall, for the CI group, all significant responses were positive except for the LDN response to changes in vowel identity at T2 in CI schoolchildren. Moreover, except for the pMMRs to vowel identity changes, for the NH group all significant responses to the difficult-to-detect changes were negative, emphasizing the crucial differences between these child groups.
In more detail, as we expected in our hypothesis 1a, and similarly to the easy-to-detect-changes, there were more significant pMMRs in the CI than the NH group. Surprisingly, for NH preschoolers, the pMMR to changes in vowel identity was significant only at T2. Surprising was also that, across CI and NH groups, for the preschoolers, the pMMR amplitude to vowel identity changes became larger by time. This contradicts with the findings that pMMRs changes to negative MMNs already at the age of 4–7 years (Morr et al., 2002; Shafer et al., 2010), and with the present findings indicating the disappearance of pMMR to changes in vowel duration and gap insertions between consonants in the children with CIs (see section “ERP responses and their development for the easy-to-detect changes”). Regarding the vowel identity changes, the previous findings are based on experiments where only one vowel change at a slow stimulus rate is presented (see, e.g., Shafer et al., 2010). In contrast, in the present study, vowel changes were presented with a fast stimulus rate, and they were embedded in the middle of a pseudoword, evidently making discrimination much more difficult than in the previous studies. All in all, it is thus possible that only for more difficult discrimination tasks such as the vowel identity discrimination in a sequence presented in a fast tempo, typical for speech in everyday life, the processing reflected by the pMMR matures in children with NH and CI only after the age of 13 years. However, further studies are needed to confirm our interpretations.
In line with our hypothesis 1b, MMN was elicited in children with NH to all difficult-to-detect change types by at least one age group at one time point. However, against our hypotheses, and contradicting with the present findings for the easy-to-detect changes, there was a complete lack of MMN to the difficult-to-detect changes in the children with CIs. Instead, in the time window for the MMN, most responses in the CI groups were not significant, the significant but positive response in the CI preschoolers to the intensity decrement at T1 being the only exception. The positivity of this response in the CI preschoolers in the time window of the MMN is consistent with the previous findings by Ortmann et al. (2013a,2017), Torppa et al. (2014a,2018), and Engström et al. (2020). One possible reason for the lack of significant responses to the intensity increments in the CI group is the activation of the automatic gain control (ACG) of the CI device (Stöbich et al., 1999). Based on the present and previous results on intensity increments, in future studies it might be better to use 60 dB–65 db SPL sound level for both child groups (children with CIs and NH) to avoid the effects of AGC.
Notably, when pMMR to changes in vowel identity appeared at T2 in the NH preschoolers, their MMN disappeared. This is in line with the findings that the polarity of mismatch responses to changes in vowels can vary in children for unknown reason (see Lee et al., 2012). Also the MMN to changes in F0 disappeared at T2 for NH preschoolers. Interestingly, this could be related to the relevance of F0 changes for younger children with NH, since F0 is the main auditory cue for sentence and word stress (see Torppa et al., 2014a), important especially for young children’s language learning (sentence stress: Männel and Friederici, 2013; word stress: Friedrich et al., 2009; Vavatzanidis et al., 2015) but not to the same extent any more for older children who already have good lexical skills (Mattys et al., 2005).
Moreover, against our hypothesis 1c, we found, at T1, P3a responses to difficult-to-detect changes only in children with CIs, more specifically to vowel identity changes, small and large changes in pitch (F0), and to intensity decrements – however, these responses were not significant any more at T2. In contrast, the NH children had significant but negative responses in the time window of the P3a, suggesting that their MMNs were prolonged to this time window, and possibly obscuring the P3as. We assume that the findings were related to interplay between increased listening effort in the CI groups, the consequences of that for activation of attention-related frontal areas of the brain, and the development of sound change discrimination and attention and related cortical networks.
It is known that the processing of degraded speech is highly dependent on auditory attention (e.g., Wild et al., 2012). Thus, when it is difficult to hear, listening becomes effortful and higher attention is needed for discrimination of speech. This has been shown for adults with hearing impairments, for whom increased listening effort leads to increased activation in cortical regions supporting executive function, attention, memory, and sensorimotor processing (Dimitrijevic et al., 2019). Moreover, for children with CIs, Ortmann et al. (2013a) found that when children with CIs showed positive responses in the MMN time window, their attention-related brain areas were activated very early. Thus, it is possible that the appearance of P3a responses at T1 and their disappearance at T2 only in children with CIs, reflects increase and decrease in listening effort. This interpretation is tentatively supported by the findings of Bertoli and Bodmer (2014), who found that in older adults with hearing loss, increases in the size of the P3a were associated with increased listening effort. The decrease in listening effort related to P3a would be consistent with the findings that the speech perception of children with CIs improves over time (Lee et al., 2002; Davidson et al., 2011; Dunn et al., 2014; Moein et al., 2017), evidently leading to less listening effort over time. Interestingly, listening effort could play a role also in the results for children with NH from Lee et al. (2012), showing that when speech stimulus contrasts were obviously harder to discriminate, positive mismatch responses were found especially in younger children. It should be noted that the responses in Lee et al. (2012) were elicited at around 300 ms, and were preceded by a small negative-going deflection at a similar latency to the MMN to the easier contrasts (see Figures 3–5 in Lee et al., 2012). This raises the question of whether the pMMRs in Lee et al. (2012) are more akin to the MMN or rather enhanced P3a responses. Disentangling the pMMRs and P3as in children with no MMNs is thus a question that warrants further research.
As we predicted (hypothesis 1d), for the NH groups, significant LDN responses were elicited to all difficult-to-detect change types and they were found at both time points of measurements for both age groups, although to the F0 deviant significant responses were elicited only to the 50% but not the 15% change. Thus, the finding was rather similar to the easy-to-detect changes, implying that for children with NH, cortical maturation, accuracy of change detection or relevance of the sound change allows further cognitive processing of changes in vowel identity, pitch (F0) and intensity. The LDN was elicited in the NH group also to several changes that did not elicit a significant MMN or P3a. However, the LDN did not show signs of disappearance in the school-aged children with NH, which is different from the present findings for gap insertions and previous findings for several other types of stimuli (Bishop et al., 2011; Kuuluvainen et al., 2014, 2016a; Liu et al., 2014). Thus, it is possible that the immaturity reflected in LDN disappears at a later age for changes in vowel identity, pitch (F0) and intensity than for the changes in duration and gap insertions. Moreover, it is possible that the auditory system of children with NH registers these changes as relevant and important until a later age than of gap insertions and duration changes, or behaves differently for syllables embedded in the middle of a word rather than for single vowels or syllables used in previous studies (Bishop et al., 2011; Kuuluvainen et al., 2014, 2016a; Liu et al., 2014). Interestingly, against our predictions based on findings from Hu et al. (2021), however in line with Singh (2005), at T2 LDN was elicited also in the CI schoolchildren to changes in vowel identity, even though the LDN was smaller for them than for children with NH. Thus, it seems that by age and time, the later cognitive processing of vowel identity changes develops in children with CIs – perhaps simultaneously with the developmental changes in discrimination of vowels or related cortical networks. Overall, further research is needed to understand the cognitive processes reflected by the LDN.
Parental singing was linked to positive mismatch response and P3a responses in children with cochlear implants
For the pMMR to vowel identity changes, the response sizes clearly increased from T1 to T2 in those children whose parental singing was average or above. These results suggest that parental singing improved vowel discrimination. First, from the responses typically reflecting sound discrimination (pMMR and MMN), only pMMR to vowel changes was elicited in the children with CIs. Second, elicitation of early positive MMR has been found to be associated with good speech perception in children with Cis (Torppa et al., 2018; Engström et al., 2020). Furthermore, singing seemed to maintain the P3a response sizes between T1 and T2, while with less singing, the P3a diminished over time. The results were similar to the results from the earlier studies with the same children, which showed that only for the children with CIs who sang regularly (“CI singers”) and whose parents sang more for them, the P3a was maintained or increased over time (Torppa et al., 2014b,2018).
However, since in the present study, the direction of the connection of vowel identity pMMR and P3a to parental singing was similar, it is possible that here pMMR and P3 reflected similar processes for children with CIs. The responses of children with CIs were completely positive, and it is possible that the P3a responses were late mismatch responses or vice versa, the pMMR is related to both discrimination and attention. P3a has been interpreted to reflect evaluative discrimination related to the activation of an attentional switch mechanism (Friedman et al., 2001; Horváth et al., 2008), it is larger and earlier in CI children with better speech recognition (Kileny et al., 1997) and becomes larger with effective auditory training (Uther et al., 2006). Thus, it reflects both discrimination and attention shift toward sound changes. This assumption is consistent with those of Ortmann et al. (2013b), indicating based on an MMN source localization study that for children with CIs who perform well in speech discrimination, attention-related frontal areas are activated more than for poor performers. Importantly, even though in the comparisons of children with CIs and NH increased P3a responses are probably signs of increased listening effort, the increase in P3a across children with CIs does not mean less effective discrimination or attention functions. As opposite, since parental singing is related to better languages skills and speech perception (Torppa et al., 2018, 2020), those who have better attention functions as reflected in P3a responses (or pMMR responses) can discriminate better the degraded signal from CIs, inherently leading to reliance on attention in speech perception. This is particularly important for deaf-born children with CIs, for whom the spoken language should be brought directly to their attention, since due to difficulties in hearing, they cannot rely on passive listening or incidental language learning as efficiently as children with normal hearing (see e.g., Cole and Flexer, 2019).
We assume that the change in the strength of cortico-cortical connections as a result of parental singing could contribute to the present results. The neural network for P3a is distributed across frontal, parietal and temporal (auditory) cortical regions (Takahashi et al., 2013), suggesting functional connectivity between them. At an early age, the frontal, attention-related areas are developing in all children (Casey et al., 2000), and the increase in white-matter in association cortices, important for the maturation of auditory orienting, is already strong before the age of 8–12 months in normal-hearing children (Kushnerenko et al., 2013). It is evident that this development is delayed for deaf-born children with CIs since congenital deafness can also lead to degradation in white-matter volume in the auditory cortex and thus fewer afferent and efferent fibers (Emmorey et al., 2003). Parental singing, particularly at an early age, could improve the connections of auditory temporal areas to attention-related, frontal brain areas since singing of normal-hearing adults is related to enhanced connectivity between frontal and temporal cortical regions (Halwani et al., 2011; Wan et al., 2014), it is known that singing arises the attention of young children with CIs (Ronkainen, 2011), parental singing evokes and keeps the attention of young children more efficiently than speech (Politimou et al., 2019), and attention toward sounds modulates activation in auditory cortical areas (Fritz et al., 2007; Woods and Alain, 2009; Woods et al., 2009).
Overall, the results indicate, that singing face to face is good training of children’s vowel perception since it allows lipreading and a multisensory context for the perception of the formant changes associated with vowel changes (for a review, see Torppa et al., 2018). In accordance with this, Torppa et al. (2018) found that in children with CIs, more singing was associated with larger and earlier P3a responses for changes in timbre, for which spectral changes are important auditory cues similarly to changes in vowel identity (for a review, Torppa et al., 2018). Furthermore, Lo et al. (2020) found that musical training including singing improves perception of spectral resolution and speech in noise of children with hearing loss. Moreover, as discussed earlier, parental singing was in the present participants related to better speech perception and language skills (Torppa et al., 2018, 2020). As singing has no foreseeable negative consequences, and it carries benefits shown by previous studies, musical activities, including singing to and with the child, has been recommended as rehabilitation for children with CIs, and hearing impairments in general (Torppa and Huotilainen, 2019). Current results endorse this recommendation. However, more studies in other languages than Finnish are needed to pinpoint the possible benefits to speech perception across languages.
Caveats and future directions
The present, mainly novel, results evoke many questions that call for answers in future studies. The sample size was small, and this was reflected as lack of statistical power in analysis for three-way interactions of clinical status, age group and time, and also regarding children’s own and parental singing, which was studied only in the children with CIs. However, it is hard to collect a large sample, not only in Finland but also in other countries, as mentioned in the review on studies on the relations of musical activities to children with CIs’s speech perception and language skills (Torppa and Huotilainen, 2019). It is clear that one should aim at larger sample sizes in the future studies related to cortical processing of speech, especially in children with CIs, possibly through joint efforts of different laboratories in different countries, and allowing also the disentangling of language-specific effects on cortical development in CI users.
The present participants were implanted with unilateral, old-generation CIs. Many of them were also implanted fairly late, some as late as at the age of three. It is known that bilateral and early implantation as well as new generation CI devices with new sound-processing technology lead to better speech perception performance (van Wieringen and Wouters, 2015; Hey et al., 2019; Warren et al., 2019). For instance, bilateral CIs allow listening with better ear, binaural summation, improved perception of speech in noise (if speech and noise sources are spatially separated), and localization of sounds, especially if the CIs are paired successfully (van Hoesel and Tyler, 2003; Litovsky et al., 2004; Hu and Dietz, 2015; van Wieringen and Wouters, 2015). These benefits are evidently important for good development of speech processing, and lack of these benefits can lead to poorer or different processing compared to bilaterally implanted children. Therefore, new research using similar paradigms are needed to assess the brain processing in today’s children with CIs, who are usually implanted at age of 1 or even before, are hearing with two CIs, and have new-generation devices. It would be especially valuable to study this group with methods better suited for source analysis, for example high-density EEG or optical imaging, which would allow pinpointing the neural processes underlying the different ERPs.
One possible issue is also the removal of the electric artifacts caused by the CI devices from the EEG signal. Even though ICA is a good method to eliminate the CI artifact from the auditory ERP responses when multichannel recordings are used as in the present study (Gilley et al., 2006; for a review, see Näätänen et al., 2017), ICA or residual artifact could affect the results between children with CIs and NH. We assume that this is not the case in the present study since the results between groups were clearly different for pMMR, MMN, P3a, and LDN, and between difficult- and easy-to-detect changes. Moreover, subtraction of responses to standards from responses to deviants was conducted which should eliminate the residual electric artifact from difference waveforms from which the responses were detected (Vavatzanidis et al., 2015). However, future studies are needed to compare ICA to other artifact elimination methods and to assess the role of other possible confounding aspects such as sampling rate or filtering in the evaluation of brain responses (see, for instance, Hu et al., 2015; Hu and Ewert, 2021; for other CI artifact elimination methods, see Wong and Gordon, 2009; Näätänen et al., 2017).
The present results also suggest that, over time, and with multisensory learning such as singing or learning to read and write, the sound change processing changes. However, in the current study, the impact of reading and writing instruction was only indirectly measured via the division of the children to preschoolers and schoolchildren. It is possible that the impact of overall cognitive development was more relevant to the observed changes in responses over time, as we know that the verbal short term memory as assessed with the digit span task increased between T1 and T2 (the increase is expected, as the score is available only in raw points and not standardized to age expectations due to the lack of up-to-date normative data for the test). However, the hypothesized impact of reading instruction is supported by the notion that the differences between preschoolers and schoolchildren emerged only for those changes which are relevant for accurate reading and writing in Finnish (gap insertions, vowel identity changes) and not to the changes in F0 or intensity. The interpretation is, in addition, supported by the results of Lovio et al. (2012) and Engström et al. (2020) using the GraphoGame intervention, which propose that reading instruction is indeed associated with changes in cortical sound processing in some children with CIs and NH.
Above all, future research is needed to confirm the interpretation that positive responses at the time window of MMN or preceding it in young children with CIs reflect early attention call or listening effort. More studies are also needed on how singing and visual or multisensory learning contribute to the development of ERP responses and underlying brain activity especially for children with CIs. It would also be important to study the children’s ability to discriminate the stimulus changes they hear in the ERP paradigm behaviorally, and thus gather data on how easy or hard the stimulus changes are to discriminate for both children with NH and CIs, taking into account the different challenges represented by different languages. Measuring the exact discrimination ability can be difficult if the participants are very young, but the present results reveal that the developmental changes in the brain processing of speech stimuli are evident starting from the age of four, and child-friendly behavioral discrimination paradigms could thus be very valuable. Finally, studies of whether the elicitation of the pMMR, MMN, P3a, and LDN at individual level is associated with the development of speech perception and discrimination, is another important area of study.
Conclusion
The present results show that the development of ERP responses to natural, fast changes in pseudowords differs between children with CIs and NH especially for the difficult-to-detect changes with CIs (F0, vowel identity and intensity), while less difference between these child groups in the development was found for the easy-to-detect changes (vowel duration and gap insertion). The present findings on parental singing and developmental changes propose that cortical processing for the difficult-to-detect changes in children with CIs is associated with attentional processes. As listening effort has been previously shown to increase in adult CI users compared to NH (Dimitrijevic et al., 2019), and it has been associated with P3a size in a context different from the current study (Bertoli and Bodmer, 2014), the importance of listening effort in children with CIs and its relationship with cortical speech processing should be investigated further.
Our results, along those in Torppa et al. (2012, 2014b,2018, 2020), show that the multifeature paradigm is a useful tool to assess cortical processing of speech and music in children with CIs, and provides a wide range of information on different processes, when all responses (pMMR, MMN, P3a, and LDN) are analyzed, as in the current study. In the future, alongside with ERP studies, brain development in children with CIs should be studied using methodologies suited for localization of cortical and subcortical activity. Understanding the developmental trajectories and their associations with speech perception, attention, and listening effort, is highly relevant for the rehabilitation of children with CIs. It is also necessary to further investigate the impact of different types of multisensory training, such as singing or reading instruction, for the perceptual skills and brain processes in these children. Finally, it is important to encourage parents of children with CIs to sing, since singing can improve their children’s perception of speech.
Data availability statement
The datasets presented in this article are not readily available because the data is sensitive. Requests to access the datasets should be directed to ritva.torppa@helsinki.fi.
Ethics statement
The studies involving human participants were reviewed and approved by the Ethical Committee of HUCH and other participating hospitals (University Hospitals of Helsinki, Kuopio, Tampere, and Turku). Written informed consent to participate in this study was provided by the participants’ legal guardian/next of kin. The study was carried out in accordance with the Declaration of Helsinki.
Author contributions
JL was responsible for statistical analyses. All authors listed have approved the work for publication.
Funding
Writing this article and data collection and analysis was supported by a grants to RT from the Signe and Ane Gyllenberg Foundation. The data collection and analyses were also supported by the Ella and Georg Ehrnrooth Foundation, the Finnish Concordia Fund, the Finnish Audiological Society, the Finnish Doctoral Program in Language Studies, funded by the Ministry of Education and Culture, and the Emil Aaltonen Foundation.
Acknowledgments
We thank the personnel, especially speech therapists (Nonna Virokannas, Sari Vikman, Satu Rimmanen, and Teija Tsupari), of university hospital CI clinics in Helsinki, Tampere, Turku, and Kuopio, and the students who participated in the data collection. We thank Johannes Pykäläinen and Hannu Loimo for conducting the artifact reduction (ICA). We thank Minna Huotilainen and Andrew Faulkner for their help and advice in the first stages of the present study. We thank Martti Vainio and Eino Partanen for constructing the stimuli, and for the latter for his helpful advice, as well as laboratory engineers Miika Leminen and Tommi Makkonen for their valuable contributions to the recording and analysis phases. Above all, we also thank the parents and children for their participation.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnins.2022.976767/full#supplementary-material
Abbreviations
CI, cochlear implant; EEG, electroencephalography; ERP, event-related potential; ICA, independent component analysis; LDN, late differentiating negativity; LMM, Linear Mixed Model; MMN, mismatch negativity; NH, normal hearing; pMMR, positive mismatch response; P3a, attention-related brain response peaking after MMN; T1 and T2, first and second time points of measurements.
References
Aro, M., and Wimmer, H. (2003). Learning to read: English in comparison to six more regular orthographies. Appl. Psycholinguist. 24, 621–635. doi: 10.1017/S0142716403000316
Bertoli, S., and Bodmer, D. (2014). Novel sounds as a psychophysiological measure of listening effort in older listeners with and without hearing loss. Clin. Neurophysiol. 125, 1030–1041. doi: 10.1016/j.clinph.2013.09.045
Bishop, D. V., Hardiman, M. J., and Barry, J. G. (2011). Is auditory discrimination mature by middle childhood? A study using time-frequency analysis of mismatch responses from 7 years to adulthood: Is auditory discrimination mature? Dev. Sci. 14, 402–416. doi: 10.1111/j.1467-7687.2010.00990.x
Boersma, P., and Weenink, T. (2001). Praat, a system for doing phonetics by computer. Glot Int. 5, 341–345.
Cai, J., Liu, Y., Yao, M., Xu, M., and Zhang, H. (2020). A Neurophysiological study of musical pitch identification in mandarin-speaking cochlear implant users. Neural Plast. 2020, 1–11. doi: 10.1155/2020/4576729
Campbell, J., and Sharma, A. (2016). Visual cross-modal re-organization in children with cochlear implants. PLoS One 11:e0147793. doi: 10.1371/journal.pone.0147793
Casey, B. J., Giedd, J. N., and Thomas, K. M. (2000). Structural and functional brain development and its relation to cognitive development. Biol. Psychol. 54, 241–257. doi: 10.1016/S0301-0511(00)00058-2
Cejas, I., Mitchell, C. M., Hoffman, M., and Quittner, A. L. (2018). Comparisons of IQ in children with and without cochlear implants: Longitudinal findings and associations with language. Ear Hear. 39, 1187–1198. doi: 10.1097/AUD.0000000000000578
Chatterjee, M., and Peng, S.-C. (2008). Processing F0 with cochlear implants: Modulation frequency discrimination and speech intonation recognition. Hear. Res. 235, 143–156. doi: 10.1016/j.heares.2007.11.004
Chatterjee, M., Kulkarni, A. M., Siddiqui, R. M., Christensen, J. A., Hozan, M., Sis, J. L., et al. (2019). Acoustics of emotional prosody produced by prelingually deaf children with cochlear implants. Front. Psychol. 10:2190. doi: 10.3389/fpsyg.2019.02190
Chatterjee, M., Zion, D. J., Deroche, M. L., Burianek, B. A., Limb, C. J., Goren, A. P., et al. (2015). Voice emotion recognition by cochlear-implanted children and their normally-hearing peers. Hear. Res. 322, 151–162. doi: 10.1016/j.heares.2014.10.003
Cole, E. B., and Flexer, C. (2019). Children with hearing loss: Developing listening and talking, birth to six. San Diego, CA: Plural Publishing.
Corbeil, M., Trehub, S. E., and Peretz, I. (2013). Speech vs. Singing: Infants choose happier sounds. Front. Psychol. 4:372. doi: 10.3389/fpsyg.2013.00372
Davidson, L. S., Geers, A. E., Blamey, P. J., Tobey, E. A., and Brenner, C. A. (2011). Factors contributing to speech perception scores in long-term pediatric cochlear implant users. Ear Hear. 32, 19S–26S. doi: 10.1097/AUD.0b013e3181ffdb8b
Dehaene-Lambertz, G., and Dehaene, S. (1994). Speed and cerebral correlates of syllable discrimination in infants. Nature 370:6487. doi: 10.1038/370292a0
Delorme, A., and Makeig, S. (2004). EEGLAB: An open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 134, 9–21. doi: 10.1016/j.jneumeth.2003.10.009
Dimitrijevic, A., Smith, M. L., Kadis, D. S., and Moore, D. R. (2019). Neural indices of listening effort in noisy environments. Sci. Rep. 9:11278. doi: 10.1038/s41598-019-47643-1
Donaldson, G. S., and Kreft, H. A. (2006). Effects of vowel context on the recognition of initial and medial consonants by cochlear implant users. Ear Hear. 27, 658–677. doi: 10.1097/01.aud.0000240543.31567.54
Drennan, W. R., and Rubinstein, J. T. (2008). Music perception in cochlear implant users and its relationship with psychophysical capabilities. J. Rehabil. Res. Dev. 45, 779–789. doi: 10.1682/jrrd.2007.08.0118
Dunn, C. C., Walker, E. A., Oleson, J., Kenworthy, M., Van Voorst, T., Tomblin, J. B., et al. (2014). Longitudinal speech perception and language performance in pediatric cochlear implant users: The effect of age at implantation. Ear Hear. 35, 148–160. doi: 10.1097/AUD.0b013e3182a4a8f0
Emmorey, K., Allen, J. S., Bruss, J., Schenker, N., and Damasio, H. (2003). A morphometric analysis of auditory brain regions in congenitally deaf adults. Proc. Natl. Acad. Sci. U.S.A. 100, 10049–10054. doi: 10.1073/pnas.1730169100
Engström, E., Kallioinen, P., Lindgren, M., Nakeva von Mentzer, C., Sahlén, B., Lyxell, B., et al. (2020). Computer-assisted reading intervention for children with hearing impairment using cochlear implants: Effects on auditory event-related potentials and mismatch negativity. Int. J. Pediatr. Otorhinolaryngol. 137:110229. doi: 10.1016/j.ijporl.2020.110229
Escera, C., and Malmierca, M. S. (2014). The auditory novelty system: An attempt to integrate human and animal research. Psychophysiology 51, 111–123. doi: 10.1111/psyp.12156
Evans, S., and Davis, M. H. (2015). Hierarchical organization of auditory and motor representations in speech perception: Evidence from searchlight similarity analysis. Cereb. Cortex 25, 4772–4788. doi: 10.1093/cercor/bhv136
Friederici, A. D. (2017). Evolution of the neural language network. Psychon. Bull. Rev. 24, 41–47. doi: 10.3758/s13423-016-1090-x
Friedman, D., Cycowicz, Y. M., and Gaeta, H. (2001). The novelty P3: An event-related brain potential (ERP) sign of the brain’s evaluation of novelty. Neurosci. Biobehav. Rev. 25, 355–373. doi: 10.1016/S0149-7634(01)00019-7
Friedrich, M., Herold, B., and Friederici, A. D. (2009). ERP correlates of processing native and non-native language word stress in infants with different language outcomes. Cortex 45, 662–676. doi: 10.1016/j.cortex.2008.06.014
Fritz, J. B., Elhilali, M., David, S. V., and Shamma, S. A. (2007). Auditory attention—Focusing the searchlight on sound. Curr. Opin. Neurobiol. 17 (Supplement), 437–455. doi: 10.1016/j.conb.2007.07.011
Geers, A., Brenner, C., and Davidson, L. (2003). Factors associated with development of speech perception skills in children implanted by age five. Ear Hear. 24, 24S–35S. doi: 10.1097/01.AUD.0000051687.99218.0F
Gilley, P. M., Sharma, A., Dorman, M., Finley, C. C., Panch, A. S., and Martin, K. (2006). Minimization of cochlear implant stimulus artifact in cortical auditory evoked potentials. Clin. Neurophysiol. 117, 1772–1782.
Glick, H., and Sharma, A. (2017). Cross-modal plasticity in developmental and age-related hearing loss: Clinical implications. Hear. Res. 343, 191–201. doi: 10.1016/j.heares.2016.08.012
Haapaniemi, J. J. (1996). The hearing threshold levels of children at school age. Ear Hear. 17, 469–477.
Halwani, G. F., Loui, P., Rüber, T., and Schlaug, G. (2011). Effects of practice and experience on the arcuate fasciculus: Comparing singers, instrumentalists, and non-musicians. Front. Psychol. 2:156. doi: 10.3389/fpsyg.2011.00156
Hämäläinen, J. A., Lohvansuu, K., Ervast, L., and Leppänen, P. H. T. (2015). Event-related potentials to tones show differences between children with multiple risk factors for dyslexia and control children before the onset of formal reading instruction. Int. J. Psychophysiol. 95, 101–112. doi: 10.1016/j.ijpsycho.2014.04.004
Hey, M., Hocke, T., Böhnke, B., and Mauger, S. J. (2019). ForwardFocus with cochlear implant recipients in spatially separated and fluctuating competing signals–introduction of a reference metric. Int. J. Audiol. 58, 869–878. doi: 10.1080/14992027.2019.1638527
Hopyan-Misakyan, T. M., Gordon, K. A., Dennis, M., and Papsin, B. C. (2009). Recognition of affective speech prosody and facial affect in deaf children with unilateral right cochlear implants. Child Neuropsychol. 15, 136–146. doi: 10.1080/09297040802403682
Horváth, J., Winkler, I., and Bendixen, A. (2008). Do N1/MMN, P3a, and RON form a strongly coupled chain reflecting the three stages of auditory distraction? Biol. Psychol. 79, 139–147. doi: 10.1016/j.biopsycho.2008.04.001
Hu, H., and Dietz, M. (2015). Comparison of interaural electrode pairing methods for bilateral cochlear implants. Trends Hear. 19:2331216515617143.
Hu, H., and Ewert, S. D. (2021). “Exploring artifact rejection for high-pulse rate electrically evoked auditory steady state responses in cochlear implants users,” in 2021 Asia-Pacific signal and information processing association annual summit and conference (APSIPA ASC), (Tokyo: IEEE), 1311–1316.
Hu, H., Kollmeier, B., and Dietz, M. (2015). Reduction of stimulation coherent artifacts in electrically evoked auditory brainstem responses. Biomed. Signal Process. Control 21, 74–81.
Hu, Z., Sun, J.-Q., Guan, R.-R., Chen, L., Sun, J.-W., and Guo, X.-T. (2021). Deficient sensory and cognitive processing in children with cochlear implants: An event-related potential study. Hear. Res. 408:108295. doi: 10.1016/j.heares.2021.108295
Kileny, P., Boerst, A., and Zwolan, T. (1997). Cognitive evoked potentials to speech and tonal stimuli in children with implants. Otolaryngol. Head Neck Surg. 117, 161–169. doi: 10.1016/S0194-5998(97)70169-4
King, L. S., Camacho, M. C., Montez, D. F., Humphreys, K. L., and Gotlib, I. H. (2021). Naturalistic language input is associated with resting-state functional connectivity in infancy. J. Neurosci. 41, 424–434. doi: 10.1523/JNEUROSCI.0779-20.2020
Kirk, S. A., McCarthy, J. J., and Kirk, W. D. (1974). Illinois test of psycholinguistic abilities ITPA-Revised edition: Examiner’s manual. Champaign IL: University of Illinois Press.
Korkeamäki, R.-L., and Dreher, M. J. (1993). Finland, phonics, and whole language: Beginning reading in a regular letter-sound correspondence language. Lang. Arts 70, 475–482.
Korpilahti, P., Krause, C. M., Holopainen, I., and Lang, A. H. (2001). Early and late mismatch negativity elicited by words and speech-like stimuli in children. Brain Lang. 76, 332–339. doi: 10.1006/brln.2000.2426
Kral, A., and Sharma, A. (2012). Developmental neuroplasticity after cochlear implantation. Trends Neurosci. 35, 111–122. doi: 10.1016/j.tins.2011.09.004
Kushnerenko, E. V., Van den Bergh, B. R., and Winkler, I. (2013). Separating acoustic deviance from novelty during the first year of life: A review of event-related potential evidence. Front. Psychol. 4:595. doi: 10.3389/fpsyg.2013.00595
Kuuluvainen, S., Alku, P., Makkonen, T., Lipsanen, J., and Kujala, T. (2016a). Cortical speech and non-speech discrimination in relation to cognitive measures in preschool children. Eur. J. Neurosci. 43, 738–750. doi: 10.1111/ejn.13141
Kuuluvainen, S., Leminen, A., and Kujala, T. (2016b). Auditory evoked potentials to speech and nonspeech stimuli are associated with verbal skills in preschoolers. Dev. Cogn. Neurosci. 19, 223–232. doi: 10.1016/j.dcn.2016.04.001
Kuuluvainen, S., Nevalainen, P., Sorokin, A., Mittag, M., Partanen, E., Putkinen, V., et al. (2014). The neural basis of sublexical speech and corresponding nonspeech processing: A combined EEG–MEG study. Brain Lang. 130, 19–32. doi: 10.1016/j.bandl.2014.01.008
Lawrence, R. J., Wiggins, I. M., Hodgson, J. C., and Hartley, D. E. H. (2021). Evaluating cortical responses to speech in children: A functional near-infrared spectroscopy (fNIRS) study. Hear. Res. 401:108155. doi: 10.1016/j.heares.2020.108155
Lee, C.-Y., Yen, H., Yeh, P., Lin, W.-H., Cheng, Y.-Y., Tzeng, Y.-L., et al. (2012). Mismatch responses to lexical tone, initial consonant, and vowel in Mandarin-speaking preschoolers. Neuropsychologia 50, 3228–3239. doi: 10.1016/j.neuropsychologia.2012.08.025
Lee, K. Y. S., van Hasselt, C. A., Chiu, S. N., and Cheung, D. M. C. (2002). Cantonese tone perception ability of cochlear implant children in comparison with normal-hearing children. Int. J. Pediatr. Otorhinolaryngol. 63, 137–147. doi: 10.1016/S0165-5876(02)00005-8
Lerkkanen, M.-K., Rasku-Puttonen, H., Aunola, K., and Nurmi, J.-E. (2004). Developmental dynamics of phonemic awareness and reading performance during the first year of primary school. J. Early Child. Res. 2, 139–156. doi: 10.1177/1476718X04042972
Limb, C. J., and Roy, A. T. (2014). Technological, biological, and acoustical constraints to music perception in cochlear implant users. Hear. Res. 308, 13–26. doi: 10.1016/j.heares.2013.04.009
Litovsky, R. Y., Parkinson, A., Arcaroli, J., Peters, R., Lake, J., Johnstone, P., et al. (2004). Bilateral cochlear implants in adults and children. Arch. Otolaryngol. Head Neck Surg. 130, 648–655.
Liu, H.-M., Chen, Y., and Tsao, F.-M. (2014). Developmental changes in mismatch responses to mandarin consonants and lexical tones from early to middle childhood. PLoS One 9:e95587. doi: 10.1371/journal.pone.0095587
Lo, C. Y., Looi, V., Thompson, W. F., and McMahon, C. M. (2020). Music training for children with sensorineural hearing loss improves speech-in-noise perception. J. Speech Lang. Hear. Res. 63, 1990–2015. doi: 10.1044/2020_JSLHR-19-00391
Lovio, R., Halttunen, A., Lyytinen, H., Näätänen, R., and Kujala, T. (2012). Reading skill and neural processing accuracy improvement after a 3-hour intervention in preschoolers with difficulties in reading-related skills. Brain Res. 1448, 42–55. doi: 10.1016/j.brainres.2012.01.071
Luo, X., Fu, Q.-J., and Galvin, J. J. (2007). Cochlear implants special issue article: Vocal emotion recognition by normal-hearing listeners and cochlear implant users. Trends Amplif. 11, 301–315. doi: 10.1177/1084713807305301
Makeig, S., Debener, S., Onton, J., and Delorme, A. (2004). Mining event-related brain dynamics. Trends Cogn. Sci. 8, 204–210. doi: 10.1016/j.tics.2004.03.008
Männel, C., and Friederici, A. D. (2013). Accentuate or repeat? Brain signatures of developmental periods in infant word recognition. Cortex 49, 2788–2798. doi: 10.1016/j.cortex.2013.09.003
Mattys, S. L., White, L., and Melhorn, J. F. (2005). Integration of multiple speech segmentation cues: A hierarchical framework. J. Exp. Psychol. Gen. 134, 477–500. doi: 10.1037/0096-3445.134.4.477
Maurer, U., Bucher, K., Brem, S., and Brandeis, D. (2003). Development of the automatic mismatch response: From frontal positivity in kindergarten children to the mismatch negativity. Clin. Neurophysiol. 114:808. doi: 10.1016/s1388-2457(03)00032-4
McConkey Robbins, A. (2020). “Music and singing in auditory-verbal therapy,” in Auditory-verbal therapy—science, research, and practice, eds W. Estabrooks, H. M. Morrison, and K. MacIver-Lux (San Diego, CA: Plural Publishing).
Moein, N., Khoddami, S. M., and Shahbodaghi, M. R. (2017). A comparison of speech intonation production and perception abilities of Farsi speaking cochlear implanted and normal hearing children. Int. J. Pediatr. Otorhinolaryngol. 101, 1–6. doi: 10.1016/j.ijporl.2017.07.018
Moore, B. C. J. (2003). Coding of sounds in the auditory system and its relevance to signal processing and coding in cochlear implants. Otol. Neurotol. 24, 243–254. doi: 10.1097/00129492-200303000-00019
Morr, M. L., Shafer, V. L., Kreuzer, J. A., and Kurtzberg, D. (2002). Maturation of mismatch negativity in typically developing infants and preschool children. Ear Hear. 23, 118–136. doi: 10.1097/00003446-200204000-00005
Näätänen, R., Paavilainen, P., Rinne, T., and Alho, K. (2007). The mismatch negativity (MMN) in basic research of central auditory processing: A review. Clin. Neurophysiol. 118, 2544–2590. doi: 10.1016/j.clinph.2007.04.026
Näätänen, R., Petersen, B., Torppa, R., Lonka, E., and Vuust, P. (2017). The MMN as a viable and objective marker of auditory development in CI users. Hear. Res. 353, 57–75. doi: 10.1016/j.heares.2017.07.007
O’Halpin, R. (2010). The perception and production of stress and intonation by children with cochlear implants. London: University of London.
Ortmann, M., Knief, A., Deuster, D., Brinkheetker, S., Zwitserlood, P., am Zehnhoff-Dinnesen, A., et al. (2013a). Neural correlates of speech processing in prelingually deafened children and adolescents with cochlear implants. PLoS One 8:e67696.
Ortmann, M., Knief, A., Deuster, D., Brinkheetker, S., Zwitserlood, P., am Zehnhoff-Dinnesen, A., et al. (2013b). Neural correlates of speech processing in prelingually deafened children and adolescents with cochlear implants. PLoS One 8:e67696. doi: 10.1371/journal.pone.0067696
Ortmann, M., Zwitserlood, P., Knief, A., Baare, J., Brinkheetker, S., am Zehnhoff-Dinnesen, A., et al. (2017). When hearing is tricky: Speech processing strategies in prelingually deafened children and adolescents with cochlear implants having good and poor speech performance. PLoS One 12:e0168655. doi: 10.1371/journal.pone.0168655
Oxenham, A. J. (2018). How we hear: The perception and neural coding of sound. Ann. Rev. Psychol. 69, 27–50. doi: 10.1146/annurev-psych-122216-011635
Paquette, N., Vannasing, P., Lefrançois, M., Lefebvre, F., Roy, M.-S., McKerral, M., et al. (2013). Neurophysiological correlates of auditory and language development: A mismatch negativity study. Dev. Neuropsychol. 38, 386–401. doi: 10.1080/87565641.2013.805218
Paquette, S., Ahmed, G. D., Goffi-Gomez, M. V., Hoshino, A. C. H., Peretz, I., and Lehmann, A. (2018). Musical and vocal emotion perception for cochlear implants users. Hear. Res. 370, 272–282. doi: 10.1016/j.heares.2018.08.009
Partanen, E., Kujala, T., Näätänen, R., Liitola, A., Sambeth, A., and Huotilainen, M. (2013a). Learning-induced neural plasticity of speech processing before birth. Proc. Natl. Acad. Sci. U.S.A. 110, 15145–15150. doi: 10.1073/pnas.1302159110
Partanen, E., Torppa, R., Pykäläinen, J., Kujala, T., and Huotilainen, M. (2013b). Children’s brain responses to sound changes in pseudo words in a multifeature paradigm. Clin. Neurophysiol. 124, 1132–1138. doi: 10.1016/j.clinph.2012.12.005
Partanen, E., Vainio, M., Kujala, T., and Huotilainen, M. (2011). Linguistic multifeature MMN paradigm for extensive recording of auditory discrimination profiles: Linguistic multifeature MMN paradigm. Psychophysiology 48, 1372–1380. doi: 10.1111/j.1469-8986.2011.01214.x
Pihko, E., Kujala, T., Mickos, A., Antell, H., Alku, P., Byring, R., et al. (2005). Magnetic fields evoked by speech sounds in preschool children. Clin. Neurophysiol. 116, 112–119. doi: 10.1016/j.clinph.2004.07.005
Pisoni, D. B., Kronenberger, W. G., Roman, A. S., and Geers, A. E. (2011). Measures of digit span and verbal rehearsal speed in deaf children after more than 10 years of cochlear implantation. Ear Hear. 32, 60S–74S. doi: 10.1097/AUD.0b013e3181ffd58e
Politimou, N., Dalla Bella, S., Farrugia, N., and Franco, F. (2019). Born to speak and sing: Musical predictors of language development in pre-schoolers. Front. Psychol. 10:948. doi: 10.3389/fpsyg.2019.00948
Price, C. J. (2012). A review and synthesis of the first 20years of PET and fMRI studies of heard speech, spoken language and reading. NeuroImage 62, 816–847. doi: 10.1016/j.neuroimage.2012.04.062
Rahko, T., and Karma, P. (1989). Pure-tone hearing thresholds in otologically healthy 5-year-old children in Finland. Arch. Otorhinolaryngol. 246, 137–141. doi: 10.1007/BF00456654
Rødvik, A. K., von Koss Torkildsen, J., Wie, O. B., Storaker, M. A., and Silvola, J. T. (2018). Consonant and vowel identification in cochlear implant users measured by nonsense words: A systematic review and meta-analysis. J. Speech Lang. Hear. Res. 61, 1023–1050. doi: 10.1044/2018_JSLHR-H-16-0463
Ronkainen, R. J. (2011). Enhancing listening and imitation skills in children with cochlear implants—The use of multimodal resources in speech therapy. J. Interact. Res. Commun. Disord. 2, 245–269.
Sagi, E., Kaiser, A. R., Meyer, T. A., and Svirsky, M. A. (2009). The effect of temporal gap identification on speech perception by users of cochlear implants. J. Speech Lang. Hear. Res. 52, 385–395. doi: 10.1044/1092-4388(2008/07-0219)
Shafer, V. L., Morr, M. L., Datta, H., Kurtzberg, D., and Schwartz, R. G. (2005). Neurophysiological indexes of speech processing deficits in children with specific language impairment. J. Cogn. Neurosci. 17, 1168–1180.
Shafer, V. L., Yu, Y. H., and Datta, H. (2010). Maturation of speech discrimination in 4- to 7-Yr-old children as indexed by event-related potential mismatch responses. Ear Hear. 31, 735–745. doi: 10.1097/AUD.0b013e3181e5d1a7
Sharma, A., Campbell, J., and Cardon, G. (2015). Developmental and cross-modal plasticity in deafness: Evidence from the P1 and N1 event related potentials in cochlear implanted children. Int. J. Psychophysiol. 95, 135–144. doi: 10.1016/j.ijpsycho.2014.04.007
Shtyrov, Y., and Lenzen, M. (2017). First-pass neocortical processing of spoken language takes only 30 msec: Electrophysiological evidence. Cogn. Neurosci. 8, 24–38. doi: 10.1080/17588928.2016.1156663
Singer, J. D., and Willett, J. B. (2003). Applied longitudinal data analysis: Modeling change and event occurrence. Oxford: Oxford university press.
Singh, S. (2005). Correlation of electrophysiological measures of auditory processing with outcome in paediatric cochlear implant patients. London: University of London.
Stöbich, B., Zierhofer, C. M., and Hochmair, E. S. (1999). Influence of automatic gain control parameter settings on speech understanding of cochlear implant users employing the continuous interleaved sampling strategy. Ear Hear. 20, 104–116. doi: 10.1097/00003446-199904000-00002
Suomi, K., Toivanen, J., and Ylitalo, R. (2008). Finnish sound structure: Phonetics, phonology, phonotactics and prosody. Oulu: University of Oulu.
Takahashi, H., Rissling, A. J., Pascual-Marqui, R., Kirihara, K., Pela, M., Sprock, J., et al. (2013). Neural substrates of normal and impaired preattentive sensory discrimination in large cohorts of nonpsychiatric subjects and schizophrenia patients as indexed by MMN and P3a change detection responses. NeuroImage 66, 594–603. doi: 10.1016/j.neuroimage.2012.09.074
Talli, I., Tsalighopoulos, M., and Okalidou, A. (2018). The relation between short-term memory and vocabulary skills in Greek children with cochlear implants: The role of hearing experience. First Lang. 38, 359–381. doi: 10.1177/0142723717749073
Torppa, R., and Huotilainen, M. (2019). Why and how music can be used to rehabilitate and develop speech and language skills in hearing-impaired children. Hear. Res. 380, 108–122. doi: 10.1016/j.heares.2019.06.003
Torppa, R., Faulkner, A., Huotilainen, M., Järvikivi, J., Lipsanen, J., Laasonen, M., et al. (2014a). The perception of prosody and associated auditory cues in early-implanted children: The role of auditory working memory and musical activities. Int. J. Audiol. 53, 182–191. doi: 10.3109/14992027.2013.872302
Torppa, R., Faulkner, A., Kujala, T., Huotilainen, M., and Lipsanen, J. (2018). Developmental links between speech perception in noise, singing, and cortical processing of music in children with cochlear implants. Music Percept. 36, 156–174. doi: 10.1525/mp.2018.36.2.156
Torppa, R., Faulkner, A., Laasonen, M., Lipsanen, J., and Sammler, D. (2020). Links of prosodic stress perception and musical activities to language skills of children with cochlear implants and normal hearing. Ear Hear. 41, 395–410. doi: 10.1097/AUD.0000000000000763
Torppa, R., Huotilainen, M., Leminen, M., Lipsanen, J., and Tervaniemi, M. (2014b). Interplay between singing and cortical processing of music: A longitudinal study in children with cochlear implants. Front. Psychol. 5:1389. doi: 10.3389/fpsyg.2014.01389
Torppa, R., Salo, E., Makkonen, T., Loimo, H., Pykäläinen, J., Lipsanen, J., et al. (2012). Cortical processing of musical sounds in children with cochlear implants. Clin. Neurophysiol. 123, 1966–1979. doi: 10.1016/j.clinph.2012.03.008
Turgeon, C., Lazzouni, L., Lepore, F., and Ellemberg, D. (2014). An objective auditory measure to assess speech recognition in adult cochlear implant users. Clin. Neurophysiol. 125, 827–835. doi: 10.1016/j.clinph.2013.09.035
Uhlén, I., Engström, E., Kallioinen, P., Nakeva von Mentzer, C., Lyxell, B., Sahlén, B., et al. (2017). Using a multi-feature paradigm to measure mismatch responses to minimal sound contrasts in children with cochlear implants and hearing aids. Scand. J. Psychol. 58, 409–421. doi: 10.1111/sjop.12391
Uther, M., Kujala, A., Huotilainen, M., Shtyrov, Y., and Näätänen, R. (2006). Training in morse code enhances involuntary attentional switching to acoustic frequency: Evidence from ERPs. Brain Res. 107, 417–424. doi: 10.1016/j.brainres.2005.12.047
van Hoesel, R. J., and Tyler, R. S. (2003). Speech perception, localization, and lateralization with bilateral cochlear implants. J. Acoust. Soc. Am. 113, 1617–1630.
van Wieringen, A., and Wouters, J. (2015). What can we expect of normally-developing children implanted at a young age with respect to their auditory, linguistic and cognitive skills? Hear. Res. 322, 171–179. doi: 10.1016/j.heares.2014.09.002
Vavatzanidis, N. K., Mürbe, D., Friederici, A. D., and Hahne, A. (2016). The perception of stress pattern in young cochlear implanted children: An EEG study. Front. Neurosci. 10:68. doi: 10.3389/fnins.2016.00068
Vavatzanidis, N. K., Mürbe, D., Friederici, A., and Hahne, A. (2015). The basis for language acquisition: congenitally deaf infants discriminate vowel length in the first months after cochlear implantation. J. Cogn. Neurosci. 27, 2427–2441. doi: 10.1162/jocn_a_00868
Wan, C. Y., Zheng, X., Marchina, S., Norton, A., and Schlaug, G. (2014). Intensive therapy induces contralateral white matter changes in chronic stroke patients with Broca’s aphasia. Brain Lang. 136, 1–7. doi: 10.1016/j.bandl.2014.03.011
Warren, C. D., Nel, E., and Boyd, P. J. (2019). Controlled comparative clinical trial of hearing benefit outcomes for users of the CochlearTM Nucleus® 7 sound processor with mobile connectivity. Cochlear Implants Int. 20, 116–126. doi: 10.1080/14670100.2019.1572984
Wechsler, D. (2010). WISC-IV–wechsler intelligence scale for children—IV. Bloomington, IN: NCS Pearson, Inc.
West, B. T. (2009). Analyzing longitudinal data with the linear mixed models procedure in SPSS. Eval. Health Prof. 32, 207–228. doi: 10.1177/0163278709338554
Wetzel, N., and Schröger, E. (2014). On the development of auditory distraction: A review. PsyCh J. 3, 72–91. doi: 10.1002/pchj.49
Wetzel, N., Widmann, A., Berti, S., and Schröger, E. (2006). The development of involuntary and voluntary attention from childhood to adulthood: A combined behavioral and event-related potential study. Clin. Neurophysiol. 117, 2191–2203. doi: 10.1016/j.clinph.2006.06.717
Wild, C. J., Yusuf, A., Wilson, D. E., Peelle, J. E., Davis, M. H., and Johnsrude, I. S. (2012). Effortful listening: The processing of degraded speech depends critically on attention. J. Neurosci. 32, 14010–14021. doi: 10.1523/JNEUROSCI.1528-12.2012
Winkler, I., Tervaniemi, M., Schröger, E., Wolff, C., and Näätänen, R. (1998). Preattentive processing of auditory spatial information in humans. Neurosci. Lett. 242, 49–52. doi: 10.1016/S0304-3940(98)00022-6
Wong, D. D., and Gordon, K. A. (2009). Beamformer suppression of cochlear implant artifacts in an electroencephalography dataset. IEEE Trans. Biomed. Eng. 56, 2851–2857. doi: 10.1109/TBME.2009.2029239
Woods, D. L., and Alain, C. (2009). Functional imaging of human auditory cortex. Curr. Opin. Otolaryngol. Head Neck Surg. 17, 407–411.
Woods, D. L., Stecker, G. C., Rinne, T., Herron, T. J., Cate, A. D., Yund, E. W., et al. (2009). Functional maps of human auditory cortex: Effects of acoustic features and attention. PLoS One 4:e5183. doi: 10.1371/journal.pone.0005183
Keywords: cochlear implant, positive mismatch responses (pMMR), attention-related brain response P3a, mismatch negativity (MMN), late differentiating negativity (LDN), parental singing
Citation: Torppa R, Kuuluvainen S and Lipsanen J (2022) The development of cortical processing of speech differs between children with cochlear implants and normal hearing and changes with parental singing. Front. Neurosci. 16:976767. doi: 10.3389/fnins.2022.976767
Received: 23 June 2022; Accepted: 04 November 2022;
Published: 18 November 2022.
Edited by:
Hongmei Hu, University of Oldenburg, GermanyReviewed by:
Xiao-Tao Guo, University of Science and Technology of China, ChinaJi-Hye Han, Hallym University, South Korea
Copyright © 2022 Torppa, Kuuluvainen and Lipsanen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Ritva Torppa, ritva.torppa@helsinki.fi
†These authors have contributed equally to this work