Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 13 October 2022
Sec. Auditory Cognitive Neuroscience
This article is part of the Research Topic Early Neural Processing of Musical Melodies View all 5 articles

Musicians show more integrated neural processing of contextually relevant acoustic features

  • 1Aarhus Institute of Advanced Studies, Aarhus University, Aarhus, Denmark
  • 2Department of Clinical Medicine, Center for Music in the Brain, Aarhus University, Royal Academy of Music Aarhus/Aalborg, Aarhus, Denmark
  • 3Department of Dramaturgy and Musicology, School of Communication and Culture, Aarhus University, Aarhus, Denmark
  • 4Department of Linguistics, Cognitive Science, and Semiotics, School of Communication and Culture, Aarhus University, Aarhus, Denmark
  • 5Department of Clinical Medicine, Faculty of Health, Center of Functionally Integrative Neuroscience, Aarhus University, Aarhus, Denmark
  • 6Department of Psychology and Behavioural Sciences, Aarhus University, Aarhus, Denmark
  • 7School of Electronic Engineering and Computer Science, Cognitive Science Research Group and Centre for Digital Music, Queen Mary University of London, London, United Kingdom

Little is known about expertise-related plasticity of neural mechanisms for auditory feature integration. Here, we contrast two diverging hypotheses that musical expertise is associated with more independent or more integrated predictive processing of acoustic features relevant to melody perception. Mismatch negativity (MMNm) was recorded with magnetoencephalography (MEG) from 25 musicians and 25 non-musicians, exposed to interleaved blocks of a complex, melody-like multi-feature paradigm and a simple, oddball control paradigm. In addition to single deviants differing in frequency (F), intensity (I), or perceived location (L), double and triple deviants were included reflecting all possible feature combinations (FI, IL, LF, FIL). Following previous work, early neural processing overlap was approximated in terms of MMNm additivity by comparing empirical MMNms obtained with double and triple deviants to modeled MMNms corresponding to summed constituent single-deviant MMNms. Significantly greater subadditivity was found in musicians compared to non-musicians, specifically for frequency-related deviants in complex, melody-like stimuli. Despite using identical sounds, expertise effects were absent from the simple oddball paradigm. This novel finding supports the integrated processing hypothesis whereby musicians recruit overlapping neural resources facilitating more integrative representations of contextually relevant stimuli such as frequency (perceived as pitch) during melody perception. More generally, these specialized refinements in predictive processing may enable experts to optimally capitalize upon complex, domain-relevant, acoustic cues.

Introduction

The ability to distinguish and combine features of sensory input guides behavior by enabling humans to engage successfully with perceptual stimuli in their environment (Hommel, 2004; Spence and Frings, 2020). While sophisticated models exist for visual feature processing (Treisman and Gelade, 1980; Nassi and Callaway, 2009; Di Lollo, 2012; Grill-Spector and Weiner, 2014), auditory objects transform over time and remain more elusive (Griffiths and Warren, 2004; Shamma, 2008). Modality-specific divergences may therefore be expected. Indeed, findings that auditory feature conjunctions are processed pre-attentively (Winkler et al., 2005) and faster than single features (Woods et al., 1998) and that the identity features pitch and timbre sometimes take precedence over location (Maybery et al., 2009; Delogu et al., 2014) deviate from findings in visual perception (Campo et al., 2010; but see Guérard et al., 2013). Although auditory feature integration mechanisms are partly congenital (Ruusuvirta, 2001; Ruusuvirta et al., 2003), yet subject to some individual variation (Allen et al., 2017) and evolutionary adaptation (Fay and Popper, 2000), it remains unknown whether they vary with auditory expertise levels. Given its early onset and persistence throughout life (Ericsson, 2006) and its strong reliance on predictive processing mechanisms (Quiroga-Martinez et al., 2021), musicianship offers an especially informative model of auditory expertise (Vuust et al., 2005, 2022; Herdener et al., 2010; Herholz and Zatorre, 2012; Schlaug, 2015).

Electroencephalography (EEG) and magnetoencephalography (MEG) enable approximate estimation of feature integration using the additivity of the mismatch negativity response (MMN) and its magnetic counterpart (MMNm), respectively. The MMN(m) itself represents a deflection in the event-related potential or field (ERP/ERF) peaking around 150–250 ms after presentation of an unexpected stimulus (Näätänen et al., 1978). It results from active cortical prediction rather than from passive synaptic habituation (Wacongne et al., 2012). By comparing empirical MMN(m)s to double or triple deviants (differing from standards on two or three features) to modeled MMN(m)s obtained by summing the MMN(m) responses for the constituent single deviants, inferences have been made about the potential overlap in neural processing (e.g., Levänen et al., 1993; Paavilainen et al., 2001). Correspondence between empirical and modeled MMN(m)s is interpreted to indicate independent processing whereas subadditivity—where modeled MMN(m)s exceed empirical MMN(m)s—suggests more overlapping, integrated processing. In neurophysiological studies of multisensory integration, this well-established approach has been referred to as “the additive model” relying on the principle that “[b]iophysical laws state that the electrical fields generated by two generators add up linearly at any point measure” (Besle et al., 2009, p. 144). Thus, any observed sub-additivity (or super-additivity, for that matter) would point to an interaction between the two unimodal processes.

Although transferring “the additive model” to a single modality requires additional assumptions in terms of separate neural sources, for example, MMN(m) additivity has successfully been assessed within audition. Segregated feature processing has thus been established for frequency, intensity, onset asynchrony, and duration (Levänen et al., 1993; Schröger, 1995; Paavilainen et al., 2001; Wolff and Schröger, 2001; Paavilainen et al., 2003b) with spatially separate neural sources (Giard et al., 1995; Levänen et al., 1996; Rosburg, 2003; Molholm et al., 2005). MMN(m) is also additive for inter-aural time and intensity differences (Schröger, 1996), phoneme quality and quantity (Ylinen et al., 2005), and attack time and even-harmonic timbral attenuation (Caclin et al., 2006). Feature conjunctions occurring infrequently in the local context, moreover, result in distinct MMN(m) responses (Gomes et al., 1997; Sussman et al., 1998) that are separable in terms of neural sources (Takegata et al., 2001a) and extent of additivity from those elicited by constituent (Takegata et al., 1999) or more abstract pattern deviants (Takegata et al., 2001b).

Conversely, subadditive MMN(m)s occur for combinations of lexical tones, vowels, and consonants in Cantonese speech and non-speech (Choi et al., 2017; Yu et al., 2022), for direction of frequency and intensity changes (Paavilainen et al., 2003a), for aspects of timbre (Caclin et al., 2006), and between frequency deviants in sung stimuli and vowels (Lidji et al., 2009) and consonants (Gao et al., 2012). Generally, subadditivity is greater for triple than double deviants (Paavilainen et al., 2001; Caclin et al., 2006). While it is known that MMN(m) responses vary with expertise (Koelsch et al., 1999; Vuust et al., 2009), it remains unknown whether the same is the case for MMN(m) additivity.

The current MEG study aims to investigate whether MMNm additivity varies as a function of musical expertise. Specifically, assuming that MMNm additivity is a reliable proxy for independent feature processing, two hypotheses are contrasted. First, the independent processing hypothesis posits that expertise is associated with specialized feature processing by separate neural populations, increasing access to lower-level representations that have higher context-specific relevance (Ahissar and Hochstein, 2004; Ahissar et al., 2009). This would result in more similar empirical and modeled MMNm responses in musicians compared to non-musicians who, by contrast, would show greater subadditivity. Second, the integrated processing hypothesis posits that expertise is associated with processing of multiple features by shared neural resources, manifesting as decreased overall neural activity (Jäncke et al., 2001; Zatorre et al., 2012). This would manifest as smaller empirical than modeled MMNm responses expressed more prominently in musicians than in non-musicians.

Since expertise produces more accurate expectations (Hansen and Pearce, 2014; Hansen et al., 2016) and shorter MMN(m) latencies (Lappe et al., 2016) in musical contexts specifically, these hypotheses will be tested using contrasting paradigms with higher and lower levels of complexity and corresponding musical relevance. Deviants on the three acoustic features frequency, intensity, and location (in terms of inter-aural time difference) will be probed as expertise-related differences in MMN(m) responses to all these particular feature deviants have previously been demonstrated, albeit to different extent and with different levels of selectivity for the type of musical specialization (Nager et al., 2003; Tervaniemi et al., 2006, 2009; Brattico et al., 2009; Putkinen et al., 2014). Because these features represent decreasing degrees of musical relevance—with frequency being syntactically more important than intensity, and intensity being expressively more important than location—their inclusion will allow us to assess feature selectivity for any observed expertise differences in auditory feature integration.

Materials and methods

Participants

Twenty-five non-musicians (11 females; mean age: 24.7 years) and 25 musicians (10 females; mean age: 25.0 years) were recruited through the local research participation system and posters at Aarhus University and The Royal Academy of Music Aarhus. Members of the musician group were full-time conservatory students or professional musicians receiving their main income from performing and/or teaching music. The non-musician group had no regular experience playing a musical instrument and had received less than 1 year of musical training beyond mandatory music lessons in school. As shown in Table 1, the two groups were matched on age and sex, and musicians scored significantly higher on all subscales of Goldsmiths Musical Sophistication Index, v. 1.0 (Müllensiefen et al., 2014).

TABLE 1
www.frontiersin.org

Table 1. Demographics and musical sophistication of research participants.

All participants were right-handed with no history of hearing difficulties. Informed written consent was provided, and participants received a taxable compensation of DKK 300. The study was approved by The Central Denmark Regional Committee on Health Research Ethics (case 1-10-72-11-15).

Stimuli

Stimuli for the experiment were constructed from standard and deviant tones derived from Wizoo Acoustic Piano samples from Halion in Cubase 7 (Steinberg Media Technologies GmbH) with a 200 ms duration including a 5 ms rise time and 10 ms fall time.

Deviants differed from standards on one or more of three acoustic features: fundamental frequency (in Hz), sound intensity (in dB), and inter-aural time difference (in μs), henceforth referred to as frequency (F), intensity (I), and location (L), respectively. There are several reasons for focusing on these specific features. First, unlike alternative features such as duration (Czigler and Winkler, 1996) and frequency slide (Winkler et al., 1998), the point of deviance can be established unambiguously for frequency, intensity, and location deviants. Second, these features have reliably evoked additive MMN(m) responses in previous research; specifically, additivity has been demonstrated for frequency and intensity (e.g., Paavilainen et al., 2001) and frequency and location (e.g., Schröger, 1995), but not yet for location and intensity or for all three features together. Third, expertise-related selectivity for some of these features over others has been demonstrated in other contexts (Symons and Tierney, 2021). Finally, the restriction to three features balances representability and generalizability with practical feasibility within the typical timeframe of an MEG experiment.

In total, seven deviant types, comprising three single deviants, three double deviants, and one triple deviant, were generated through modification in Adobe Audition v. 3.0 (Adobe Systems Inc.). Specifically, frequency deviants (F) were shifted down 35 cents using the Stretch function configured to “high precision.” Intensity deviants (I) were decreased by 12 dB in both left and right channels using the Amplify function. Location deviants (L) resulted from delaying the right stereo track by 200μs compared to the left one. These parameter values were found to produce robust and relatively similar ERP amplitudes in a previous EEG study (Vuust et al., 2016). Double and triple deviants combining deviants in frequency and intensity (FI), intensity and location (IL), location and frequency (LF), and frequency and intensity and location (FIL) were generated by applying two or three of the operations just described, always in the order of frequency, intensity, and location.

Procedure

Two MMN paradigms were used in the experiment. Specifically, as shown in Figure 1A, four blocks (M1–M4) of the main complex paradigm, based on the musical multi-feature paradigm (cf., Vuust et al., 2011), were interleaved with three blocks (C1–C3) of a simpler control paradigm (cf., Näätänen et al., 2004) with lower degrees of musical relevance.

FIGURE 1
www.frontiersin.org

Figure 1. (A) Experimental procedure. Stimuli were presented in alternating blocks using the complex musical multi-feature paradigm (M1–M4) and the simple control paradigm (C1–C3). Each block comprised three (M) or four (C) iterations of a sequence of standards and deviants presented at twelve varying pitch levels (i.e., chromatic range of A#2–A3). At each pitch level, seven deviant types were used in permutation. These comprised three single deviants differing in frequency (-35 cents), intensity (-12 dB), or location (right stereo track delayed by 200 μs), double and triple deviants differing in frequency and intensity, intensity and location, location and frequency, or frequency and intensity and location. Epoching of the magnetoencephalographic recordings was time-locked to the onset of the musical notes marked with S for standards and D for deviants. (B) Additivity analysis. The additivity of neural mismatch responses was computed by subtracting modeled MMNms—corresponding to the sum of the two or three constituent single-deviant MMNms—from the empirical MMNm obtained in response to the relevant double or triple deviant. (C) The 18 combined gradiometers from the Elekta Neuromag TRIUX system that were used in analysis centered on the two sensors—MEG1342 + 1343 and MEG0232 + 0233—exhibiting the peak grand-average MMNm amplitude across both paradigms and across all participants and conditions.

The complex paradigm consists of repetitions of a characteristic four-note pattern referred to as the Alberti bass. In this pattern, the notes of a chord are arpeggiated in the order “lowest-highest-middle-highest” (Figure 1A). Although named after an Italian composer who used it extensively in early 18th-century keyboard accompaniment, the Alberti bass occurs widely across historical periods, instruments, and musical genres (Fuller, 2015).

The studies introducing this paradigm (Vuust et al., 2011, 2016, 2012a,b; Timm et al., 2014; Petersen et al., 2015) have typically modified every second occurrence of the third note in the pattern (termed “middle” above) by changing its frequency, intensity, perceived location, timbre, timing, or by introducing frequency slides. In the present implementation, two extra occurrences of the standard pattern were introduced between each deviant pattern (Figure 1A) to minimize spill-over effects from consecutive deviants some of which made use of the same constituent deviant types due to the inclusion of double and triple deviants. Independence of deviant types is an underlying assumption of multi-feature mismatch negativity paradigms (Näätänen et al., 2004).

Consistent with previous studies, individual notes of the pattern were presented with a constant stimulus onset asynchrony (SOA) of 205 ms. After each occurrence of all seven deviants, the pitch height of the pattern changed pseudo-randomly across all 12 notes of the chromatic scale. This resulted in standard tones with frequency values ranging from 116.54 Hz (A#2) to 220.00 Hz (A3). Each block of the musical multi-feature paradigm comprised three such iterations of the 12 chromatic notes (Figure 1A), resulting in a total of 144 trials of each deviant type across the four blocks.

The same number of trials per deviant type was obtained across the three blocks of the simple control paradigm, inspired by Näätänen et al. (2004). This was achieved by incorporating four rather than three iterations of the 12 chromatic notes (A#2 to A3) for each deviant array in each block (Figure 1A). Instead of the Alberti bass, the simple paradigm used standard and deviant notes presented with a constant SOA of 400 ms, corresponding to a classical oddball paradigm. This was a multi-feature paradigm in the sense that it incorporated all seven deviant types in each block; however, the number of standards presented before each deviant varied randomly between 3 and 5, again to minimize spill-over effects from consecutive deviants with the same constituent deviant types (Figure 1A).

During the complete experimental procedure (∼100 mins), participants were seated in a magnetically shielded room watching a silent movie with the soundtrack muted. Stimulus sounds were presented binaurally through Etymotic ER2 insert earphones using the Presentation software (Neurobehavioral Systems, San Francisco, USA). Sound pressure level was set to 50 dB above individual hearing threshold as determined by a staircase procedure implemented in PsychoPy (Peirce, 2007). Participants were instructed to stay still while the sounds were playing, to ignore them, and to focus on the movie. Complex musical multi-feature blocks (M1–M4) lasted 13 mins 47 sec whereas simple control blocks (C1–C3) lasted ∼11 mins. Between each block, short breaks of ∼1–2 mins were provided during which participants could stretch and move slightly while staying seated. Prior to the lab session, participants completed an online questionnaire that ensured eligibility and assessed their level of musical experience using Goldsmiths Musical Sophistication Index, v.1.0 (Müllensiefen et al., 2014).

Magnetoencephalography recording and pre-processing

MEG data were sampled at 1,000 Hz (with a passband of 0.03–330 Hz) using the Elekta Neuromag TRIUX system hosted by the MINDLab Core Experimental Facility at Aarhus University Hospital. This MEG system contains 102 magnetometers and 204 planar gradiometers. Head position was recorded continuously throughout the experimental session using four head position indicator coils (cHPI). Additionally, vertical and horizontal EOG as well as ECG recordings were obtained using bipolar surface electrodes positioned above and below the right eye, at the outer canthi of both eyes, and on the left clavicle and right rib.

Data were pre-processed using the temporal extension of the signal space separation (tSSS) technique (Taulu et al., 2004; Taulu and Simola, 2006) implemented in Elekta’s MaxFilter software (Version 2.2.15). This included head movement compensation using cHPI, removing noise from electromagnetic sources outside the head, and down-sampling by a factor of 4–250 Hz. EOG and ECG artifacts were removed with independent component analysis (ICA) using the find_bads_eog and find_bads_ecg algorithms in MNE Python (Gramfort et al., 2013, 2014). These algorithms detect artifactual components based on either the Pearson correlation between the identified ICA components and the EOG/ECG channels (for EOG) or the significance value from a Kuiper’s test using cross-trial phase statistics (Dammers et al., 2008; for ECG). Topographies and averaged epochs for the EOG- and ECG-related components were visually inspected for all participants to ensure the validity of rejected components.

Further pre-processing and statistical analysis was performed in FieldTrip (Oostenveld et al., 20111; RRID:SCR_004849). Data were epoched into trials of 500 ms duration including a 100 ms pre-stimulus interval. As indicated in Figure 1A, for both paradigms, only the third occurrence of the standard tone after each deviant was included as a standard trial. Trials containing SQUID jumps were discarded using automatic artifact rejection with a z-value cutoff of 30. The remaining 98.2% of trials on average (ranging from 91.0 to 99.8% for individual participants) were band-pass filtered at 1–40 Hz using a two-pass Butterworth filter (data-padded to 3 s to avoid filter artifacts). Planar gradiometer pairs were combined by taking the root-mean-square of the two gradients at each sensor, resulting in a single positive value. Baseline correction was performed based on the 50 ms pre-stimulus interval.

Experimental design and statistical analysis

To reiterate the experimental design, 25 musicians and 25 non-musicians completed a total of seven blocks distributed between the complex musical multi-feature paradigm (M) and simple control paradigm (C) in the order M1–C1–M2–C2–M3–C3–M4 (Figure 1A). Trials were averaged for each condition (i.e., one standard and seven deviant types) separately for each participant and separately for each paradigm. Magnetic mismatch negativity responses (MMNm) were computed by subtracting the same average standard (the third after each deviant, cf. Figure 1A) originating from the relevant paradigm from each of the deviant responses. These were the empirical MMNms. Consistent with previous studies (Levänen et al., 1993; Schröger, 1995, 1996; Takegata et al., 1999; Paavilainen et al., 2001, 2003b; Ylinen et al., 2005; Caclin et al., 2006; Lidji et al., 2009; Choi et al., 2017; Yu et al., 2022), modeled MMNms were computed for the three double deviants and one triple deviant by adding the two or three empirical MMNms obtained from the relevant single deviants (Figure 1B).

The statistical analysis reported here focused on data from the combined planar gradiometers. Our focus was on the modulation of the neural signals as indexed by the MMNm response. Seeing that previous research has shown little to no added benefit in signal-to-noise ratio for source-derived estimates of the MMN(m) response, likely because the tangentially-oriented MMNm sources are optimally detected by the planar gradiometers (Tervaniemi et al., 2005; Recasens and Uhlhaas, 2017), we deemed that our research question could be fully addressed in sensor space without making further assumptions as required for source reconstruction. Non-parametric, cluster-based permutation statistics (Maris and Oostenveld, 2007) were used to test the prediction that the additivity of the MMNm response would differ between musicians and non-musicians. Because violation of the additive model is assessed through demonstration of a difference between empirical and modeled MMNm responses, this hypothesis predicts an interaction effect between degree of additivity and musical expertise. This was tested by running cluster-based permutation tests on the difference between empirical and modeled MMNm responses, comparing between musicians and non-musicians. This approach for testing interaction effects within the non-parametric permutation framework is advocated by FieldTrip.2

The test statistic was computed in the following way: Independent-samples t-statistics were computed for all timepoint-by-sensor samples. Samples that were neighbors in time and/or space and exceeded the pre-determined alpha level of 0.05 were included in the same cluster. Despite its similarity to uncorrected mass-univariate testing, this step does not represent hypothesis testing, but only serves the purpose of cluster definition. To this end, a neighborhood structure was generated that defined which sensors were considered neighbors based on the linear distance between sensors. Only assemblies containing minimum two neighbor sensors were regarded as clusters. A cluster-level statistic was computed by summing the t-statistics within each cluster and taking the maximum value of these summed t-statistics. This process was subsequently repeated for 10,000 random permutations of the group labels (musicians vs. non-musicians) giving rise to a Monte Carlo approximation of the null distribution. The final p-value resulted from comparing the initial test statistic with this distribution. Bonferroni-correction was applied to the two-sided alpha level to correct for the four comparisons of modeled and empirical double and triple MMNms, resulting in an alpha level of α = 0.025/4 = 0.00625. When significant additivity-by-expertise interactions were discovered, simple effects of additivity were assessed for musicians and non-musicians separately.

Previous research shows that the MMNm usually peaks in the 150–250 ms post-stimulus time range and is maximally detected with gradiometers bilaterally at supratemporal sites (Levänen et al., 1996; Näätänen et al., 2007). This was confirmed for the current dataset by computing the grand-average MMNm across all participants, all conditions, and both paradigms. This grand-average MMNm peaked in the combined gradiometer sensors MEG1342 + 1343 (right) and MEG0232 + 0233 (left) at ∼156 ms post-stimulus (with secondary peaks extending into the 200–300 ms range). However, to account for possible differences in peak latency and source location between participants and between the various deviant types, the analysis was extended to the 100–300 ms post-stimulus interval and to also include the eight neighboring sensors around the peak sensor in each hemisphere (Figure 1C). The final 2 × 9 sensors were located approximately over the superior temporal lobe in each hemisphere. In this way, prior knowledge was incorporated to increase the sensitivity of the statistical tests without compromising their validity (Maris and Oostenveld, 2007). In that respect, it should be noted that the cluster-based permutation framework does not allow inferences on the onset or offset of specific effects nor of their exact spatial distribution (Sassenhagen and Draschkow, 2019).

Before the main analysis, however, two sets of tests were conducted to ensure the validity of the main analysis. To this end, cluster-based permutation tests were run using the parameters, sensors, and time window specified above (except for the permuted labels, which were changed according to the contrast of interest). First, the difference between standard and deviant responses was assessed to determine whether MMNm effects were present, i.e., whether standard and deviant responses differed significantly in the 100–300 ms post-stimulus range. These analyses were carried out for each deviant type separately for musicians and non-musicians and separately for the two paradigms. Second, to establish that the possible MMNm effects were potentially additive, nine pairwise comparisons were conducted between relevant single and double deviants as well as between relevant double and triple deviants (i.e., FI-FIL, IL-FIL, LF-FIL, F-FI, I-FI, I-IL, L-IL, L-LF, F-LF). This was done across all participants regardless of expertise level. No Bonferroni-correction was applied to these secondary validity checks.

Results

To anticipate our results, significantly greater MMNm subadditivity was found in musicians compared to non-musicians for frequency-related features in the complex musical multi-feature paradigm. These expertise effects were absent from the simple control paradigm. Further details are reported separately for the two paradigms below.

Complex musical multi-feature paradigm

In the main experimental paradigm—the complex musical multi-feature paradigm—all seven single, double, and triple deviants elicited significantly larger responses than the standards for musicians as well as for non-musicians (Table 2), indicative of significant MMNm responses in all conditions of this paradigm. Moreover, the triple deviant resulted in a significantly larger MMNm than each of the double deviants (Table 3). Except for a single comparison between the location deviant and the double deviant combining location and frequency, MMNms to all double deviants were significantly larger than MMNms to single deviants, indicating that the addition of an extra feature significantly increased the MMNm amplitude.

TABLE 2
www.frontiersin.org

Table 2. Magnetic mismatch negativity response (MMNm).

TABLE 3
www.frontiersin.org

Table 3. Potential additivity of the magnetic mismatch negativity response (MMNm) response.

The non-significant LF vs. L comparison as well as the somewhat larger p-values for the comparisons between responses elicited with and without the frequency component (i.e., FIL vs. IL, FI vs. I) already indicated that a certain extent of subadditivity was present specifically for the frequency component. This is also suggested by the dissimilarity between modeled and empirical responses in the first, third, and fourth rows of Figure 2. The subsequent main analysis addressed how this possible effect interacted with musical expertise (see Supplementary Figure 1 for depictions of modeled MMNm responses together with their constituent single-deviant MMNms).

FIGURE 2
www.frontiersin.org

Figure 2. Modeled and empirical mismatch negativity (MMNm) in musicians and non-musicians in the complex musical multi-feature paradigm in response to deviants differing in either frequency and intensity (FI), intensity and location (IL), location and frequency (LF), or frequency and intensity and location (FIL). Modeled MMNms (dashed gray lines) correspond to the sum of the MMNms to two or three single deviants whereas empirical MMNms (solid black lines) correspond to MMNms obtained for double and triple deviants. Comparing the plots for modeled and empirical MMNms, significantly greater subadditivity was evident in musicians compared to non-musicians specifically in the double deviants involving frequency (i.e., FI and LF). This expertise-by-additivity effect was marginally non-significant for the triple deviant (i.e., FIL). The event-related field (ERF) plots depict the mean of the data from the peak sensor and eight surrounding sensors in each hemisphere [the peak gradiometer pairs were MEG1342 + 1343 (Right) and MEG0242 + 0243 (Left)]. Low-pass filtering at 20 Hz was applied for visualization purposes only. The topographical distributions depict the 100–300 ms post-stimulus time range for the difference waves between standard and deviant responses, and p-values outside the boxes reflect simple effects of additivity separately for musicians and non-musicians.

Indeed, the differences between modeled and empirical responses were significantly greater in musicians than in non-musicians for the FI and LF deviants (Table 4). This demonstrated subadditivity was consistent with the integrated processing hypothesis as formulated above. For the triple deviant (FIL), this interaction effect approached significance whereas it was absent for the IL deviant, which did not include the frequency component. Consistent with this picture, follow-up analyses of the significant interactions found simple additivity effects only for musicians (Table 4 and Figure 2).

TABLE 4
www.frontiersin.org

Table 4. Subadditivity of the magnetic mismatch negativity response (MMNm) response.

Simple control paradigm

In the simple control paradigm, significant MMNm effects were also present for all deviant types (Table 2). Here, additivity was more uniformly present in terms of significant differences between all single and double deviants as well as between the double deviants and the triple deviant (Table 3, see Supplementary Figure 2 for depictions of modeled MMNm responses together with their constituent single-deviant MMNms). The main analysis showed no differences in additivity between the two groups and across the various deviant types (Table 4). This pattern also emerges from the plots of event-related fields and scalp topographies in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3. Modeled and empirical mismatch negativity (MMNm) in musicians and non-musicians in the simple control paradigm in response to double deviants differing in either frequency and intensity (FI), intensity and location (IL), or location and frequency (LF), as well as to triple deviants differing in frequency and intensity and location (FIL). Modeled MMNms (dashed gray lines) correspond to the sum of MMNms to two or three single deviants whereas empirical MMNms (solid black lines) correspond to the actual MMNms obtained with double and triple deviants. In contrast to the musical multi-feature paradigm, there were no significant differences in additivity between musicians and non-musicians when comparing the modeled and empirical MMNms. The event-related field (ERF) plots depict the mean of the data from the peak sensor and eight surrounding sensors in each hemisphere [the peak gradiometer pairs were MEG1342 + 1343 (Right) and MEG0242 + 0243 (Left)]. Low-pass filtering at 20 Hz was applied for visualization purposes only. The topographical distributions depict the 100–300 ms post-stimulus time range for the difference waves between standard and deviant responses, and p-values reflect simple effects of additivity separately for musicians and non-musicians.

Discussion

Expertise as integrated feature processing

The current results provide the first neurophysiological evidence consistent with more integrated processing of multiple auditory feature deviants in musical experts compared to non-experts. In a complex musical paradigm, magnetic mismatch negativity responses (MMNm) elicited by double deviants differing in multiple features (including frequency) were lower than the sum of MMNm responses to the constituent single-feature deviants. This selective subadditivity for the musically relevant frequency feature was either absent or present to a smaller extent in non-musicians, despite significant MMNms in all conditions. In addition to the main finding of expertise-related differences in frequency-related subadditivity, the pattern of results between the two paradigms suggests that these differences were context-specific. Specifically, the expertise differences were absent when using identical sounds in a simpler and less musically-relevant configuration as a control paradigm.

Following the additive model used extensively in neurophysiological studies of multisensory interactions (Besle et al., 2009) and adapted for studies of unimodal feature integration (e.g., Levänen et al., 1993; Paavilainen et al., 2001), we interpret these findings as support for the integrated processing hypothesis by which musical expertise is associated with enhanced processing by recruiting shared neural resources for more complex representations of domain-relevant stimuli. Importantly, our cross-sectional study design does not allow us to determine whether musical training causally induces more holistic processing or whether pre-existing integration in processing benefits musicianship. A causal interpretation, however, would contrast with formulations of feature integration theory regarding feature processing as innate or acquired through normal neurodevelopment and therefore largely immutable (Treisman and Gelade, 1980; Quinlan, 2003). Findings that perceptual learning modulates attention, thus determining whether specific features are considered for binding (Colzato et al., 2006), have already questioned this view. Indeed, Neuhaus and Knösche (2008) demonstrated more integrative pitch and rhythm processing in musicians than non-musicians manifested in expertise-related differences in the P1 and P2 components. By extending these findings to intensity and location and relating them to MMNm responses, we consolidate the association of musical training with changes in predictive neural processing (Skoe and Kraus, 2012).

Evidence for expertise-related effects on multimodal integration of audiovisual and audiomotor stimuli has steadily accumulated (Paraskevopoulos et al., 2012, 2014; Paraskevopoulos and Herholz, 2013; Bishop and Goebl, 2014; Proverbio et al., 2014, 2015; Pantev et al., 2015; Møller et al., 2018; Sorati and Behne, 2020; Wilbiks and O’Brien, 2020). While integration across modalities appears less prominent in experts who may be better at segregating auditory features from audiovisual compounds (Møller et al., 2018, 2021), our study suggests that feature integration within the relevant modality may simultaneously increase with training.

Consistent with a causal interpretation of our data in the context of the integrated processing hypothesis, decreased neural activity is observed for perceptual learning in audition (Jäncke et al., 2001; Berkowitz and Ansari, 2010; Zatorre et al., 2012) and vision (Schiltz et al., 1999; van Turennout et al., 2000; Kourtzi et al., 2005; Yotsumoto et al., 2008). These changes are sometimes associated with enhanced effective connectivity between distinct cortical regions (Büchel et al., 1999). Enhanced connectivity in auditory areas may also enable musicians to rely more heavily on auditory than visual information in audiovisual tasks (Paraskevopoulos et al., 2015; Møller et al., 2018). This seems adaptive assuming that auditory information is better integrated and thus potentially more informative and relevant to musicians. Correspondingly, greater cortical thickness correlation between visual and auditory areas have been found in non-musicians compared to musicians (Møller et al., 2021).

Potentially more integrative auditory processing may indeed benefit musicians behaviorally. For instance, this may manifest in terms of enhanced verbal and visual memory (Chan et al., 1998; Jakobson et al., 2008). This would be consistent with feature-based theories of visual short-term memory positing that integrative processing allows experts to incorporate multiple features into object representations, thus improving discrimination of highly similar exemplars (Curby et al., 2009).

Frequency-related feature selectivity

Whereas MMNm subadditivity was present for double and triple deviants comprising frequency, intensity, and location, expertise only interacted with additivity for frequency-related deviants. This aligns well with the different levels of musical relevance embodied by these features, with frequency, intensity, and location representing decreasingly common parameters of syntactic organization in music. More specifically, frequency is usually predetermined by composers in terms of unambiguous notation of intended pitch categories, the spontaneous changing of which would produce distinctly different melodies. Intensity is subject to more flexible expressive performance decisions (Palmer, 1996; Widmer and Goebl, 2004), and changing it would usually not compromise syntax or melodic identity. Sound source localization is relatively unimportant in the professional lives of musicians and non-musicians, with the possible exception of orchestral and choral conductors whose more advanced sound localization skills are evident from specialized neuroplasticity (Münte et al., 2001; Nager et al., 2003). Given that our musician group, however, only comprised a single individual with conducting experience, we can assume that all members of this group had intensive pitch-related training, that most members regularly used intensity as an expressive means, but that sound source localization was overall of secondary importance.

The prominence of pitch over intensity and location may relate to the diverging potential for learning effects to emerge within these features. For example, while learning of frequency discrimination is ubiquitous (Kishon-Rabin et al., 2001), adaptation to altered sound-localization cues is only partial in the horizontal plane (Hofman et al., 2002; Wright and Zhang, 2006). Indeed, recognizing and producing pitch accurately is rehearsed intensively in musical practice (Besson et al., 2007), and voice pedagogues regard pitch intonation as the most important factor in determining singing talent (Watts et al., 2002). Even beyond musical contexts, musicians up-weight pitch compared to duration when determining linguistic focus in speech (Symons and Tierney, 2021).

One possible interpretation following from the present study is that musical learning may involve changes in feature integration which, accordingly, would manifest most clearly in the musically relevant and perceptually plastic pitch-related domain. Indeed, musical pitch information has been shown to be integrated with intensity (Grau and Kemler-Nelson, 1988; McBeath and Neuhoff, 2002) and timbre (Krumhansl and Iverson, 1992; Hall et al., 2000; Lidji et al., 2009; Allen and Oxenham, 2014), at least in parietal cortex (Sohoglu et al., 2020), manifesting as nearly complete spatial overlap in neural processing of frequency and timbre at the inter-subject level (Allen et al., 2017). More broadly, the three features included here exemplify both the “what” (frequency, intensity) and “where” (location, intensity) streams of auditory processing, which have been found to follow different patterns of feature integration (Delogu et al., 2014).3 By mostly recruiting participants with uniform expertise levels, many previous studies have avoided systematic group comparisons capable of demonstrating expertise-related differences in feature integration.

In sum, although frequency, intensity, and location all have relevance in music, frequency represents the most salient cue for identifying and distinguishing musical pieces and their component themes and motifs. Therefore, when musicians show selectively affected frequency processing, we infer this to be more likely due to their intensive training than to innate predispositions (Hyde et al., 2009; Herdener et al., 2010). Even if aspects of innately enhanced frequency-specific processing are present, they are likely to require reinforcement through intensive training, possibly motivated by greater success experiences as a young musician. While these learning effects may transfer to prosodic production and decoding (Besson et al., 2007; Pastuszek-Lipińska, 2008; Lima and Castro, 2011), they are of limited use outside musical contexts where sudden changes in intensity or location are usually more likely than frequency to constitute environmentally relevant cues.

Context-dependency of expert feature processing

It is plausible that because superior frequency processing is merely adaptive in relevant contexts, musicians showed greater frequency-related feature integration for more complex, musically relevant stimuli. This corresponds with stronger MMNm amplitudes, shorter MMNm latencies, and more widespread cortical activation observed in musically complex compared to simple oddball paradigms (Lappe et al., 2016). Crucially, “complexity” in this regard refers to degrees of musical relevance (i.e., ecological validity) whereas increasing the statistical complexity of the context in which deviants are presented—for example, in terms of entropy—attenuates mismatch responses for both musicians and non-musicians (Quiroga-Martinez et al., 2020). Similarly, greater MMNm amplitudes and shorter latencies typically emerge for spectrally rich tones (e.g., realistic piano tones) compared to pure tones that only rarely occur naturally (Tervaniemi et al., 2000). Musical expertise, moreover, produces greater MMNm amplitudes (Fujioka et al., 2004) and more widespread activation (Pantev et al., 1998) for complex than for pure tones. Enhanced processing of audiovisual asynchronies in musicians is also restricted to music-related tasks, remaining absent for language (Lee and Noppeney, 2014). Research in other expertise domains has found comparable levels of domain-specificity (de Groot, 1978; Ericsson and Towne, 2010).

Consistent with the context-specificity demonstrated here, we have previously found in behavioral experiments that musical stimuli activate more specific predictions in musicians and thus evoke stronger reactions when expectations are violated (Hansen and Pearce, 2014; Hansen et al., 2016). In what may first appear to diverge from the current results, this previous work showed that expertise effects were sometimes only detectable in stylistically simple conditions (Hansen and Pearce, 2014). The notion of complexity does, however, not directly translate to the current study in that the simple melodies used by Hansen and Pearce (2014) were always substantially more complex than the complex stimuli in the current MEG paradigm. The psychological processes under investigation were also much higher level (i.e., conscious expectations for discrete pitches within a scale vs. pre-attentive, early neural responses to mistuned pitches) and thus would have relied on (at least partially) distinct neural mechanisms. Because the expertise-by-complexity interaction already disappeared in a more implicit task where listener uncertainty was inferred from probe-tone ratings for multiple melody continuations, we would also not expect this effect to transfer to early neural responses.

Observations that pitch and duration in melodies are integrated when these dimensions co-vary—but not when they contrast—and that such integration increases with exposure (Boltz, 1999) suggest an interpretation of our results where musicians are predisposed to capitalize more optimally upon contextual cues. Note, though, that the two present paradigms did not only differ on musical relevance, but also on stimulus onset asynchrony, predictability of deviant location, and general textural complexity. Although these differences between the two paradigms do not compromise our main finding of expertise-related subadditivity for frequency-related features, future research should try to disentangle the individual contribution of these factors.

Future research and unresolved issues

Some clinical studies have indeed measured neural responses to multiple co-occurring auditory feature deviants (e.g., Hay et al., 2015) and have found, for example, that MMNs from double deviants combining duration and frequency–but not their constitutive single-deviant MMNs–predict the time to psychosis onset in individuals at high risk for schizophrenia (Perez et al., 2014; Hamilton et al., 2022). These studies have, however, not typically assessed MMN(m) additivity per se. Thus, there may be an underused potential for extending the methods and results applied and developed here to relevant clinical populations in future work.

It is worth noting that the increased frequency-related feature integration in musicians observed here seems inconsistent with other behavioral and neuroimaging results. For example, a behavioral study found that interference of pitch and timbre was unaffected by expertise (Allen and Oxenham, 2014). Conversely, findings that pitch and consonants in speech produce sub-additive MMN responses (Gao et al., 2012), but do not interact behaviorally (Kolinsky et al., 2009), suggest that measures of behavioral and neural integration do not always correspond (Musacchia et al., 2007). Additionally, cross-modal integration often results in increased rather than decreased neural activity (Stein, 2012). For instance, trumpeters’ responses to sounds and tactile lip stimulation exceed the sum of constituent unimodal responses (Schulz et al., 2003). When reading musical notation, musicians dissociate processing of spatial pitch in the dorsal visual stream from temporal movement preparation in the ventral visual stream (Bengtsson and Ullén, 2006) and show no neurophysiological or behavioral interference of frequency and duration (Schön and Besson, 2002). This would suggest greater feature independence in experts.

The Reverse Hierarchy Theory of perceptual learning supports this view by asserting that expertise entails more segregated representations of ecologically relevant stimuli, thus providing access to more detailed lower-level representations (Ahissar and Hochstein, 2004; Ahissar et al., 2009). This has some bearing on musical intuitions. Specifically, segregated processing would presumably better enable musicians to capitalize upon the covariance of acoustic features, as demonstrated for tonal and metrical hierarchies (Prince and Schmuckler, 2014). Phenomenal accents in music are indeed intensified by coinciding cues (Lerdahl and Jackendoff, 1983; Hansen, 2011), which may underlie musicians’ superiority in decoding emotion from speech (Lima and Castro, 2011) and segmenting speech (François et al., 2013, 2014) and music (Deliège, 1987; Peebles, 2011, 2012). Given that individuals with better frequency discrimination thresholds more capably segregate auditory features from audiovisual stimuli (Møller et al., 2018), it remains unknown whether musicians suppress frequency-related feature integration whenever constituent features are irrelevant to the task at hand. The interaction of expertise with frequency-related feature-selectivity and context-specificity observed here contributes crucial empirical data to ongoing discussions of seemingly diverging findings pertaining to integrated and independent musical feature processing (Palmer and Krumhansl, 1987; Boltz, 1999; Waters and Underwood, 1999; Tillmann and Lebrun-Guillaud, 2006).

Data availability statement

To the extent that local ethical and legal regulations allow, the data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving human participants were reviewed and approved by The Central Denmark Regional Committee on Health Research Ethics (case 1-10-72-11-15). The participants provided their written informed consent to participate in this study.

Author contributions

NH: conceptualization, methodology, software, validation, formal analysis, investigation, data curation, writing—original draft and review and editing, visualization, project administration, and funding acquisition. AH: conceptualization, methodology, software, validation, formal analysis, investigation, writing—review and editing, and supervision. CM: conceptualization, methodology, investigation, and writing—review and editing. MP: conceptualization, writing—review and editing, and supervision. PV: conceptualization, resources, writing—review and editing, supervision, and funding acquisition. All authors contributed to the article and approved the submitted version.

Funding

This work received funding from the Ministry of Culture Denmark; Royal Academy of Music Aarhus/Aalborg; Graduate School Arts Aarhus University; Ministry of Higher Education and Science Denmark (EliteForsk Travel Grant 3071-00008B); European Union’s Horizon 2020 Research and Innovation Programme under the Marie Skłodowska-Curie grant agreement No 754513; Aarhus University Research Foundation; and Danish National Research Foundation (DNRF117). Funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Acknowledgments

We would like to thank Rebeka Bodak, Kira Vibe Jespersen, and the staff at MINDLab Core Experimental Facility, Aarhus University Hospital, for help with data collection and technical assistance. Center for Music in the Brain is funded by the Danish National Research Foundation (DNRF117).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnins.2022.907540/full#supplementary-material

Footnotes

  1. ^ http://www.ru.nl/neuroimaging/fieldtrip
  2. ^ http://www.fieldtriptoolbox.org/faq/how_can_i_test_an_interaction_effect_using_cluster-based_permutation_tests
  3. ^ The coupling of frequency and intensity cues for determining the identity of auditory objects may seem obvious. With regards to the coupling of intensity and location cues, research by Hansen and Huron (2019), for example, demonstrates how manipulating intensity can alter perception of location and motion in physical space.

References

Ahissar, M., and Hochstein, S. (2004). The reverse hierarchy theory of visual perceptual learning. Trends Cogn. Sci. 8, 457–464. doi: 10.1016/j.tics.2004.08.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Ahissar, M., Nahum, M., Nelken, I., and Hochstein, S. (2009). Reverse hierarchies and sensory learning. Philos. Trans. R. Soc. Lond. B Biol. Sci. 364, 285–299. doi: 10.1098/rstb.2008.0253

PubMed Abstract | CrossRef Full Text | Google Scholar

Allen, E. J., and Oxenham, A. J. (2014). Symmetric interactions and interference between pitch and timbre. J. Acoust. Soc. Am. 135, 1371–1379. doi: 10.1121/1.4863269

CrossRef Full Text | Google Scholar

Allen, E. J., Burton, P. C., Olman, C. A., and Oxenham, A. J. (2017). Representations of pitch and timbre variation in human auditory cortex. J. Neurosci. 37:1284. doi: 10.1523/JNEUROSCI.2336-16.2016

PubMed Abstract | CrossRef Full Text | Google Scholar

Bengtsson, S. L., and Ullén, F. (2006). Dissociation between melodic and rhythmic processing during piano performance from musical scores. Neuroimage 30, 272–284. doi: 10.1016/j.neuroimage.2005.09.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Berkowitz, A. L., and Ansari, D. (2010). Expertise-related deactivation of the right temporoparietal junction during musical improvisation. Neuroimage 49, 712–719. doi: 10.1016/j.neuroimage.2009.08.042

PubMed Abstract | CrossRef Full Text | Google Scholar

Besle, J., Bertrand, O., and Giard, M. H. (2009). Electrophysiological (EEG, sEEG, MEG) evidence for multiple audiovisual interactions in the human auditory cortex. Hear. Res. 258, 143–151. doi: 10.1016/j.heares.2009.06.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Besson, M., Schön, D., Moreno, S., Santos, A., and Magne, C. (2007). Influence of musical expertise and musical training on pitch processing in music and language. Restor. Neurol. Neurosci. 25, 399–410.

Google Scholar

Bishop, L., and Goebl, W. (2014). Context-specific effects of musical expertise on audiovisual integration. Front. Psychol. 5:1123. doi: 10.3389/fpsyg.2014.01123

PubMed Abstract | CrossRef Full Text | Google Scholar

Boltz, M. G. (1999). The processing of melodic and temporal information: Independent or unified dimensions? J. New Music Res. 28, 67–79. doi: 10.1076/jnmr.28.1.67.3121

CrossRef Full Text | Google Scholar

Brattico, E., Pallesen, K. J., Varyagina, O., Bailey, C., Anourova, I., Järvenpää, M., et al. (2009). Neural discrimination of nonprototypical chords in music experts and laymen: An MEG study. J. Cogn. Neurosci. 21, 2230–2244. doi: 10.1162/jocn.2008.21144

PubMed Abstract | CrossRef Full Text | Google Scholar

Büchel, C., Coull, J. T., and Friston, K. J. (1999). The predictive value of changes in effective connectivity for human learning. Science 283, 1538–1541. doi: 10.1126/science.283.5407.1538

PubMed Abstract | CrossRef Full Text | Google Scholar

Caclin, A., Brattico, E., Tervaniemi, M., Näätänen, R., Morlet, D., Giard, M.-H., et al. (2006). Separate neural processing of timbre dimensions in auditory sensory memory. J. Cogn. Neurosci. 18, 1959–1972. doi: 10.1162/jocn.2006.18.12.1959

PubMed Abstract | CrossRef Full Text | Google Scholar

Campo, P., Poch, C., Parmentier, F. B. R., Moratti, S., Elsley, J. V., Castellanos, N. P., et al. (2010). Oscillatory activity in prefrontal and posterior regions during implicit letter-location binding. Neuroimage 49, 2807–2815. doi: 10.1016/j.neuroimage.2009.10.024

PubMed Abstract | CrossRef Full Text | Google Scholar

Chan, A. S., Ho, Y. C., and Cheung, M. C. (1998). Music training improves verbal memory. Nature 396:128. doi: 10.1038/24075

PubMed Abstract | CrossRef Full Text | Google Scholar

Choi, W., Tong, X., Gu, F., Tong, X., and Wong, L. (2017). On the early neural perceptual integrality of tones and vowels. J. Neurolinguist. 41, 11–23. doi: 10.1016/j.jneuroling.2016.09.003

CrossRef Full Text | Google Scholar

Colzato, L. S., Raffone, A., and Hommel, B. (2006). What do we learn from binding features? Evidence for multilevel feature integration. J. Exp. Psychol. Hum. Percept. Perform. 32:705. doi: 10.1037/0096-1523.32.3.705

PubMed Abstract | CrossRef Full Text | Google Scholar

Curby, K. M., Glazek, K., and Gauthier, I. (2009). A visual short-term memory advantage for objects of expertise. J. Exp. Psychol. Hum. Percept. Perform. 35:94. doi: 10.1037/0096-1523.35.1.94

PubMed Abstract | CrossRef Full Text | Google Scholar

Czigler, I., and Winkler, I. (1996). Preattentive auditory change detection relies on unitary sensory memory representations. Neuroreport 7, 2413–2418. doi: 10.1097/00001756-199611040-00002

PubMed Abstract | CrossRef Full Text | Google Scholar

Dammers, J., Schiek, M., Boers, F., Silex, C., Zvyagintsev, M., Pietrzyk, U., et al. (2008). Integration of amplitude and phase statistics for complete artifact removal in independent components of neuromagnetic recordings. IEEE Trans. Biomed. Eng. 55, 2353–2362. doi: 10.1109/TBME.2008.926677

PubMed Abstract | CrossRef Full Text | Google Scholar

de Groot, A. D. (1978). Thought And Choice In Chess. The Hague: Mouton Publishers.

Google Scholar

Deliège, I. (1987). Grouping conditions in listening to music: An approach to Lerdahl Jackendoff’s grouping preference rules. Music Percept. 4:325. doi: 10.2307/40285378

CrossRef Full Text | Google Scholar

Delogu, F., Gravina, M., Nijboer, T., and Postma, A. (2014). Binding “what” and “where” in auditory working memory: An asymmetrical association between sound identity and sound location. J. Cogn. Psychol. 26, 788–798. doi: 10.1080/20445911.2014.959448

CrossRef Full Text | Google Scholar

Di Lollo, V. (2012). The feature-binding problem is an ill-posed problem. Trends Cogn. Sci. 16, 317–321. doi: 10.1016/j.tics.2012.04.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Ericsson, K. A. (2006). “The influence of experience and deliberate practice on the development of superior expert performance,” in The Cambridge Handbook of Expertise and Expert Performance, eds K. A. Ericsson, P. J. Feltovich, and R. R. Hoffman (Cambridge, UK: Cambridge University Press), 683–703. doi: 10.1111/j.1365-2923.2007.02946.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Ericsson, K. A., and Towne, T. J. (2010). Expertise. Wires Cogn. Sci. 1, 404–416. doi: 10.1002/wcs.47

PubMed Abstract | CrossRef Full Text | Google Scholar

Fay, R. R., and Popper, A. N. (2000). Evolution of hearing in vertebrates: The inner ears and processing. Hear. Res. 149, 1–10. doi: 10.1016/S0378-5955(00)00168-4

CrossRef Full Text | Google Scholar

François, C., Chobert, J., Besson, M., and Schön, D. (2013). Music training for the development of speech segmentation. Cereb. Cortex 23, 2038–2043. doi: 10.1093/cercor/bhs180

PubMed Abstract | CrossRef Full Text | Google Scholar

François, C., Jaillet, F., Takerkart, S., and Schön, D. (2014). Faster sound stream segmentation in musicians than in nonmusicians. PLoS One 9:e101340. doi: 10.1371/journal.pone.0101340

PubMed Abstract | CrossRef Full Text | Google Scholar

Fujioka, T., Trainor, L. J., Ross, B., Kakigi, R., and Pantev, C. (2004). Musical training enhances automatic encoding of melodic contour and interval structure. J. Cogn. Neurosci. 16, 1010–1021. doi: 10.1162/0898929041502706

PubMed Abstract | CrossRef Full Text | Google Scholar

Fuller, D. (2015). “Alberti bass,” in Grove Music Online, ed. L. Macy (Oxford: Oxford University Press).

Google Scholar

Gao, S., Hu, J., Gong, D., Chen, S., Kendrick, K. M., and Yao, D. (2012). Integration of consonant and pitch processing as revealed by the absence of additivity in mismatch negativity. PLoS One 7:e38289. doi: 10.1371/journal.pone.0038289

PubMed Abstract | CrossRef Full Text | Google Scholar

Giard, M. H., Lavikahen, J., Reinikainen, K., Perrin, F., Bertrand, O., Pernier, J., et al. (1995). Separate representation of stimulus frequency, intensity, and duration in auditory sensory memory: An event-related potential and dipole-model analysis. J. Cogn. Neurosci. 7, 133–143. doi: 10.1162/jocn.1995.7.2.133

PubMed Abstract | CrossRef Full Text | Google Scholar

Gomes, H., Bernstein, R., Ritter, W., Vaughan, H. G., and Miller, J. (1997). Storage of feature conjunction in transient auditor memory. Psychophysiology 34, 712–716. doi: 10.1111/j.1469-8986.1997.tb02146.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Gramfort, A., Luessi, M., Larson, E., Engemann, D. A., Strohmeier, D., Brodbeck, C., et al. (2013). MEG and EEG data analysis with MNE-Python. Front. Neurosci. 7:267. doi: 10.3389/fnins.2013.00267

PubMed Abstract | CrossRef Full Text | Google Scholar

Gramfort, A., Luessi, M., Larson, E., Engemann, D. A., Strohmeier, D., Brodbeck, C., et al. (2014). MNE software for processing MEG and EEG data. Neuroimage 86, 446–460. doi: 10.1016/j.neuroimage.2013.10.027

PubMed Abstract | CrossRef Full Text | Google Scholar

Grau, J. A., and Kemler-Nelson, D. G. (1988). The distinction between integral and separable dimensions: Evidence for the integrality of pitch and loudness. J. Exp. Psychol. Gen. 117, 347–370. doi: 10.1037/0096-3445.117.4.347

PubMed Abstract | CrossRef Full Text | Google Scholar

Griffiths, T., and Warren, J. (2004). What is an auditory object? Nat. Rev. Neurosci. 5, 887–892. doi: 10.1038/nrn1538

PubMed Abstract | CrossRef Full Text | Google Scholar

Grill-Spector, K., and Weiner, K. S. (2014). The functional architecture of the ventral temporal cortex and its role in categorization. Nat. Rev. Neurosci. 15, 536–548. doi: 10.1038/nrn3747

PubMed Abstract | CrossRef Full Text | Google Scholar

Guérard, K., Morey, C. C., Lagacé, S., and Tremblay, S. (2013). Asymmetric binding in serial memory for verbal and spatial information. Mem. Cogn. 41, 378–391. doi: 10.3758/s13421-012-0275-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Hall, M. D., Pastore, R. E., Acker, B. E., and Huang, W. (2000). Evidence for auditory feature integration with spatially distributed items. Percept. Psychophys. 62, 1243–1257. doi: 10.3758/BF03212126

PubMed Abstract | CrossRef Full Text | Google Scholar

Hamilton, H. K., Roach, B. J., Bachman, P. M., Belger, A., Carrión, R. E., Duncan, E., et al. (2022). Mismatch negativity in response to auditory deviance and risk for future psychosis in youth at clinical high risk for psychosis. JAMA Psychiatry 79, 780–789. doi: 10.1001/jamapsychiatry.2022.1417

PubMed Abstract | CrossRef Full Text | Google Scholar

Hansen, N. C. (2011). The legacy of Lerdahl and Jackendoff’s A Generative Theory of Tonal Music: Bridging a significant event in the history of music theory and recent developments in cognitive music research. Danish Yearb. Musicol. 38, 33–55.

Google Scholar

Hansen, N. C., and Huron, D. (2019). Twirling triplets: The qualia of rotation and musical rhythm. Music. Sci. 2, 1–20. doi: 10.1177/2059204318812243

CrossRef Full Text | Google Scholar

Hansen, N. C., and Pearce, M. T. (2014). Predictive uncertainty in auditory sequence processing. Front. Psychol. 5:1052. doi: 10.3389/fpsyg.2014.01052

PubMed Abstract | CrossRef Full Text | Google Scholar

Hansen, N. C., Vuust, P., and Pearce, M. (2016). ‘If you have to ask, you’ll never know’: Effects of specialised stylistic expertise on predictive processing of music. PLoS One 11:e0163584. doi: 10.1371/journal.pone.0163584

PubMed Abstract | CrossRef Full Text | Google Scholar

Hay, R. A., Roach, B. J., Srihari, V. H., Woods, S. W., Ford, J. M., and Mathalon, D. H. (2015). Equivalent mismatch negativity deficits across deviant types in early illness schizophrenia-spectrum patients. Biol. Psychol. 105, 130–137. doi: 10.1016/j.biopsycho.2015.01.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Herdener, M., Esposito, F., di Salle, F., Boller, C., Hilti, C. C., Habermeyer, B., et al. (2010). Musical training induces functional plasticity in human hippocampus. J. Neurosci. 30:1377. doi: 10.1523/jneurosci.4513-09.2010

PubMed Abstract | CrossRef Full Text | Google Scholar

Herholz, S. C., and Zatorre, R. J. (2012). Musical training as a framework for brain plasticity: Behavior, function, and structure. Neuron 76, 486–502. doi: 10.1016/j.neuron.2012.10.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Hofman, P. M., Vlaming, M. S. M. G., Termeer, P. J. J., and Van Opstal, A. J. (2002). A method to induce swapped binaural hearing. J. Neurosci. Methods 113, 167–179. doi: 10.1016/S0165-0270(01)00490-3

CrossRef Full Text | Google Scholar

Hommel, B. (2004). Event files: Feature binding in and across perception and action. Trends Cogn. Sci. 8, 494–500. doi: 10.1016/j.tics.2004.08.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Hyde, K. L., Lerch, J., Norton, A., Forgeard, M., Winner, E., Evans, A. C., et al. (2009). Musical training shapes structural brain development. J. Neurosci. 29, 3019–3025. doi: 10.1523/jneurosci.5118-08.2009

PubMed Abstract | CrossRef Full Text | Google Scholar

Jakobson, L. S., Lewycky, S. T., and Kilgour, A. R. (2008). Memory for verbal and visual material in highly trained musicians. Music Percept. 26, 41–55. doi: 10.1525/mp.2008.26.1.41

CrossRef Full Text | Google Scholar

Jäncke, L., Gaab, N., Wüstenberg, T., Scheich, H., and Heinze, H. (2001). Short-term functional plasticity in the human auditory cortex: An fMRI study. Cogn. Brain Res. 12, 479–485. doi: 10.1016/S0926-6410(01)00092-1

CrossRef Full Text | Google Scholar

Kishon-Rabin, L., Amir, O., Vexler, Y., and Zaltz, Y. (2001). Pitch discrimination: Are professional musicians better than non-musicians? J. Basic Clin. Physiol. Pharmacol. 12, 125–144. doi: 10.1515/JBCPP.2001.12.2.125

PubMed Abstract | CrossRef Full Text | Google Scholar

Koelsch, S., Schröger, E., and Tervaniemi, M. (1999). Superior pre-attentive auditory processing in musicians. Neuroreport 10, 1309–1313.

Google Scholar

Kolinsky, R., Lidji, P., Peretz, I., Besson, M., and Morais, J. (2009). Processing interactions between phonology and melody: Vowels sing but consonants speak. Cognition 112, 1–20. doi: 10.1016/j.cognition.2009.02.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Kourtzi, Z., Betts, L. R., Sarkheil, P., and Welchman, A. E. (2005). Distributed neural plasticity for shape learning in the human visual cortex. PLoS Biol. 3:e204. doi: 10.1371/journal.pbio.0030204

PubMed Abstract | CrossRef Full Text | Google Scholar

Krumhansl, C. L., and Iverson, P. (1992). Perceptual interactions between musical pitch and timbre. J. Exp. Psychol. Hum. Percept. Perform. 18:739. doi: 10.1037/0096-1523.18.3.739

PubMed Abstract | CrossRef Full Text | Google Scholar

Lappe, C., Lappe, M., and Pantev, C. (2016). Differential processing of melodic, rhythmic and simple tone deviations in musicians-an MEG study. Neuroimage 124, 898–905. doi: 10.1016/j.neuroimage.2015.09.059

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, H., and Noppeney, U. (2014). Temporal prediction errors in visual and auditory cortices. Curr. Biol. 24, R309–R310. doi: 10.1016/j.cub.2014.02.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Lerdahl, F., and Jackendoff, R. (1983). A Generative Theory Of Tonal Music. Cambridge, MA: MIT Press.

Google Scholar

Levänen, S., Ahonen, A., Hari, R., McEvoy, L., and Sams, M. (1996). Deviant auditory stimuli activate human left and right auditory cortex differently. Cereb. Cortex 6, 288–296. doi: 10.1093/cercor/6.2.288

PubMed Abstract | CrossRef Full Text | Google Scholar

Levänen, S., Hari, R., McEvoy, L., and Sams, M. (1993). Responses of the human auditory cortex to changes in one versus two stimulus features. Exp. Brain Res. 97, 177–183. doi: 10.1007/BF00228828

PubMed Abstract | CrossRef Full Text | Google Scholar

Lidji, P., Jolicśur, P., Moreau, P., Kolinsky, R., and Peretz, I. (2009). Integrated preattentive processing of vowel and pitch. Ann. N. Y. Acad. Sci. 1169, 481–484. doi: 10.1111/j.1749-6632.2009.04770.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Lima, C. F., and Castro, S. L. (2011). Speaking to the trained ear: Musical expertise enhances the recognition of emotions in speech prosody. Emotion 11:1021. doi: 10.1037/a0024521

PubMed Abstract | CrossRef Full Text | Google Scholar

Maris, E., and Oostenveld, R. (2007). Nonparametric statistical testing of EEG-and MEG-data. J. Neurosci. Methods 164, 177–190. doi: 10.1016/j.jneumeth.2007.03.024

PubMed Abstract | CrossRef Full Text | Google Scholar

Maybery, M. T., Clissa, P. J., Parmentier, F. B. R., Leung, D., Harsa, G., Fox, A. M., et al. (2009). Binding of verbal and spatial features in auditory working memory. J. Mem. Lang. 61, 112–133. doi: 10.1016/j.jml.2009.03.001

CrossRef Full Text | Google Scholar

McBeath, M. K., and Neuhoff, J. G. (2002). The Doppler effect is not what you think it is: Dramatic pitch change due to dynamic intensity change. Psychon. Bull. Rev. 9, 306–313. doi: 10.3758/BF03196286

PubMed Abstract | CrossRef Full Text | Google Scholar

Molholm, S., Martinez, A., Ritter, W., Javitt, D. C., and Foxe, J. J. (2005). The neural circuitry of pre-attentive auditory change-detection: An fMRI study of pitch and duration mismatch negativity generators. Cereb. Cortex 15, 545–551. doi: 10.1093/cercor/bhh155

PubMed Abstract | CrossRef Full Text | Google Scholar

Møller, C., Garza-Villarreal, E. A., Hansen, N. C., Højlund, A., Bærentsen, K. B., Chakravarty, M. M., et al. (2021). Audiovisual structural connectivity in musicians and non-musicians: A cortical thickness and diffusion tensor imaging study. Sci. Rep. 11:4324. doi: 10.1038/s41598-021-83135-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Møller, C., Højlund, A., Bærentsen, K. B., Hansen, N. C., Skewes, J. C., and Vuust, P. (2018). Visually induced gains in pitch discrimination: Linking audio-visual processing with auditory abilities. Atten. Percept. Psychophys. 80, 999–1010. doi: 10.3758/s13414-017-1481-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Müllensiefen, D., Gingras, B., Musil, J., and Stewart, L. (2014). The musicality of non-musicians: An index for assessing musical sophistication in the general population. PLoS One 9:e101091. doi: 10.1371/journal.pone.0101091

CrossRef Full Text | Google Scholar

Münte, T. F., Kohlmetz, C., Nager, W., and Altenmüller, E. (2001). Superior auditory spatial tuning in conductors. Nature 409:580. doi: 10.1038/35054668

PubMed Abstract | CrossRef Full Text | Google Scholar

Musacchia, G., Sams, M., Skoe, E., and Kraus, N. (2007). Musicians have enhanced subcortical auditory and audiovisual processing of speech and music. Proc. Natl. Acad. Sci. U.S.A. 104, 15894–15898. doi: 10.1073/pnas.0701498104

PubMed Abstract | CrossRef Full Text | Google Scholar

Näätänen, R., Gaillard, A. W. K., and Mäntysalo, S. (1978). Early selective-attention effect on evoked potential reinterpreted. Acta Psychol. 42, 313–329. doi: 10.1016/0001-6918(78)90006-9

CrossRef Full Text | Google Scholar

Näätänen, R., Paavilainen, P., Rinne, T., and Alho, K. (2007). The mismatch negativity (MMN) in basic research of central auditory processing: A review. Clin. Neurophysiol. 118, 2544–2590. doi: 10.1016/j.clinph.2007.04.026

PubMed Abstract | CrossRef Full Text | Google Scholar

Näätänen, R., Pakarinen, S., Rinne, T., and Takegata, R. (2004). The mismatch negativity (MMN): Towards the optimal paradigm. Clin. Neurophysiol. 115, 140–144. doi: 10.1016/j.clinph.2003.04.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Nager, W., Kohlmetz, C., Altenmuller, E., Rodriguez-Fornells, A., and Münte, T. F. (2003). The fate of sounds in conductors’ brains: An ERP study. Cogn. Brain Res. 17, 83–93. doi: 10.1016/S0926-6410(03)00083-1

CrossRef Full Text | Google Scholar

Nassi, J. J., and Callaway, E. M. (2009). Parallel processing strategies of the primate visual system. Nat. Rev. Neurosci. 10, 360–372. doi: 10.1038/nrn2619

PubMed Abstract | CrossRef Full Text | Google Scholar

Neuhaus, C., and Knösche, T. R. (2008). Processing of pitch and time sequences in music. Neurosci. Lett. 441, 11–15. doi: 10.1016/j.neulet.2008.05.101

PubMed Abstract | CrossRef Full Text | Google Scholar

Oostenveld, R., Fries, P., Maris, E., and Schoffelen, J.-M. (2011). FieldTrip: Open source software for advanced analysis of MEG, EEG, and invasive electrophysiological data. Comput. Intell. Neurosci. 2011:156869. doi: 10.1155/2011/156869

PubMed Abstract | CrossRef Full Text | Google Scholar

Paavilainen, P., Mikkonen, M., Kilpeläinen, M., Lehtinen, R., Saarela, M., and Tapola, L. (2003b). Evidence for the different additivity of the temporal and frontal generators of mismatch negativity: A human auditory event-related potential study. Neurosci. Lett. 349, 79–82. doi: 10.1016/S0304-3940(03)00787-0

CrossRef Full Text | Google Scholar

Paavilainen, P., Degerman, A., Takegata, R., and Winkler, I. (2003a). Spectral and temporal stimulus characteristics in the processing of abstract auditory features. Neuroreport 14, 715–718. doi: 10.1097/01.wnr.0000064985.96259.ce

CrossRef Full Text | Google Scholar

Paavilainen, P., Valppu, S., and Näätänen, R. (2001). The additivity of the auditory feature analysis in the human brain as indexed by the mismatch negativity: 1+ 1≈ 2 but 1+ 1+ 1< 3. Neurosci. Lett. 301, 179–182. doi: 10.1016/S0304-3940(01)01635-4

CrossRef Full Text | Google Scholar

Palmer, C. (1996). Anatomy of a performance: Sources of musical expression. Music Percept. 13, 433–453. doi: 10.2307/40286178

CrossRef Full Text | Google Scholar

Palmer, C., and Krumhansl, C. L. (1987). Independent temporal and pitch structures in determination of musical phrases. J. Exp. Psychol. Hum. Percept. Perform. 13:116. doi: 10.1037/0096-1523.13.1.116

PubMed Abstract | CrossRef Full Text | Google Scholar

Pantev, C., Oostenveld, R., Engelien, A., Ross, B., Roberts, L. E., and Hoke, M. (1998). Increased auditory cortical representation in musicians. Nature 392, 811–814. doi: 10.1038/33918

PubMed Abstract | CrossRef Full Text | Google Scholar

Pantev, C., Paraskevopoulos, E., Kuchenbuch, A., Lu, Y., and Herholz, S. C. (2015). Musical expertise is related to neuroplastic changes of multisensory nature within the auditory cortex. Eur. J. Neurosci. 41, 709–717. doi: 10.1111/ejn.12788

PubMed Abstract | CrossRef Full Text | Google Scholar

Paraskevopoulos, E., and Herholz, S. (2013). Multisensory integration and neuroplasticity in the human cerebral cortex. Transl. Neurosci. 4, 337–348. doi: 10.2478/s13380-013-0134-1

CrossRef Full Text | Google Scholar

Paraskevopoulos, E., Kraneburg, A., Herholz, S. C., Bamidis, P. D., and Pantev, C. (2015). Musical expertise is related to altered functional connectivity during audiovisual integration. Proc. Natl. Acad. Sci. U.S.A. 112, 12522–12527. doi: 10.1073/pnas.1510662112

PubMed Abstract | CrossRef Full Text | Google Scholar

Paraskevopoulos, E., Kuchenbuch, A., Herholz, S. C., and Pantev, C. (2012). Musical expertise induces audiovisual integration of abstract congruency rules. J. Neurosci. 32, 18196–18203. doi: 10.1523/jneurosci.1947-12.2012

PubMed Abstract | CrossRef Full Text | Google Scholar

Paraskevopoulos, E., Kuchenbuch, A., Herholz, S. C., and Pantev, C. (2014). Multisensory integration during short-term music reading training enhances both uni-and multisensory cortical processing. J. Cogn. Neurosci. 26, 2224–2238. doi: 10.1162/jocn_a_00620

CrossRef Full Text | Google Scholar

Pastuszek-Lipińska, B. (2008). “Musicians outperform nonmusicians in speech imitation,” in Computer Music Modeling And Retrieval: Sense Of Sounds, eds R. Kronland-Martinet, S. Ystad, and K. Jensen (Berlin: Springer), 56–73.

Google Scholar

Peebles, C. A. (2011). The Role Of Segmentation And Expectation In The Perception Of Closure. Ph.D. thesis. Tallahassee, FL: Florida State University.

Google Scholar

Peebles, C. A. (2012). “Closure and expectation: listener segmentation of Mozart minuets,” in Proceedingsof the 12th International Conference on Music Perception and Cognition, eds E. Cambouropoulos, C. Tsougras, P. Mavromatis, and K. Pastiadis (Thessaloniki: Aristotle University), 786–791.

Google Scholar

Peirce, J. W. (2007). PsychoPy: Psychophysics software in Python. J. Neurosci. Methods 162, 8–13. doi: 10.1016/j.jneumeth.2006.11.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Perez, V. B., Woods, S. W., Roach, B. J., Ford, J. M., McGlashan, T. H., Srihari, V. H., et al. (2014). Automatic auditory processing deficits in schizophrenia and clinical high-risk patients: Forecasting psychosis risk with mismatch negativity. Biol. Psychiatry 75, 459–469. doi: 10.1016/j.biopsych.2013.07.038

PubMed Abstract | CrossRef Full Text | Google Scholar

Petersen, B., Weed, E., Sandmann, P., Brattico, E., Hansen, M., Sørensen, S. D., et al. (2015). Brain responses to musical feature changes in adolescent cochlear implant users. Front. Hum. Neurosci. 9:7. doi: 10.3389/fnhum.2015.00007

PubMed Abstract | CrossRef Full Text | Google Scholar

Prince, J. B., and Schmuckler, M. A. (2014). The tonal-metric hierarchy: A corpus analysis. Music Percept. 31, 254–270. doi: 10.1525/mp.2014.31.3.254

CrossRef Full Text | Google Scholar

Proverbio, A. M., Attardo, L., Cozzi, M., and Zani, A. (2015). The effect of musical practice on gesture/sound pairing. Front. Psychol. 6:376. doi: 10.3389/fpsyg.2015.00376

PubMed Abstract | CrossRef Full Text | Google Scholar

Proverbio, A. M., Calbi, M., Manfredi, M., and Zani, A. (2014). Audio-visuomotor processing in the Musician’s brain: An ERP study on professional violinists and clarinetists. Sci. Rep. 4:5866. doi: 10.1038/srep05866

PubMed Abstract | CrossRef Full Text | Google Scholar

Putkinen, V., Tervaniemi, M., Saarikivi, K., de Vent, N., and Huotilainen, M. (2014). Investigating the effects of musical training on functional brain development with a novel Melodic MMN paradigm. Neurobiol. Learn. Mem. 110, 8–15. doi: 10.1016/j.nlm.2014.01.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Quinlan, P. T. (2003). Visual feature integration theory: Past, present, and future. Psychol. Bull. 129:643. doi: 10.1037/0033-2909.129.5.643

PubMed Abstract | CrossRef Full Text | Google Scholar

Quiroga-Martinez, D. R., Hansen, N. C., Højlund, A., Pearce, M., Brattico, E., Holmes, E., et al. (2021). Musicianship and melodic predictability enhance neural gain in auditory cortex during pitch deviance detection. Hum. Brain Mapp. 42, 5595–5608. doi: 10.1002/hbm.25638

PubMed Abstract | CrossRef Full Text | Google Scholar

Quiroga-Martinez, D. R., Hansen, N. C., Højlund, A., Pearce, M., Brattico, E., and Vuust, P. (2020). Musical prediction error responses similarly reduced by predictive uncertainty in musicians and non-musicians. Eur. J. Neurosci. 51, 2250–2269. doi: 10.1111/ejn.14667

PubMed Abstract | CrossRef Full Text | Google Scholar

Recasens, M., and Uhlhaas, P. J. (2017). Test-retest reliability of the magnetic mismatch negativity response to sound duration and omission deviants. Neuroimage 157, 184–195. doi: 10.1016/j.neuroimage.2017.05.064

PubMed Abstract | CrossRef Full Text | Google Scholar

Rosburg, T. (2003). Left hemispheric dipole locations of the neuromagnetic mismatch negativity to frequency, intensity and duration deviants. Cogn. Brain Res. 16, 83–90. doi: 10.1016/S0926-6410(02)00222-7

CrossRef Full Text | Google Scholar

Ruusuvirta, T. (2001). Differences in pitch between tones affect behaviour even when incorrectly indentified in direction. Neuropsychologia 39, 876–879. doi: 10.1016/S0028-3932(00)00137-8

CrossRef Full Text | Google Scholar

Ruusuvirta, T., Huotilainen, M., Fellman, V., and Näätänen, R. (2003). The newborn human brain binds sound features together. Neuroreport 14, 2117–2119. doi: 10.1097/00001756-200311140-00021

CrossRef Full Text | Google Scholar

Sassenhagen, J., and Draschkow, D. (2019). Cluster-based permutation tests of MEG/EEG data do not establish significance of effect latency or location. Psychophysiology 56:e13335. doi: 10.1111/psyp.13335

PubMed Abstract | CrossRef Full Text | Google Scholar

Schiltz, C., Bodart, J. M., Dubois, S., Dejardin, S., Michel, C., Roucoux, A., et al. (1999). Neuronal mechanisms of perceptual learning: Changes in human brain activity with training in orientation discrimination. Neuroimage 9, 46–62. doi: 10.1006/nimg.1998.0394

PubMed Abstract | CrossRef Full Text | Google Scholar

Schlaug, G. (2015). Musicians and music making as a model for the study of brain plasticity. Prog. Brain Res. 217, 37–55. doi: 10.1016/bs.pbr.2014.11.020

PubMed Abstract | CrossRef Full Text | Google Scholar

Schön, D., and Besson, M. (2002). Processing pitch and duration in music reading: A RT-ERP study. Neuropsychologia 40, 868–878. doi: 10.1016/S0028-3932(01)00170-1

CrossRef Full Text | Google Scholar

Schröger, E. (1995). Processing of auditory deviants with changes in one versus two stimulus dimensions. Psychophysiology 32, 55–65. doi: 10.1111/j.1469-8986.1995.tb03406.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Schröger, E. (1996). Interaural time and level differences: Integrated or separated processing? Hear. Res. 96, 191–198. doi: 10.1016/0378-5955(96)00066-4

CrossRef Full Text | Google Scholar

Schulz, M., Ross, B., and Pantev, C. (2003). Evidence for training-induced crossmodal reorganization of cortical functions in trumpet players. Neuroreport 14, 157–161. doi: 10.1097/01.wnr.0000053061.10406.c7

CrossRef Full Text | Google Scholar

Shamma, S. (2008). On the emergence and awareness of auditory objects. PLoS Biol. 6:e155. doi: 10.1371/journal.pbio.0060155

PubMed Abstract | CrossRef Full Text | Google Scholar

Skoe, E., and Kraus, N. (2012). A little goes a long way: How the adult brain is shaped by musical training in childhood. J. Neurosci. 32, 11507–11510. doi: 10.1523/jneurosci.1949-12.2012

PubMed Abstract | CrossRef Full Text | Google Scholar

Sohoglu, E., Kumar, S., Chait, M., and Griffiths, T. D. (2020). Multivoxel codes for representing and integrating acoustic features in human cortex. Neuroimage 217:116661. doi: 10.1016/j.neuroimage.2020.116661

PubMed Abstract | CrossRef Full Text | Google Scholar

Sorati, M., and Behne, D. M. (2020). Audiovisual modulation in music perception for musicians and non-musicians. Front. Psychol. 11:1094. doi: 10.3389/fpsyg.2020.01094

PubMed Abstract | CrossRef Full Text | Google Scholar

Spence, C., and Frings, C. (2020). Multisensory feature integration in (and out) of the focus of spatial attention. Atten. Percept. Psychophys. 82, 363–376. doi: 10.3758/s13414-019-01813-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Stein, B. E. (2012). The New Handbook of Multisensory Processing. Cambridge, MA: MIT Press.

Google Scholar

Sussman, E., Gomes, H., Nousak, J. M. K., Ritter, W., and Vaughan, H. G. (1998). Feature conjunctions and auditory sensory memory. Brain Res. 793, 95–102. doi: 10.1016/S0006-8993(98)00164-4

CrossRef Full Text | Google Scholar

Symons, A. E., and Tierney, A. (2021). Musicians upweight pitch during prosodic categorization. PsyArxiv [Preprint]. doi: 10.31234/osf.io/c4396

CrossRef Full Text | Google Scholar

Takegata, R., Huotilainen, M., Rinne, T., Näätänen, R., and Winkler, I. (2001a). Changes in acoustic features and their conjunctions are processed by separate neuronal populations. Neuroreport 12, 525–529. doi: 10.1097/00001756-200103050-00019

PubMed Abstract | CrossRef Full Text | Google Scholar

Takegata, R., Paavilainen, P., Näätänen, R., and Winkler, I. (2001b). Preattentive processing of spectral, temporal, and structural characteristics of acoustic regularities: A mismatch negativity study. Psychophysiology 38, 92–98. doi: 10.1111/1469-8986.3810092

CrossRef Full Text | Google Scholar

Takegata, R., Paavilainen, P., Näätänen, R., and Winkler, I. (1999). Independent processing of changes in auditory single features and feature conjunctions in humans as indexed by the mismatch negativity. Neurosci. Lett. 266, 109–112. doi: 10.1016/S0304-3940(99)00267-0

CrossRef Full Text | Google Scholar

Taulu, S., and Simola, J. (2006). Spatiotemporal signal space separation method for rejecting nearby interference in MEG measurements. Phys. Med. Biol. 51:1759. doi: 10.1088/0031-9155/51/7/008

CrossRef Full Text | Google Scholar

Taulu, S., Kajola, M., and Simola, J. (2004). Suppression of interference and artifacts by the signal space separation method. Brain Topogr. 16, 269–275. doi: 10.1023/B:BRAT.0000032864.93890.f9

CrossRef Full Text | Google Scholar

Tervaniemi, M., Castaneda, A., Knoll, M., and Uther, M. (2006). Sound processing in amateur musicians and nonmusicians: Event-related potential and behavioral indices. Neuroreport 17, 1225–1228. doi: 10.1097/01.wnr.0000230510.55596.8b

CrossRef Full Text | Google Scholar

Tervaniemi, M., Ilvonen, T., Sinkkonen, J., Kujala, A., Alho, K., Huotilainen, M., et al. (2000). Harmonic partials facilitate pitch discrimination in humans: Electrophysiological and behavioral evidence. Neurosci. Lett. 279, 29–32. doi: 10.1016/S0304-3940(99)00941-6

CrossRef Full Text | Google Scholar

Tervaniemi, M., Kruck, S., De Baene, W., Schröger, E., Alter, K., and Friederici, A. D. (2009). Top-down modulation of auditory processing: Effects of sound context, musical expertise and attentional focus. Eur. J. Neurosci. 30, 1636–1642. doi: 10.1111/j.1460-9568.2009.06955.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Tervaniemi, M., Sinkkonen, J., Virtanen, J., Kallio, J., Ilmoniemi, R. J., Salonen, O., et al. (2005). Test-retest stability of the magnetic mismatch response (MMNm). Clin. Neurophysiol. 116, 1897–1905. doi: 10.1016/j.clinph.2005.03.025

PubMed Abstract | CrossRef Full Text | Google Scholar

Tillmann, B., and Lebrun-Guillaud, G. (2006). Influence of tonal and temporal expectations on chord processing and on completion judgments of chord sequences. Psychol. Res. 70, 345–358. doi: 10.1007/s00426-005-0222-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Timm, L., Vuust, P., Brattico, E., Agrawal, D., Debener, S., Büchner, A., et al. (2014). Residual neural processing of musical sound features in adult cochlear implant users. Front. Hum. Neurosci. 8:181. doi: 10.3389/fnhum.2014.00181

PubMed Abstract | CrossRef Full Text | Google Scholar

Treisman, A. M., and Gelade, G. (1980). A feature-integration theory of attention. Cogn. Psychol. 12, 97–136. doi: 10.1016/0010-0285(80)90005-5

CrossRef Full Text | Google Scholar

van Turennout, M., Ellmore, T., and Martin, A. (2000). Long-lasting cortical plasticity in the object naming system. Nat. Neurosci. 3, 1329–1334. doi: 10.1038/81873

PubMed Abstract | CrossRef Full Text | Google Scholar

Vuust, P., Brattico, E., Glerean, E., Seppänen, M., Pakarinen, S., Tervaniemi, M., et al. (2011). New fast mismatch negativity paradigm for determining the neural prerequisites for musical ability. Cortex 47, 1091–1098. doi: 10.1016/j.cortex.2011.04.026

PubMed Abstract | CrossRef Full Text | Google Scholar

Vuust, P., Brattico, E., Seppänen, M., Näätänen, R., and Tervaniemi, M. (2012a). Practiced musical style shapes auditory skills. Ann. N. Y. Acad. Sci. 1252, 139–146. doi: 10.1111/j.1749-6632.2011.06409.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Vuust, P., Brattico, E., Seppänen, M., Näätänen, R., and Tervaniemi, M. (2012b). The sound of music: Differentiating musicians using a fast, musical multi-feature mismatch negativity paradigm. Neuropsychologia 50, 1432–1443. doi: 10.1016/j.neuropsychologia.2012.02.028

PubMed Abstract | CrossRef Full Text | Google Scholar

Vuust, P., Heggli, O. A., Friston, K., and Kringelbach, M. (2022). Music in the brain. Nat. Neurosci. 23, 287–305.

Google Scholar

Vuust, P., Liikala, L., Näätänen, R., Brattico, P., and Brattico, E. (2016). Comprehensive auditory discrimination profiles recorded with a fast parametric musical multi-feature mismatch negativity paradigm. Clin. Neurophysiol. 127, 2065–2077. doi: 10.1016/j.clinph.2015.11.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Vuust, P., Ostergaard, L., Pallesen, K. J., Bailey, C., and Roepstorff, A. (2009). Predictive coding of music-brain responses to rhythmic incongruity. Cortex 45, 80–92. doi: 10.1016/j.cortex.2008.05.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Vuust, P., Pallesen, K. J., Bailey, C., van Zuijen, T. L., Gjedde, A., Roepstorff, A., et al. (2005). To musicians, the message is in the meter pre-attentive neuronal responses to incongruent rhythm are left-lateralized in musicians. Neuroimage 24, 560–564. doi: 10.1016/j.neuroimage.2004.08.039

PubMed Abstract | CrossRef Full Text | Google Scholar

Wacongne, C., Changeux, J.-P., and Dehaene, S. (2012). A neuronal model of predictive coding accounting for the mismatch negativity. J. Neurosci. 32, 3665–3678. doi: 10.1523/jneurosci.5003-11.2012

PubMed Abstract | CrossRef Full Text | Google Scholar

Waters, A. J., and Underwood, G. (1999). Processing pitch and temporal structures in music reading: Independent or interactive processing mechanisms? Eur. J. Cogn. Psychol. 11, 531–553. doi: 10.1080/095414499382282

CrossRef Full Text | Google Scholar

Watts, C., Barnes-Burroughs, K., Andrianopoulos, M., and Carr, M. (2002). Potential factors related to untrained singing talent: A survey of singing pedagogues. J. Voice 17, 298–307. doi: 10.1067/S0892-1997(03)00068-7

CrossRef Full Text | Google Scholar

Widmer, G., and Goebl, W. (2004). Computational models of expressive music performance: The state of the art. J. New Music Res. 33, 203–216. doi: 10.1080/0929821042000317804

CrossRef Full Text | Google Scholar

Wilbiks, J. M., and O’Brien, C. (2020). Musical training improves audiovisual integration capacity under conditions of high perceptual load. Vision 4:9. doi: 10.3390/vision4010009

PubMed Abstract | CrossRef Full Text | Google Scholar

Winkler, I., Czigler, I., Jaramillo, M., and Paavilainen, P. (1998). Temporal constraints of auditory event synthesis: Evidence from ERPs. Neuroreport 9, 495–499. doi: 10.1097/00001756-199802160-00025

PubMed Abstract | CrossRef Full Text | Google Scholar

Winkler, I., Czigler, I., Sussman, E., Horváth, J., and Balázs, L. (2005). Preattentive binding of auditory and visual stimulus features. J. Cogn. Neurosci. 17, 320–339. doi: 10.1162/0898929053124866

PubMed Abstract | CrossRef Full Text | Google Scholar

Wolff, C., and Schröger, E. (2001). Human pre-attentive auditory change-detection with single, double, and triple deviations as revealed by mismatch negativity additivity. Neurosci. Lett. 311, 37–40. doi: 10.1016/S0304-3940(01)02135-8

CrossRef Full Text | Google Scholar

Woods, D. L., Alain, C., and Ogawa, K. H. (1998). Conjoining auditory and visual features during high-rate serial presentation: Processing and conjoining two features can be faster than processing one. Percept. Psychophys. 60, 239–249. doi: 10.3758/BF03206033

PubMed Abstract | CrossRef Full Text | Google Scholar

Wright, B. A., and Zhang, Y. (2006). A review of learning with normal and altered sound-localization cues in human adults. Int. J. Audiol. 45, 92–98. doi: 10.1080/14992020600783004

PubMed Abstract | CrossRef Full Text | Google Scholar

Ylinen, S., Huotilainen, M., and Näätänen, R. (2005). Phoneme quality and quantity are processed independently in the human brain. Neuroreport 16, 1857–1860. doi: 10.1097/01.wnr.0000185959.11465.9b

CrossRef Full Text | Google Scholar

Yotsumoto, Y., Watanabe, T., and Sasaki, Y. (2008). Different dynamics of performance and brain activation in the time course of perceptual learning. Neuron 57, 827–833. doi: 10.1016/j.neuron.2008.02.034

PubMed Abstract | CrossRef Full Text | Google Scholar

Yu, K., Chen, Y., Wang, M., Wang, R., and Li, L. (2022). Distinct but integrated processing of lexical tones, vowels, and consonants in tonal language speech perception: Evidence from mismatch negativity. J. Neurolinguist. 61:101039. doi: 10.1016/j.jneuroling.2021.101039

CrossRef Full Text | Google Scholar

Zatorre, R. J., Delhommeau, K., and Zarate, J. M. (2012). Modulation of auditory cortex response to pitch variation following training with microtonal melodies. Front. Psychol. 3:544. doi: 10.3389/fpsyg.2012.00544

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: auditory perception, music, expertise, feature integration, mismatch negativity, melody, pitch, magnetoencephalography

Citation: Hansen NC, Højlund A, Møller C, Pearce M and Vuust P (2022) Musicians show more integrated neural processing of contextually relevant acoustic features. Front. Neurosci. 16:907540. doi: 10.3389/fnins.2022.907540

Received: 29 March 2022; Accepted: 08 September 2022;
Published: 13 October 2022.

Edited by:

Andre Rupp, University Hospital Heidelberg, Germany

Reviewed by:

Claude Alain, Rotman Research Institute (RRI), Canada
Eino Partanen, University of Helsinki, Finland

Copyright © 2022 Hansen, Højlund, Møller, Pearce and Vuust. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Niels Chr. Hansen, bmkzbHNDaHJIYW5zZW5AZ21haWwuY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.