Skip to main content

ORIGINAL RESEARCH article

Front. Cognit., 17 January 2024
Sec. Perception
This article is part of the Research Topic Neurocognitive Bases of Music Reading View all 4 articles

The key role of the right posterior fusiform gyrus in music reading: an electrical neuroimaging study on 90 readers

  • 1Cognitive Electrophysiology Lab, Department of Psychology, University of Milano-Bicocca, Milan, Italy
  • 2NeuroMI, Milan Center for Neuroscience, Milan, Italy
  • 3School of Psychology, Vita-Salute San Raffaele University, Milan, Italy
  • 4Department of Psychology, Psychologisches Institut, University of Zurich, Zürich, Switzerland

Introduction: In this study, we employed a combined electromagnetic recording technique, i.e., electroencephalogram (EEG)/event-related potentials (ERPs) plus standardized weighted low-resolution electromagnetic tomography (swLORETA), to investigate the neural mechanism subserving the orthographic processing of symbols in language and music. While much is known about word processing, the current literature remains inconclusive regarding music reading, as its mechanisms appear to be left lateralized in some cases (as suggested by music-alexia clinical case reports) and either right-sided or bilateral in others, depending on the study and the methodology used.

Methods: In this study, 90 right-handed participants with varying musical abilities and sexes performed an attentional selection task that involved the recognition of target letters and musical notes, while their EEG signals were recorded from 128 sites.

Results: The occipito/temporal N170 component of ERPs (170–210 ms) was found strictly left-sided during letter selection and bilateral (with a right-hemispheric tendency) during note selection. Source reconstruction data indicated the preponderant engagement of the right posterior fusiform gyrus (BA19) for processing musical notes. Also involved were other brain regions belonging to the word reading circuit, including the left-sided visual word form area (VWFA) and frontal eye-fields (FEFs).

Discussion: This finding provides an explanation for the infrequent appearance of musical alexia cases (previously observed only in patients with left hemispheric lesions). It also suggests how musical literacy could be a rehabilitative and preventive factor for dyslexia, by promoting neuroplasticity and bilaterality in the reading areas.

Introduction

Knowledge of orthographic mechanisms for recognizing alphabetic characters, such as letters, is highly developed. Conversely, visual mechanisms for reading and recognizing musical notation exhibit more ambiguity and inconsistency. Many of the inconsistencies relate to the exact location of the mechanism, which should be in the left hemisphere based on the limited data from music alexia. However, numerous studies have proposed a right posterior region that is highly responsive to notation. In this study, we have attempted to possibly identify the electromagnetic sources of notation processing (the visual note form area, VNFA) by testing a very large group of participants, including professional musicians, but also people with elementary musical education. It is known that the Visual Word Form Area (VWFA), located in the medial portion of the left fusiform gyrus, plays a key role in letter and word recognition, being more responsive to strings of letters rather than other symbols (Kuriki et al., 1998; Cohen et al., 2000, 2003; Garrett et al., 2000; Polk et al., 2002; Flowers et al., 2004; Pernet et al., 2005; Vigneau et al., 2005; Kronbichler et al., 2007).

The development and specialization of this area for recognition of written words enables rapid reading, as it increases the perceptual ability of words, making it sensitive to recurrent features of the writing system (McCandliss et al., 2003). Although the VWFA processes structurally invariant representations of visually presented words as abstract letter sequences, independent of size, style, and character (Dehaene et al., 2002), several studies have shown that the regions of the left fusiform gyrus, corresponding to the VWFA, are sensitive to sublexical properties of words such as their frequency of use and orthographic familiarity (Mechelli et al., 2003; Kronbichler et al., 2004; Devlin et al., 2006; Bruno et al., 2008). Indeed, data have shown higher activations for words than strings of letters and for words with high- than low-frequency of use (e.g., Proverbio et al., 2008).

The precise localization of VWFA in the left occipitotemporal sulcus has been explained by its anatomical proximity to the left perisylvian area, which is deputed to the processing of spoken language; in this context, the VWFA could play the role of providing direct connections between the visual and linguistic areas, and in particular between the ventral visual recognition system and the perisylvian areas of language (Bouhali et al., 2014). The early study by Cohen et al. (2000) found a left lateralization of hemispheric activation in control subjects, regardless of the visual field in which the stimulus was presented. The role of the left occipito-temporal area in word reading was later supported by Devlin et al. (2006), who identified its specific components in the posterior occipito-temporal sulcus and the inferior ventral fusiform and temporal gyrus. The results found an activation of the left posterior fusiform gyrus in both word and pseudoword reading, with peak activation in response to words located in the left occipito-temporal sulcus, and extending to the convexity of the posterior fusiform gyrus and the inferior temporal gyrus. In addition, with regard to the left fusiform gyrus, the more posterior portion would appear to be more responsive to single letters, while the anterior portion would activate more extensively in response to letter strings (James et al., 2005). Thesen et al. (2012) through an fMRI study found that the selective activation for letters in the left posterior fusiform gyrus was prior to more anterior selective activation for words.

The N170 ERP component is the electromagnetic manifestation of VWFA activation, with its negative peak exhibiting the highest amplitude to letter/words than other objects at ~170 ms in the left occipito/temporal region (Bentin et al., 1999; Rossion et al., 2003; Proverbio et al., 2008). The N170 component has also been shown to be sensitive to musical notation, being larger to target than non-target notes at bilateral occipito/temporal sites, in selective attention tasks requiring visual recognition and response to tones of different pitch (Proverbio et al., 2013).

As for the neural mechanism subserving the ability to read musical notation, the pioneeristic study by Sergent et al. (1992) found that reading musical notations and translating these notations into musical gestures resulted in the activation of cortical areas distinct from, but adjacent to, those devoted to orthographic reading. Reading a musical score (without listening or playing) was associated with the activation of bilateral extra-striate visual areas (BA18). However, the left occipito-parietal junction (BA19), rather than the left lingual gyrus (LG) and fusiform gyrus (FG), was activated by musical notation. This was interpreted as a reflection of the role of the dorsal visual system in spatial processing (pentagram, spatial position of notes). In an ERP study conducted with control subjects and musicians, which included notation reading and a selective attention task with notes vs. letters, we discovered that there exists only a partial overlap between reading words and reading notes (Proverbio et al., 2013). In musicians, musical notation recognition strongly activated the left FG (BA27), but also the right inferior occipital gyrus (BA18) and the right FG, BA37). Similarly, Nakada et al. (1998) identified the right transverse occipital sulcus as being more active when reading musical scores compared to reading English or Japanese texts, which activated the VWFA. Additionally, Schön et al. (2001) emphasized the role of the right occipitotemporal junction in music reading, drawing a comparison to word reading. This finding was explained in light of the hypothesis that note pitch is processed with regards spatial position within the pentagram (a right hemispheric function), while words and, more in general, alphanumeric characters, are encoded in relation to their shape. Of course, other musical accidentals such as flats, sharps, pauses, and duration indications are also recognized based on their shape rather than their position. Schön et al. (2001) therefore suggested that the right occipitotemporal junction might represent the musical functional homolog of the VWFA for written words. However, there is not a clear agreement in the literature about the role of right visual areas in notation processing. Mongelli et al. (2017) used fMRI to assess activations induced by musical notation in professional musicians and naive controls, comparing them to activations by written words and other classes of objects. They found that both words and music activated the left occipitotemporal cortex (VWFA), but music-related activations in the left occipitotemporal cortex peaked posterior and lateral to word-related activations. In addition, a bilateral network including the right occipito/temporal cortex was more connected during music than word processing. In another study, Lu et al. (2019), through magnetoencephalography, compared the activations of brain areas during silent word and music reading in professional musicians and found an activation of left sided regions in both music and speech tasks. In addition, it was found the bilateral activation of the superior parietal cortex, which was interpreted as reflecting encoding and visuomotor transformation (as also found by Meister et al., 2004; Proverbio et al., 2014). The parietal cortex was also observed to be active in an fMRI study by Stewart et al. (2003), in which a group of non-musicians underwent training to learn to play the piano and read music for 15 weeks. Post-training scans revealed increased activation of the right superior parietal cortex and supramarginal gyrus, consistent with other imaging studies (Sergent et al., 1992; Schön et al., 2001; Roux et al., 2007). Finally, Wong and Gauthier (2010) tested the ability to selectively attend to part of a four-note sequence in music reading experts and novices in a same/different task and found a correlation between right FG activation and holistic strategy.

All in all, while its seems that reading musical activates right posterior areas to some extent, there is not complete agreement in the literature about the precise network involved in the orthographic stage of music processing (150–200 ms), that from study to study would involve the transverse occipital sulcus, the occipital gyrus, the inferior occipital gyrus, the superior occipital gyrus, the occipito/temporal junction, the FG, the superior parietal and supramarginal cortices. These differences may be due to the different tasks or methodological constraints used to test reading of musical notation: experimental paradigms range from reading while playing, reading in silence, reading at first sight, attentional selection of notes, reading for learning to play, etc. For example, Lu et al. (2019) used MEG to compare reading of musical notation vs. English letters in professional musicians. Both note and letter reading tasks required translation to phonological codes and activated left hemisphere language areas and bilateral parietal cortex. In addition, the authors measured a laterality index to determine which of the two hemispheres was more active during a 1–letter sounding and a 1-note reading tasks, and found that in about half of the participants the left fusiform gyrus was more activated than the right fusiform gyrus (and vice versa) in the 150–200 ms time window: these results do not provide a convincing pattern of results.

Overall, some music reading studies indicate greater visual activation on the right side (Nakada et al., 1998; Li et al., 2017; Lu et al., 2021), whereas others suggest a more bilateral distribution (Proverbio et al., 2013) or smaller activation on the right side compared to the left (Mongelli et al., 2017). One potential theoretical problem with the neuroimaging literature is that music alexia, a relatively rare disorder that consists of an inability to recognize musical notes, is almost invariably linked to left-hemispheric lesions. Indeed, Peretz and Zatorre (2005) assumed that music sight-reading would involve left occipito/temporal areas, whose lesion would lead to music alexia (e.g., Schön et al., 2001).

There are three variants of music alexia that have been delineated over the years: difficulty in reading individual notes (Fasanaro et al., 1990; Cappelletti et al., 2000), misreading the pitch of notes rather than the rhythm (Brust, 1980), and inability to recognize both rhythm and correct pitch (Brust, 1980; Midorikawa et al., 2003). In all quoted cases, and in others, music alexia was consistently associated with a left hemispheric lesion; more specifically this was observed in the left temporoparieto-occipital region in Fasanaro et al. (1990), in the left posterior temporal lobe (along with a small right occipitotemporal lesion) in Cappelletti et al. (2000), in the left temporal lobe in Brust (1980), in the left superior temporal gyrus in Midorikawa et al. (2003), in the left angular gyrus in Kawamura et al. (2000), and in the left ventral temporo-occipital cortex in Starrfelt and Wong (2017). Again, one of the earliest cases of musical alexia reported by Horikoshi et al. (1997) regarded the case of a 26-year-old pianist who, due to an intracerebral hematoma in the left occipital lobe, experienced difficulty in reading music, especially in relation to the pitch of notes, while maintaining preserved rhythm reading, auditory recognition, and playing ability.

Note and word reading exhibit related processing and production processes, although functionally the areas involved are close but, for the most part, not overlapping. This, in fact, provides an explanation with respect to the possibility of finding clinical case descriptions that report associations or dissociations between language disorders and musical disorders. Despite cases of music alexia being associated with damage to the left occipito/temporal or occipito/parietal cortex, recent neurometabolic and electromagnetic studies have demonstrated the crucial role of the right visual cortex in reading notes and playing music scores. We aim to clarify the matter by gauging participants' capacity to discern and identify written words and musical notes. This will be achieved by observing an unprecedentedly large sample of strictly right-handed participants and comparing their ability to discern and identify written words and notes. In order to measure the neurological response to both stimuli, we'll be considering N170 responses for electromagnetic manifestation. These responses will be studied through brain source reconstruction for participants with varying degrees of musical experience.

Methods

Participants

Ninety healthy, right-handed post-graduate students from the Milan area participated in the experiment. EEG data from four participants were disregarded due to an abundance of EEG artifacts, and two participants were disqualified due to amplifier blocking. The final sample consisted of 36 male and 48 female participants, with a mean age of 25.6 years (SD = 6.3). Half of the participants were professional musicians with a Conservatory degree and typically read music scores in the Violin and/or Bass clefs (44 musicians), whereas the other half had studied music for ~3 years in high school (40 non-musicians). They were matched for socio-cultural status, and were all native Italian speakers. Their lateral preference was assessed through the administration of the Edinburgh Inventory (Oldfield, 1971) whose scale ranges from −1 (indicating left-handedness) to +1 (indicating right-handedness). Musicians' average laterality score was 0.714 (standard deviation = 1.18), compared to non-musicians' average score of 0.80 (standard deviation = 1.145), indicating that all participants were fully right-handed (average score = 0.76). Inclusion criteria were having never suffered from psychiatric or neurologic disorders, not being currently assuming drugs or narcoleptics, or being epilepsy-predisposed. All participants had normal or glass-corrected vision and hearing. No participant suffered or ever suffered from learning or reading disorder (e.g., developmental dyslexia, alexia, autism, ADHD, etc...). All participants provided written informed consent and were unaware of the research purpose. The study lasted approximately 3 h, encompassing breaks, training, and administering the questionnaire. Participants voluntarily offered their participation or received academic credits. The Ethics Committee of University of Milano-Bicocca (protocol number RM-2021-370) approved the project.

Stimulus and procedure

The research included 300 Italian words and 300 music bars with different lengths and complexities. These were presented randomly on a computer screen. The stimuli used were identical to those utilized in an earlier ERP study (Proverbio et al., 2013). The words were typed in capital letters and varied in length from 4 to 10 letters, while the music bars were 4–8 notes in length. Two experimental conditions were used: a note recognition task and a letter recognition task. They consisted in pressing a key at the sight of a specific letter contained in a word, or note contained in a musical measure, depending on the experimental condition, while ignoring non-target letters or notes. The participants completed one task in the first half, and the other in the second half of the experiment, with stimulus presentation lasting 1,600 ms and an ISI of 1,000 to 1,200 ms. The stimuli were matched for duration, frequency of use, and target/non-target categories. For the note recognition task, the targets were “mi,” “fa,” “sol,” “la,” and “si” of the middle piano octave, while the orthographic task used the letters B, G, L, M, and S. Participants received verbal and visual cues before each trial, and responded by pressing a joystick button with their index finger. Stimuli were randomly presented in 12 runs of 50, with halves of targets and non-targets. The two experimental sessions were preceded by two training sequences, and a fixation point was used to minimize movement. The response hand and presentation order were both counterbalanced across participants.

EEG recordings

EEG data were recorded using a 128-electrode cap, at a sampling rate of 512 Hz. hEOG and vEOG eye movements were recorded and linked mastoids were used as a reference lead. Electrode impedance was kept below 5 KOhm and artifact rejection methods removed contaminated EEG segments. A computerized criterion for rejecting artifacts was a peak-to-peak amplitude exceeding 50 μV. Evoked response potentials (ERPs) were then averaged off-line and a band-pass 0.16–30 Hz filter was applied to them. ERPs were averaged offline with a time epoch of 1,500 ms. N170 orthographic response was identified and measured in the time window and scalp location where it reached its maximum amplitude, and according to previous literature (e.g., Proverbio et al., 2013; Cnudde et al., 2021). N170 mean area amplitude of the N170 was quantified at occipito-temporal sites (PPO9h-PPO10h) in between 170 and 210 ms in response to notes and words, in musicians and controls, and across the different attentional conditions. Behavioral data (response times in ms) were additionally recorded.

Individual N170 amplitude values were subjected to repeated measures ANOVA. Factors of variance were: 1 between-group factor “group” (musicians vs. controls) plus 3 within-group factors: “stimulus type” (notes vs. words), “target” (non-targets, targets); “hemisphere” (left, right). A further repeated-measure ANOVA was performed on response times (RTs) data. Between group factor was “group” (musicians vs. controls). Within-group factors were: “stimulus type” (note vs. word), “target” (non-target, target), and “response hand” (left, right).

Tukey post-hoc comparisons were carried out to test differences among means. The effect size for the statistically significant factors was estimated using partial Eta Squared (ηp2) and the Greenhouse-Geisser correction was applied to account for non-sphericity of the data.

Source reconstruction

In order to identify the intracranial sources explaining the surface electrical potentials Standardized low-resolution electromagnetic tomography (sLORETA; Pascual-Marqui, 2002) was performed on ERP voltages. In particular, LORETA was applied to mean voltages recorded in the 170–190 ms time range, and corresponding to N170 response. LORETA is a discrete linear solution to the inverse EEG problem, and it corresponds to the 3D distribution of neural electric activity that maximizes similarity (i.e., maximizes synchronization) in terms of orientation and strength between neighboring neuronal populations (represented by adjacent voxels). In this study, an improved version of standardized weighted low-resolution brain electromagnetic tomography was used (swLORETA, Palmero-Soler et al., 2007), which incorporates a singular value decomposition-based lead field weighting method. The source space properties included: localization within the gray matter; a grid spacing of five points (the distance between 2 calculation points) and an estimated signal-to-noise ratio (SNR) of 3, which defines the regularization (higher values indicating less regularization and therefore less blurred results). Using a value of 3–4 for SNR computation in Tikhonov's regularization results in superior accuracy for all assessed inverse problems. swLORETA was performed on the grand-averaged data to identify statistically significant electromagnetic dipoles (p < 0.05) in which larger magnitudes correlated with more significant activation. The data were automatically re-referenced to the average reference as part of the LORETA analysis. A realistic boundary element model (BEM) was derived from a T1-weighted 3D MRI dataset through segmentation of the brain tissue. This BEM model consisted of a homogeneous compartment comprising 3,446 vertices and 6,888 triangles. Advanced source analysis (ASA) employs a realistic head model of three layers (scalp, skull, and brain) created using the boundary element model. This realistic head model comprises a set of irregularly shaped boundaries and the conductivity values for the compartments between them. Each boundary is represented by a series of interconnected points, forming plane triangles to create an approximation. The triangulation leads to a more or less evenly distributed mesh of triangles as a function of the chosen grid value. A smaller value for the grid spacing results in finer meshes and vice versa. With the aforementioned realistic head model of three layers, the segmentation is assumed to include current generators of brain volume, including both gray and white matter. Scalp, skull, and brain region conductivities were assumed to be 0.33, 0.0042, and 0.33, respectively. The source reconstruction solutions were projected onto the 3D MRI of the Collins brain, which was supplied by the Montreal Neurological Institute. The probabilities of source activation based on Fisher's F test were provided for each independent EEG source, whose values are indicated in a “unit” scale in nA (the larger the value, the more significant the activation). It should be noted, however, that the spatial resolution of swLORETA is somewhat limited compared to other neuroimaging techniques like MEG or fMRI. Both the head model's segmentation and generation were conducted through the use of Advanced Neuro Technology, a software program developed by ASA (Zanow and Knösche, 2004).

Results

Behavioral data

The ANOVA yielded the significance of group [F(1, 82) = 28.7, p < 0.00001; ηp2 = 0.26], with faster RTs in musicians than controls (Figure 1, left). Also significant was the factor stimulus type [F(1, 82) = 785, p < 0.00001; ηp2 = 0.91], with faster responses to words than notes. The significant interaction of stimulus type x group [F(1, 82) = 40.35, p < 0.00001; ηp2 = 0.33], and relative post-hoc comparisons, showed faster RTs to words than to notes in both groups, faster RTs to notes in musicians than in controls (p < 0.00001), but not significantly faster RTs to words in musicians (550 ms, SE = 10.5) than in controls (570.3 ms, SE = 10.8), as can be seen in Figure 1, left. The further interaction of response hand x stimulus type [F(1, 82) = 4.1, p < 0.05; ηp2 = 0.04] showed faster (p < 0.00001) RTs to words with the right hand (Figure 1, right), but a non-significant trend (p = 0.1) for faster RTs to notes with the left than with the right hand.

Figure 1
www.frontiersin.org

Figure 1. (Left) Mean reaction times (with standard errors) recorded in response to the two types of stimuli in musicians and non-musicians. (Right) Mean reaction times (with standard errors) performed with the left vs. the right hand, in response to stimuli of the two types.

Electrophysiological data

Figure 2 shows the grand-average ERP waveforms recorded in response to the two types of stimuli over the left and right occipito-temporal cortex, where the N170 showed its maximum distribution, and over the midline centro-parietal and frontal areas. It can be observed the presence of consistent hemispheric asymmetries, dependent on stimulus type, which were statistically analyzed.

Figure 2
www.frontiersin.org

Figure 2. Grand-average ERP waveforms recorded at left and right posterior temporal and occipito/temporal sites, and midline cento/parietal and fronto/central sites as a function of stimulus type. Hemispheric asymmetries in favor of the left hemisphere for word processing and the right hemisphere for note processing are highlighted.

The ANOVA yielded the significance of group [F(1, 82) = 40.23, p < 0.001; ηp2 = 0.33], with much larger N170 responses recorded in musicians than in controls (see Figure 3 left, for mean and SE values). Also significant was the stimulus type factor [F(1, 82) = 61, p < 0.001; ε = 1; ηp2 = 0.44], with larger N170 to words than to notes. The interaction of group x stimulus type [F(1, 82) = 23.78, p < 0.001; ε= 1; ηp2 = 0.74], and relative post-hoc comparisons, showed significantly larger N170 to words than to notes in both groups of controls (p < 0.001) and musicians (p < 0.039). Furthermore, N170 was markedly larger in musicians than controls in response to both notation (p < 0.001), and words (p < 0.001). The ANOVA also yielded the significance of hemisphere factor [F(1, 82) = 4.5, p < 0.035; ε = 1; ηp2 = 0.05] with larger N170 amplitudes over the left than the right hemisphere (Figure 2, right). The further interaction of stimulus type x hemisphere [F(1, 82) = 24.25, p < 0.001; ε = 1; ηp2 = 0.23], and relative post-hoc comparisons, showed that while N170 was larger over the left than right hemisphere to words (p < 0.001), it tended to be larger over the right than left hemisphere to notes (p < 0.07) as shown in Figure 3 (right) and in the scatterplot distribution of individual values displayed in Figure 4.

Figure 3
www.frontiersin.org

Figure 3. (Left) Mean amplitude values of the occipito/temporal N170 response (170–210 ms) as a function of participants' musical education and stimulus type. (Right) Mean amplitude values of the occipito/temporal N170 response as a function of stimulus type and cerebral hemisphere.

Figure 4
www.frontiersin.org

Figure 4. Distribution of individual values of N170 amplitudes recorded in individual participants as a function of stimulus type and cerebral hemisphere. Note the macroscopic left lateralization of the N170 in response to words and the more subtle inverse pattern (consistent across participants) for notes.

Figure 5 displays the topographical distribution of N170 voltage response to words and notes within the 170–210 ms latency range. The left-sided distribution of the orthographic response during word processing is evident, as well as the right-sided distribution during musical notation processing. An inversion solution was applied to N170 mean voltages recorded in the 170–190 ms time range, in order to identify the inner sources of surface potentials. Table 1 provides a comprehensive list of active electromagnetic dipoles separately for word and notation processing, while the corresponding neuroimages are shown in Figure 6. During word reading, the most active source was undoubtedly the left fusiform gyrus (BA37), with a lesser activation of the contralateral counterpart. During note reading, the most active source was the right fusiform gyrus (BA19), with smaller amplitudes in the left FG BA37 and in the left inferior temporal gyrus (BA20).

Figure 5
www.frontiersin.org

Figure 5. Iso-Color topographical maps (back view) of N170 voltage recorded in the 170–210 ms time window during processing of word and musical notation.

Table 1
www.frontiersin.org

Table 1. List of active electromagnetic dipoles (along with their Talairach coordinates and relative Brodmann areas) explaining the scalp-recorded potentials measured in the 170–210 ms time windows in response to words and notes in the whole sample of participants.

Figure 6
www.frontiersin.org

Figure 6. Coronal, axial and sagittal brain sections showing the location and strength of electromagnetic dipoles explaining the surface voltage of N170 response (170–210 ms) in response to words (upper row) or notes (lower row) in the whole sample of participants. The selected sections were centered on the sources of the fusiform gyrus in both cases. The electromagnetic dipoles are shown as arrows and indicate the position, orientation and magnitude of dipole modeling solution.

Common areas of activation during reading of words and music were observed. They were the superior frontal gyrus (BA8), devoted to attention and gaze shifting; the right cingulate cortex (BA23) devoted to non-target suppression and response selection; and the superior frontal gyri (BA9 and 10), involved in working memory and selective attentional control. Interestingly, some sources were uniquely active during word reading, namely the parahippocampal area (BA35), which is active during reading tasks, the right inferior occipital gyrus (BA18) and the left parietal area (BA3), which are involved in letter shape extraction during reading. On the other hand, other sources were uniquely active during reading music, namely the left inferior temporal gyrus (BA20), involved in shape recognition and selection, and the right inferior frontal gyrus (BA44), involved in encoding musical metrics and tempo.

Discussion

This study aimed to investigate the neural mechanisms involved in visually recognizing notes and letters (in the context of music bars and words) in a large sample of right-handed individuals with varying levels of musical familiarity. We also aimed to address the seeming inconsistency of music alexia having a left hemispheric basis, despite neuroimaging data suggesting a privileged role or involvement (depending on the study) of the right occipital/temporal region in music reading. Overall, the whole sample here considered showed a right hand/left hemisphere preference for responding to target letters, and a bilaterality (with a tendency in favor of the left hand/right hemisphere) for responding to target notes. The N170 component of ERPs showed a predominantly left-sided response to words, consistent with prior electrophysiological literature (Bentin et al., 1999; Rossion et al., 2003; Proverbio et al., 2008). In contrast, the N170 response was bilateral and tended to be larger over the right hemisphere in response to music stimuli, as shown in the topographical maps of surface voltage plotted across all electrodes. LORETA source reconstruction showed how word encoding during letter selection strongly activated the left fusiform gyrus (BA37) corresponding to the putative VWFA (Bentin et al., 1999; Rossion et al., 2003; Proverbio et al., 2008; Selpien et al., 2015; Dehaene-Lambertz et al., 2018). The left mid-fusiform cortex is thought to specialize in letter and word visual recognition with the acquisition of literacy (Cohen et al., 2002; McCandliss et al., 2003; Cohen and Dehaene, 2004; Szwed et al., 2011; Glezer et al., 2015). The left lateralization (as opposed to the right one) can be attributed to this region's inclination to analyze local elements, such as high spatial frequencies and details. This is in contrast to the holistic characteristics of neurons in the homologous right visual area, i.e., the fusiform face area, which prefer to recognize holistic patterns such as face schemata (Davies-Thompson et al., 2016; Takamiya et al., 2020).

The activity observed in the right fusiform gyrus during letter selection in this study could be attributed to the fact that around half of the participants were musicians. It is recognized that musicianship may enhance the development of the right reading area for both notation and language reading, potentially explaining this phenomenon (Proverbio et al., 2013, 2024). Other regions that were uniquely active during word reading included the parahippocampal gyrus (BA35), which is strongly connected to the VWFA (van der Mark et al., 2011; Fan et al., 2014) and involved in reading (Sefcikova et al., 2022), as well as the right inferior occipital gyrus (BA18) and left parietal area (BA3), which are thought to be involved in letter shape extraction during reading (Corbetta and Shulman, 2002; Zhang et al., 2013). In addition, other areas were found to be commonly active during reading of words and music. They were the superior frontal gyrus (BA8), also known as frontal eye-fields (FEF) controlling attentional and gaze shifting (Japee et al., 2015; Wang et al., 2019); the right cingulate cortex (BA23), involved in non-target suppression and response selection (Braver et al., 2001); and the superior frontal gyrus (BA9 and BA10), involved in working memory (Yee et al., 2010) and selective attentional control (Proverbio et al., 2009; Szczepanski et al., 2013).

Note reading activated only partially overlapping circuits largely extended over the right hemisphere. The strongest focus of activation during note reading was the right posterior fusiform gyrus (BA19). This area lies in similar regions or coincides with those described in previous literature on notation reading (Sergent et al., 1992; Nakada et al., 1998; Wong and Gauthier, 2010; Li et al., 2017; Lu et al., 2021). Along with the right FG were also strongly active during note selection the left fusiform gyrus (BA37) and the left inferior temporal gyrus (BA20), possibly involved in shape recognition of notation symbols, such those indicating key, duration, sharps, flats, pauses, agogic or tonal articulation. On the other hand, right hemispheric visual areas, such as the right fusiform gyrus, would preferentially process holistic information and be activated during selective attention to global configurations (Proverbio et al., 1998). In fact, the activation of the fusiform face area (FFA) during face processing is mostly right-sided (Kanwisher et al., 1997), especially in males. Holistic face processing has been consistently demonstrated across various paradigms, such as the part-whole (Tanaka and Simonyi, 2016) and composite paradigms (Rossion, 2013). We believe that the right fusiform gyrus cell's broader receptive field could potentially learn to identify notes' pitch across the rows of the pentagram as a whole, while taking into account the global visuospatial coordinates. Consistently, bilateral fusiform gyrus has shown increased selectivity for single notes in experts (Wong and Gauthier, 2010). Interestingly Wong and Gauthier (2010) measured holistic processing of music sequences using a selective attention task in participants who varied in music reading expertise. The authors found that holistic effects were strategic in novices but relatively automatic in experts. Correlational analyses revealed that individual holistic effects were predicted by both individual music reading ability and neural responses for musical notation in the right fusiform face area (rFFA).

In the present study, other neural sources that were uniquely active during reading music and not words (at this early orthographic stage), were the left inferior temporal gyrus (BA20), involved in shape recognition and selection, and the right inferior frontal gyrus (BA44), involved in encoding musical metrics and tempo (Thaut et al., 2014).

As for the effect of musicianship (not the focus of this investigation) the data showed how musicians (N = 44) were faster in detecting letters and notes than Non-Musicians (N = 40). This finding correlates with electrophysiological data indicating how N170 was markedly larger in musicians than controls in response to both notation and words. This pattern of results fits with previous investigations showing larger N170 words in musicians than non-musicians (Proverbio et al., 2013, 2024) and providing evidence of enhanced reading skills in musicians (Gordon et al., 2015). Correlational studies conducted in adults have demonstrated that individuals who are musicians possess higher sensitivity to speech rhythm (Marie et al., 2011) and exhibit better skills related to reading (such as phoneme discrimination, as observed by Zuk et al., 2013). Whilst the advantageous effects of musical education may lead to neuroplastic changes that affect phonological processing, awareness and sensitivity of the auditory cortex, the observed effect at N170 level reflects a more proficient orthographic encoding, as N170 is believed to manifest VWFA activity. A study conducted by Proverbio et al. (2024) demonstrated that musical education promotes the development of a right-sided area responsible for coding global and spatial properties of musical notation. This area seems to correspond to the right middle occipital/fusiform gyrus, while in musicians and proficient readers, the right middle and inferior temporal cortices (BA20/BA21) nearby would also be activated during word reading. The activity of this right orthographic area would be correlated with the word reading proficiency as measured in independent reading tests (in syllables per second). Consistently Li and Hsiao (2018) showed that musicians named English words significantly faster than non-musicians when words were presented in the left visual field/right hemisphere (RH) or the center position, suggesting an advantage of RH processing due to music reading experience.

Overall, the bilateralism of notation reading mechanism might possibly explain why music dyslexia is so rare, and usually linked to word dyslexia (Brust, 1980; Fasanaro et al., 1990; Cappelletti et al., 2000; Midorikawa et al., 2003). Indeed, the presence of a double mechanism for coding fine grain details and whole spatial information related to music writing, might allow reorganization and neuroplastic compensations leading to preserved music literacy in case of unilateral lesion. Furthermore, the presence of a bilateral redundant reading mechanism, although differently specialized in the analysis of more detailed vs. spatially distributed patterns, in the left and right hemispheres, respectively, might serve as a protective measure against tackling reading disorders or dyslexia. Indeed, orthographic coding mostly engages the left VWFA in non-musicians or non-bilinguals, and studies have shown an atypical/insufficiency activity of the VWFA in surface dyslexic readers (Wilson et al., 2013; Amora et al., 2022) or in poor readers (Proverbio et al., 2024).

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Ethics statement

The studies involving humans were approved by Ethical Committee of University of Milano-Bicocca, Research Evaluation Committee of the Department of Psychology (CRIP). The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author contributions

AP: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Resources, Supervision, Writing – original draft, Writing – review & editing. GA: Data curation, Formal analysis, Investigation, Methodology, Validation, Visualization, Writing – original draft. MP: Data curation, Formal analysis, Investigation, Methodology, Validation, Visualization, Writing – original draft. AZ: Formal analysis, Methodology, Software, Visualization, Writing – original draft. MM: Data curation, Formal analysis, Investigation, Methodology, Writing – original draft.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Acknowledgments

We express gratitude to all participating students for their kind support and cooperation throughout this study.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Amora, K. K., Tretow, A., Verwimp, C., Tijms, J., Leppänen, P. H. T., Csépe, V., et al. (2022). Typical and atypical development of visual expertise for print as indexed by the visual word N1 (N170w): a systematic review. Front. Neurosci. 16:898800. doi: 10.3389/fnins.2022.898800

Crossref Full Text | Google Scholar

Bentin, S., Mouchetant-Rostaing, Y., Giard, M. H., Echallier, J. F., and Pernier, J. (1999). ERP manifestations of processing printed words at different psycholinguistic levels: time course and scalp distribution. J. Cogn. Neurosci. 11, 235–260. doi: 10.1162/089892999563373

PubMed Abstract | Crossref Full Text | Google Scholar

Bouhali, F., de Thiebaut Schotten, M., Pinel, P., Poupon, C., Mangin, J. F., Dehaene, S., et al. (2014). Anatomical connections of the visual word form area. J. Neurosci. 34, 15402–15414. doi: 10.1523/JNEUROSCI.4918-13.2014

PubMed Abstract | Crossref Full Text | Google Scholar

Braver, T. S., Barch, D. M., Gray, J. R., Molfese, D. L., and Snyder, A. (2001). Anterior cingulate cortex and response conflict: effects of frequency, inhibition and errors. Cereb. Cortex. 11, 825–836. doi: 10.1093/cercor/11.9.825

PubMed Abstract | Crossref Full Text | Google Scholar

Bruno, J. L., Zumberge, A., Manis, F. R., Lu, Z. L., and Goldman, J. G. (2008). Sensitivity to orthographic familiarity in the occipito-temporal region. Neuroimage 39, 1988–2001. doi: 10.1016/j.neuroimage.2007.10.044

PubMed Abstract | Crossref Full Text | Google Scholar

Brust, J. C. (1980). Music and language: musical alexia and agraphia. Brain 103, 367–392. doi: 10.1093/brain/103.2.367

PubMed Abstract | Crossref Full Text | Google Scholar

Cappelletti, M., Waley-Cohen, H., Butterworth, B., and Kopelman, M. A. (2000). selective loss of the ability to read and to write music. Neurocase 6, 321–332. doi: 10.1080/13554790008402780

Crossref Full Text | Google Scholar

Cnudde, K., van Hees, S., Brown, S., van der Wijk, G., Pexman, P. M., Protzner, A. B., et al. (2021). Increased neural efficiency in visual word recognition: evidence from alterations in event-related potentials and multiscale entropy. Entropy 23:304. doi: 10.3390/e23030304

PubMed Abstract | Crossref Full Text | Google Scholar

Cohen, L., and Dehaene, S. (2004). Specialization within the ventral stream: the case for the visual word form area. Neuroimage 22, 466–476. doi: 10.1016/j.neuroimage.2003.12.049

PubMed Abstract | Crossref Full Text | Google Scholar

Cohen, L., Dehaene, S., Naccache, L., Lehéricy, S., Dehaene-Lambertz, G., Hénaff, M. A., et al. (2000). The visual word form area: spatial and temporal characterization of an initial stage of reading in normal subjects and posterior split-brain patients. Brain 123, 291–307. doi: 10.1093/brain/123.2.291

PubMed Abstract | Crossref Full Text | Google Scholar

Cohen, L., Lehéricy, S., Chochon, F., Lemer, C., Rivaud, S., Dehaene, S., et al. (2002). Language-specific tuning of visual cortex? Functional properties of the Visual Word Form Area. Brain 125, 1054–1069. doi: 10.1093/brain/awf094

PubMed Abstract | Crossref Full Text | Google Scholar

Cohen, L., Martinaud, O., Lemer, C., Lehéricy, S., Samson, Y., Obadia, M., et al. (2003). Visual word recognition in the left and right hemispheres: anatomical and functional correlates of peripheral alexias. Cereb. Cortex. 13, 1313–1333. doi: 10.1093/cercor/bhg079

PubMed Abstract | Crossref Full Text | Google Scholar

Corbetta, M., and Shulman, G. L. (2002). Control of goal-directed and stimulus-driven attention in the brain. Nat. Rev. Neurosci. 3, 201–215. doi: 10.1038/nrn755

PubMed Abstract | Crossref Full Text | Google Scholar

Davies-Thompson, J., Johnston, S., Tashakkor, Y., Pancaroglu, R., and Barton, J. J. (2016). The relationship between visual word and face processing lateralization in the fusiform gyri: a cross-sectional study. Brain Res. 1644, 88–97. doi: 10.1016/j.brainres.2016.05.009

PubMed Abstract | Crossref Full Text | Google Scholar

Dehaene, S., Le Clec'H, G., Poline, J. B., Le Bihan, D., and Cohen, L. (2002). The visual word form area: a prelexical representation of visual words in the fusiform gyrus. Neuroreport 13, 321–325. doi: 10.1097/00001756-200203040-00015

PubMed Abstract | Crossref Full Text | Google Scholar

Dehaene-Lambertz, G., Monzalvo, K., and Dehaene, S. (2018). The emergence of the visual word form: longitudinal evolution of category-specific ventral visual areas during reading acquisition. PLoS Biol. 16:e2004103. doi: 10.1371/journal.pbio.2004103

PubMed Abstract | Crossref Full Text | Google Scholar

Devlin, J. T., Jamison, H. L., Gonnerman, L. M., and Matthews, P. M. (2006). The role of the posterior fusiform gyrus in reading. J. Cogn. Neurosci. 18, 911–922. doi: 10.1162/jocn.2006.18.6.911

PubMed Abstract | Crossref Full Text | Google Scholar

Fan, Q., Anderson, A. W., Davis, N., and Cutting, L. E. (2014). Structural connectivity patterns associated with the putative visual word form area and children's reading ability. Brain Res. 1586, 118–129. doi: 10.1016/j.brainres.2014.08.050

PubMed Abstract | Crossref Full Text | Google Scholar

Fasanaro, A. M., Spitaleri, D. L. A., and Valiani, R. (1990). Dissociation in musical reading: a musician affected by alexia without agraphia. Music Percept. 7, 259–272. doi: 10.2307/40285464

Crossref Full Text | Google Scholar

Flowers, D. L., Jones, K., Noble, K., VanMeter, J., Zeffiro, T. A., Wood, F. B., et al. (2004). Attention to single letters activates left extrastriate cortex. Neuroimage 21, 829–839. doi: 10.1016/j.neuroimage.2003.10.002

PubMed Abstract | Crossref Full Text | Google Scholar

Garrett, A. S., Flowers, D. L., Absher, J. R., Fahey, F. H., Gage, H. D., Keyes, J. W., et al. (2000). Cortical activity related to accuracy of letter recognition. Neuroimage 11, 111–123. doi: 10.1006/nimg.1999.0528

PubMed Abstract | Crossref Full Text | Google Scholar

Glezer, L. S., Kim, J., Rule, J., Jiang, X., and Riesenhuber, M. (2015). Adding words to the brain's visual dictionary: novel word learning selectively sharpens orthographic representations in the VWFA. J. Neurosci. 35, 4965–4972. doi: 10.1523/JNEUROSCI.4031-14.2015

PubMed Abstract | Crossref Full Text | Google Scholar

Gordon, R. L., Fehd, H. M., and McCandliss, B. D. (2015). Does music training enhance literacy skills? A meta-analysis. Front. Psychol. 6:1777. doi: 10.3389/fpsyg.2015.01777

PubMed Abstract | Crossref Full Text | Google Scholar

Horikoshi, T., Asari, Y., Watanabe, A., Nagaseki, Y., Nukui, H., Sasaki, H., et al. (1997). Music alexia in a patient with mild pure alexia: disturbed visual perception of nonverbal meaningful figures. Cortex 33, 187–194. doi: 10.1016/S0010-9452(97)80014-7

PubMed Abstract | Crossref Full Text | Google Scholar

James, K. H., James, T. W., Jobard, G., Wong, A. C., and Gauthier, I. (2005). Letter processing in the visual system: different activation patterns for single letters and strings. Cogn. Affect. Behav. Neurosci. 5, 452–466. doi: 10.3758/CABN.5.4.452

PubMed Abstract | Crossref Full Text | Google Scholar

Japee, S., Holiday, K., Satyshur, M. D., Mukai, I., and Ungerleider, L. G. A. (2015). role of right middle frontal gyrus in reorienting of attention: a case study. Front. Syst. Neurosci. 9:23. doi: 10.3389/fnsys.2015.00023

PubMed Abstract | Crossref Full Text | Google Scholar

Kanwisher, N., McDermott, J., and Chun, M. M. (1997). The fusiform face area: a module in human extrastriate cortex specialized for face perception. J. Neurosci. 17, 4302–4311. doi: 10.1523/JNEUROSCI.17-11-04302.1997

PubMed Abstract | Crossref Full Text | Google Scholar

Kawamura, M., Midorikawa, A., and Kezuka, M. (2000). Cerebral localization of the center for reading and writing music. Neuroreport 11, 3299–3303. doi: 10.1097/00001756-200009280-00050

PubMed Abstract | Crossref Full Text | Google Scholar

Kronbichler, M., Bergmann, J., Hutzler, F., Staffen, W., Mair, A., Ladurner, G., et al. (2007). Taxi vs. taksi: on orthographic word recognition in the left ventral occipitotemporal cortex. J. Cogn. Neurosci. 19, 1584–1594. doi: 10.1162/jocn.2007.19.10.1584

PubMed Abstract | Crossref Full Text | Google Scholar

Kronbichler, M., Hutzler, F., Wimmer, H., Mair, A., Staffen, W., Ladurner, G., et al. (2004). The visual word form area and the frequency with which words are encountered: evidence from a parametric fMRI study. Neuroimage 21, 946–953. doi: 10.1016/j.neuroimage.2003.10.021

PubMed Abstract | Crossref Full Text | Google Scholar

Kuriki, S., Takeuchi, F., and Hirata, Y. (1998). Neural processing of words in the human extrastriate visual cortex. Brain Res. Cogn. Brain Res. 6, 193–203. doi: 10.1016/S0926-6410(97)00030-X

PubMed Abstract | Crossref Full Text | Google Scholar

Li, S. T. K., and Hsiao, J. H. (2018). Music reading expertise modulates hemispheric lateralization in English word processing but not in Chinese character processing. Cognition 176, 159–173. doi: 10.1016/j.cognition.2018.03.010

PubMed Abstract | Crossref Full Text | Google Scholar

Li, T. K., Chan, H. Y. V., Li, L., and Hsiao, J. H. W. (2017). “How does music reading expertise modulate visual processing of English words? An ERP study,” in Annual Conference of the Cognitive Science Society 2017. Cognitive Science Society.

Google Scholar

Lu, C. I., Greenwald, M. L., Lin, Y. Y., and Bowyer, S. M. (2019). Reading musical notation versus English letters: mapping brain activation with MEG. Psychol. Music 47, 255–269. doi: 10.1177/0305735617744886

Crossref Full Text | Google Scholar

Lu, C. I., Greenwald, M. L., Lin, Y. Y., and Bowyer, S. M. (2021). Musical transposing versus sight-reading: mapping brain activation with magnetoencephalography. Psychol. Music 49, 581–599. doi: 10.1177/0305735619883692

Crossref Full Text | Google Scholar

Marie, C., Magne, C., and Besson, M. (2011). Musicians and the metric structure of words. J. Cogn. Neurosci. 23, 294–305. doi: 10.1162/jocn.2010.21413

PubMed Abstract | Crossref Full Text | Google Scholar

McCandliss, B. D., Cohen, L., and Dehaene, S. (2003). The visual word form area: expertise for reading in the fusiform gyrus. Trends Cogn. Sci. 7, 293–299. doi: 10.1016/S1364-6613(03)00134-7

PubMed Abstract | Crossref Full Text | Google Scholar

Mechelli, A., Gorno-Tempini, M. L., and Price, C. J. (2003). Neuroimaging studies of word and pseudoword reading: consistencies, inconsistencies, and limitations. J. Cogn. Neurosci. 15, 260–271. doi: 10.1162/089892903321208196

PubMed Abstract | Crossref Full Text | Google Scholar

Meister, I. G., Krings, T., Foltys, H., Boroojerdi, B., Müller, M., Töpper, R., et al. (2004). Playing piano in the mind–an fMRI study on music imagery and performance in pianists. Brain Res. Cogn. Brain Res. 19, 219–228. doi: 10.1016/j.cogbrainres.2003.12.005

PubMed Abstract | Crossref Full Text | Google Scholar

Midorikawa, A., Kawamura, M., and Kezuka, M. (2003). Musical alexia for rhythm notation: a discrepancy between pitch and rhythm. Neurocase 9, 232–238. doi: 10.1076/neur.9.3.232.15558

PubMed Abstract | Crossref Full Text | Google Scholar

Mongelli, V., Dehaene, S., Vinckier, F., Peretz, I., Bartolomeo, P., Cohen, L., et al. (2017). Music and words in the visual cortex: the impact of musical expertise. Cortex 86, 260–274. doi: 10.1016/j.cortex.2016.05.016

PubMed Abstract | Crossref Full Text | Google Scholar

Nakada, T., Fujii, Y., Suzuki, K., and Kwee, I. L. (1998). 'Musical brain' revealed by high-field (3 Tesla) functional MRI. Neuroreport 9, 3853–3856. doi: 10.1097/00001756-199812010-00016

PubMed Abstract | Crossref Full Text | Google Scholar

Oldfield, R. C. (1971). The assessment and analysis of handedness: the Edinburgh inventory. Neuropsychologia 9, 97–113. doi: 10.1016/0028-3932(71)90067-4

PubMed Abstract | Crossref Full Text | Google Scholar

Palmero-Soler, E., Dolan, K., Hadamschek, V., and Tass, P. A. (2007). swLORETA: a novel approach to robust source localization and synchronization tomography. Phys. Med. Biol. 52, 1783–1800. doi: 10.1088/0031-9155/52/7/002

PubMed Abstract | Crossref Full Text | Google Scholar

Pascual-Marqui, R. D. (2002). Standardized low-resolution brain electromagnetic tomography (sLORETA): technical details. Methods Find Exp. Clin. Pharmacol. 24(Suppl.D), 5–12.

PubMed Abstract | Google Scholar

Peretz, I., and Zatorre, R. J. (2005). Brain organization for music processing. Annu. Rev. Psychol. 56, 89–114. doi: 10.1146/annurev.psych.56.091103.070225

PubMed Abstract | Crossref Full Text | Google Scholar

Pernet, C., Celsis, P., and Démonet, J. F. (2005). Selective response to letter categorization within the left fusiform gyrus. Neuroimage 28, 738–744. doi: 10.1016/j.neuroimage.2005.06.046

PubMed Abstract | Crossref Full Text | Google Scholar

Polk, T. A., Stallcup, M., Aguirre, G. K., Alsop, D. C., D'Esposito, M., Detre, J. A., et al. (2002). Neural specialization for letter recognition. J. Cogn. Neurosci. 14, 145–159. doi: 10.1162/089892902317236803

Crossref Full Text | Google Scholar

Proverbio, A. M., Calbi, M., Manfredi, M., and Zani, A. (2014). Audio-visuomotor processing in the musician's brain: an ERP study on professional violinists and clarinetists. Sci. Rep. 4:5866. doi: 10.1038/srep05866

PubMed Abstract | Crossref Full Text | Google Scholar

Proverbio, A. M., Del Zotto, M., Crotti, N., and Zani, A. (2009). A no-go related prefrontal negativity larger to irrelevant stimuli that are difficult to suppress. Behav. Brain Funct. 5:25. doi: 10.1186/1744-9081-5-25

PubMed Abstract | Crossref Full Text | Google Scholar

Proverbio, A. M., Manfredi, M., Zani, A., and Adorni, R. (2013). Musical expertise affects neural bases of letter recognition. Neuropsychologia 51, 538–549. doi: 10.1016/j.neuropsychologia.2012.12.001

PubMed Abstract | Crossref Full Text | Google Scholar

Proverbio, A. M., Minniti, A., and Zani, A. (1998). Electrophysiological evidence of a perceptual precedence of global vs. local visual information. Brain Res. Cogn. Brain Res. 6, 321–334. doi: 10.1016/S0926-6410(97)00039-6

PubMed Abstract | Crossref Full Text | Google Scholar

Proverbio, A. M., Pantaleo, M. M., Arcuri, G., and Manfredi, M. (2024). Neuroplastic effects of musical notation literacy on reading skills: bilateral word processing predicts enhanced proficiency. Res. Square. doi: 10.21203/rs.3.rs-3480980/v1

Crossref Full Text | Google Scholar

Proverbio, A. M., Zani, A., and Adorni, R. (2008). The left fusiform area is affected by written frequency of words. Neuropsychologia 46, 2292–2299. doi: 10.1016/j.neuropsychologia.2008.03.024

PubMed Abstract | Crossref Full Text | Google Scholar

Rossion, B. (2013). The composite face illusion: a whole window into our understanding of holistic face perception. Vis. Cogn. 21, 139–253. doi: 10.1080/13506285.2013.772929

Crossref Full Text | Google Scholar

Rossion, B., Joyce, C. A., Cottrell, G. W., and Tarr, M. J. (2003). Early lateralization and orientation tuning for face, word, and object processing in the visual cortex. Neuroimage 20, 1609–1624. doi: 10.1016/j.neuroimage.2003.07.010

PubMed Abstract | Crossref Full Text | Google Scholar

Roux, F. E., Lubrano, V., Lotterie, J. A., Giussani, C., Pierroux, C., Démonet, J. F., et al. (2007). When “abegg” is read and (“A, B, E, G, G”) is not: a cortical stimulation study of musical score reading. J. Neurosurg. 106, 1017–1027. doi: 10.3171/jns.2007.106.6.1017

PubMed Abstract | Crossref Full Text | Google Scholar

Schön, D., Semenza, C., and Denes, G. (2001). Naming of musical notes: a selective deficit in one musical clef. Cortex 37, 407–421. doi: 10.1016/S0010-9452(08)70581-1

PubMed Abstract | Crossref Full Text | Google Scholar

Sefcikova, V., Sporrer, J. K., Juvekar, P., Golby, A., and Samandouras, G. (2022). Converting sounds to meaning with ventral semantic language networks: integration of interdisciplinary data on brain connectivity, direct electrical stimulation and clinical disconnection syndromes. Brain Struct. Funct. 227, 1545–1564. doi: 10.1007/s00429-021-02438-x

PubMed Abstract | Crossref Full Text | Google Scholar

Selpien, H., Siebert, C., Genc, E., Beste, C., Faustmann, P. M., Güntürkün, O., et al. (2015). Left dominance for language perception starts in the extrastriate cortex: an ERP and sLORETA study. Behav. Brain Res. 291, 325–333. doi: 10.1016/j.bbr.2015.05.050

PubMed Abstract | Crossref Full Text | Google Scholar

Sergent, J., Zuck, E., Terriah, S., and MacDonald, B. (1992). Distributed neural network underlying musical sight-reading and keyboard performance. Science 257, 106–109. doi: 10.1126/science.1621084

PubMed Abstract | Crossref Full Text | Google Scholar

Starrfelt, R., and Wong, Y. K. (2017). Musical notation reading in pure alexia. J. Vis. 17, 1038–1038. doi: 10.1167/17.10.1038

Crossref Full Text | Google Scholar

Stewart, L., Henson, R., Kampe, K., Walsh, V., Turner, R., Frith, U., et al. (2003). Becoming a pianist. An fMRI study of musical literacy acquisition. Ann. N. Y. Acad. Sci. 999, 204–208. doi: 10.1196/annals.1284.030

PubMed Abstract | Crossref Full Text | Google Scholar

Szczepanski, S. M., Pinsk, M. A., Douglas, M. M., Kastner, S., and Saalmann, Y. B. (2013). Functional and structural architecture of the human dorsal frontoparietal attention network. Proc. Natl. Acad. Sci. U. S. A. 110, 15806–15811. doi: 10.1073/pnas.1313903110

PubMed Abstract | Crossref Full Text | Google Scholar

Szwed, M., Dehaene, S., Kleinschmidt, A., Eger, E., Valabrègue, R., Amadon, A., et al. (2011). Specialization for written words over objects in the visual cortex. Neuroimage 56, 330–344. doi: 10.1016/j.neuroimage.2011.01.073

PubMed Abstract | Crossref Full Text | Google Scholar

Takamiya, N., Maekawa, T., Yamasaki, T., Ogata, K., Yamada, E., Tanaka, M., et al. (2020). Different hemispheric specialization for face/word recognition: a high-density ERP study with hemifield visual stimulation. Brain Behav. 10:e01649. doi: 10.1002/brb3.1649

PubMed Abstract | Crossref Full Text | Google Scholar

Tanaka, J. W., and Simonyi, D. (2016). The “parts and wholes” of face recognition: a review of the literature. Q. J. Exp. Psychol. 69, 1876–1889. doi: 10.1080/17470218.2016.1146780

Crossref Full Text | Google Scholar

Thaut, M. H., Trimarchi, P. D., and Parsons, L. M. (2014). Human brain basis of musical rhythm perception: common and distinct neural substrates for meter, tempo, and pattern. Brain Sci. 4, 428–452. doi: 10.3390/brainsci4020428

PubMed Abstract | Crossref Full Text | Google Scholar

Thesen, T., McDonald, C. R., Carlson, C., Doyle, W., Cash, S., Sherfey, J., et al. (2012). Sequential then interactive processing of letters and words in the left fusiform gyrus. Nat. Commun. 3:1284. doi: 10.1038/ncomms2220

PubMed Abstract | Crossref Full Text | Google Scholar

van der Mark, S., Klaver, P., Bucher, K., Maurer, U., Schulz, E., Brem, S., et al. (2011). The left occipitotemporal system in reading: disruption of focal fMRI connectivity to left inferior frontal and inferior parietal language areas in children with dyslexia. Neuroimage 54, 2426–2436. doi: 10.1016/j.neuroimage.2010.10.002

PubMed Abstract | Crossref Full Text | Google Scholar

Vigneau, M., Jobard, G., Mazoyer, B., and Tzourio-Mazoyer, N. (2005). Word and non-word reading: what role for the Visual Word Form Area? Neuroimage 27, 694–705. doi: 10.1016/j.neuroimage.2005.04.038

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, K., Leopold, D. R., Banich, M. T., Reineberg, A. E., Willcutt, E. G., Cutting, L. E., et al. (2019). Characterizing and decomposing the neural correlates of individual differences in reading ability among adolescents with task-based fMRI. Dev. Cogn. Neurosci. 37:100647. doi: 10.1016/j.dcn.2019.100647

PubMed Abstract | Crossref Full Text | Google Scholar

Wilson, S. M., Rising, K., Stib, M. T., Rapcsak, S. Z., and Beeson, P. M. (2013). Dysfunctional visual word form processing in progressive alexia. Brain 136, 1260–1273. doi: 10.1093/brain/awt034

PubMed Abstract | Crossref Full Text | Google Scholar

Wong, Y. K., and Gauthier, I. (2010). Holistic processing of musical notation: dissociating failures of selective attention in experts and novices. Cogn. Affect. Behav. Neurosci. 10, 541–551. doi: 10.3758/CABN.10.4.541

PubMed Abstract | Crossref Full Text | Google Scholar

Yee, L. T., Roe, K., and Courtney, S. M. (2010). Selective involvement of superior frontal cortex during working memory for shapes. J. Neurophysiol. 103, 557–563. doi: 10.1152/jn.91299.2008

PubMed Abstract | Crossref Full Text | Google Scholar

Zanow, F., and Knösche, T. R. (2004). ASA–Advanced Source Analysis of continuous and event-related EEG/MEG signals. Brain Topogr. 16, 287–290. doi: 10.1023/B:BRAT.0000032867.41555.d0

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, Y., Whitfield-Gabrieli, S., Christodoulou, J. A., and Gabrieli, J. D. (2013). Atypical balance between occipital and fronto-parietal activation for visual shape extraction in dyslexia. PLoS ONE 8:e67331. doi: 10.1371/journal.pone.0067331

PubMed Abstract | Crossref Full Text | Google Scholar

Zuk, J., Ozernov-Palchik, O., Kim, H., Lakshminarayanan, K., Gabrieli, J. D., Tallal, P., et al. (2013). Enhanced syllable discrimination thresholds in musicians. PLoS ONE 8:e80546. doi: 10.1371/journal.pone.0080546

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: fusiform gyrus (FG), VWFA, reading, language, music, N170, alexia, dyslexia

Citation: Proverbio AM, Arcuri G, Pantaleo MM, Zani A and Manfredi M (2024) The key role of the right posterior fusiform gyrus in music reading: an electrical neuroimaging study on 90 readers. Front. Cognit. 3:1323220. doi: 10.3389/fcogn.2024.1323220

Received: 21 October 2023; Accepted: 03 January 2024;
Published: 17 January 2024.

Edited by:

Claude Alain, Rotman Research Institute (RRI), Canada

Reviewed by:

Fang Wang, Stanford University, United States
Luca Bischetti, University Institute of Higher Studies in Pavia, Italy

Copyright © 2024 Proverbio, Arcuri, Pantaleo, Zani and Manfredi. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Alice Mado Proverbio, bWFkby5wcm92ZXJiaW8mI3gwMDA0MDt1bmltaWIuaXQ=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.