Skip to main content

ORIGINAL RESEARCH article

Front. Hum. Neurosci., 21 October 2015
Sec. Speech and Language

Identifying environmental sounds: a multimodal mapping study

  • 1Istituto di Ricovero e Cura a Carattere Scientifico “E. Medea”, Polo Regionale del Friuli Venezia Giulia, Udine, Italy
  • 2Fisica Medica A.O.S. Maria della Misericordia, Udine, Italy
  • 3Unità Operativa di Neuroradiologia, A.O.S. Maria della Misericordia, Udine, Italy
  • 4Unità Operativa di Neurochirurgia, A.O.S. Maria della Misericordia, Udine, Italy

Our environment is full of auditory events such as warnings or hazards, and their correct recognition is essential. We explored environmental sounds (ES) recognition in a series of studies. In study 1 we performed an Activation Likelihood Estimation (ALE) meta-analysis of neuroimaging experiments addressing ES processing to delineate the network of areas consistently involved in ES processing. Areas consistently activated in the ALE meta-analysis were the STG/MTG, insula/rolandic operculum, parahippocampal gyrus and inferior frontal gyrus bilaterally. Some of these areas truly reflect ES processing, whereas others are related to design choices, e.g., type of task, type of control condition, type of stimulus. In study 2 we report on 7 neurosurgical patients with lesions involving the areas which were found to be activated by the ALE meta-analysis. We tested their ES recognition abilities and found an impairment of ES recognition. These results indicate that deficits of ES recognition do not exclusively reflect lesions to the right or to the left hemisphere but both hemispheres are involved. The most frequently lesioned area is the hippocampus/insula/STG. We made sure that any impairment in ES recognition would not be related to language problems, but reflect impaired ES processing. In study 3 we carried out an fMRI study on patients (vs. healthy controls) to investigate how the areas involved in ES might be functionally deregulated because of a lesion. The fMRI evidenced that controls activated the right IFG, the STG bilaterally and the left insula. We applied a multimodal mapping approach and found that, although the meta-analysis showed that part of the left and right STG/MTG activation during ES processing might in part be related to design choices, this area was one of the most frequently lesioned areas in our patients, thus highlighting its causal role in ES processing. We found that the ROIs we drew on the two clusters of activation found in the left and in the right STG overlapped with the lesions of at least 4 out of the 7 patients' lesions, indicating that the lack of STG activation found for patients is related to brain damage and is crucial for explaining the ES deficit.

Introduction

Sound recognition such as a telephone ringing or a dog barking seems such an effortless task. The ability to process environmental sounds (ES) such as warnings (e.g., a siren), threats (e.g., a rattlesnake), recognize when a device is functioning correctly (e.g., clicking of a stapler) or incorrectly (e.g., water dripping), locate an event in space (e.g., an explosion), monitor a change in status (e.g., chiming of a cuckoo clock), communicate an emotional (e.g., scream), or physical condition (e.g., a burp) (Marcell et al., 2000) is essential for everyday life.

The impaired capacity to recognize sounds despite adequate speech comprehension and hearing is defined as auditory agnosia. Auditory agnosia is a rare neuropsychological disorder; the literature about auditory agnosia mainly consists of case studies. The lesions related to this disorder are not particularly consistent as they can involve the temporal or temporo-parietal cortex (Vignolo, 1982; Fujii et al., 1990; Schnider et al., 1994; Engelien et al., 1995; Clarke et al., 2000, 2002; Saygin et al., 2003), subcortical areas (Kazui et al., 1990) such as the thalamus (Clarke et al., 2000) as well as the putamen (Taniwaki et al., 2000), the right hemisphere (e.g., Vignolo, 1982; Fujii et al., 1990; Schnider et al., 1994; Clarke et al., 1996), the left hemisphere (e.g., Vignolo, 1982; Schnider et al., 1994; Clarke et al., 1996, 2000), and both hemispheres (Rosati et al., 1982; Vignolo, 1982; Mendez and Geehan, 1988; Engelien et al., 1995; Clarke et al., 1996; Nové-Josserand et al., 1998). Left hemisphere (and bilateral) lesions tend to produce additional deficits in verbal comprehension. Thus, so far no precise anatomical locations have been correlated with auditory agnosia (Lewis et al., 2004) and knowledge of the brain regions and processing pathways that make up the non-verbal sound recognition system is still fragmentary (Lewis et al., 2004) as many regions have been found to be involved in such processing.

The network of areas involved in ES processing has been investigated also by functional imaging studies (see, for example, the ES processing model by Lewis et al., 2004). In particular, activations in the posterior middle temporal gyrus (MTG) as well as areas like the inferior frontal gyrus (IFG) (Lewis et al., 2004), which is anatomically connected with the auditory cortex (Hackett et al., 1999; Romanski et al., 1999a,b; Romanski and Goldman-Rakic, 2002) have been reported. Areas of activation were found in the MTG and the precuneus bilaterally and in the posterior portion of the left IFG—activation in this area was higher for sound recognition vs. sound localization (Maeder et al., 2001). In addition, activation can be found in the insula (Sharda and Singh, 2012), an area which has numerous direct connections with the auditory cortex (Bamiou et al., 2003) and can cause auditory agnosia if lesioned bilaterally (Engelien et al., 1995). Also the parahippocampal gyri (Sharda and Singh, 2012) can be activated by sound recognition, possibly reflecting the “imageability” of ES sounds (Engel et al., 2009). Lastly, sound activations were found in various subcortical regions like the thalamus (Sharda and Singh, 2012)—which is part of the auditory pathway—as well as the caudate and putamen. Some studies suggested that activation is rather right lateralized, especially in the (non-primary) auditory cortex such as the superior temporal gyrus (STG) (Bergerbest et al., 2004) and the inferior prefrontal cortex (Bergerbest et al., 2004). A PET study (Zatorre et al., 1992) showed that cerebral blood flow (CBF) in the inferior prefrontal cortex depends on the type of cognitive operation involved in ES, see also (Specht and Reul, 2003). In a similar vein, some authors (Thierry et al., 2003) proposed that the connectivity to the left lateralized semantic network is primarily right-sided for ES and left-sided for words. Others (Specht and Reul, 2003) argued that activation in the right STG and superior temporal sulcus (STS) plays the same crucial role in the analysis of non-speech sounds as does the left STS in speech perception (Specht and Reul, 2003). Last, Dick et al. (2007) found that language and ES stimuli evoked very similar volumes of activation in their language-based regions of interest in the left hemisphere, whereas they found greater activation for ES stimuli in the right hemisphere. Studies also showed that the areas involved in ES processing can be modulated by type of stimulus. Like semantic processing, ES recognition is characterized by category specificity: vocalizations (Fecteau et al., 2004; Lewis et al., 2009; Rauschecker and Scott, 2009; Staeren et al., 2009; Leaver and Rauschecker, 2010) and human-produced action sounds (Lewis, 2006; Lewis et al., 2006; Altmann et al., 2007) are different classes of stimuli triggering different activations. To sum up, inconsistencies as to the areas involved in the network supporting ES processing are found across studies, similarly to what emerges from the analysis of the neuropsychological data reported above.

Considering these discrepancies, the aim of our study was to investigate which nodes of the network triggered by ES processing are essential for ES processing and which areas are accessory. We thus compared neuroimaging and neuropsychological data between ES patients and normal controls. In addition, being a correlation-based method, fMRI can delineate brain networks engaged in ES processing; critical mechanisms can only be reported by studying patients with a deficit in ES following a brain damage. We first performed an Activation Likelihood Estimation (ALE) Meta-Analysis to investigate the regions which were found to be consistently activated in neuroimaging studies of ES recognition. In study 2 we selected neurosurgical patients with lesions involving these areas and tested their ES abilities. In study 3 we performed an fMRI study to understand how the key areas involved in ES recognition are functionally deregulated as compared to those of control subjects. In our patients we investigated which parts of the network involved in ES processing that were found to be activated by the meta-analysis of fMRI studies are critically involved in the task. We also tested which part of the ES processing network which was found to be activated by the meta-analysis would be deregulated in the patients' fMRI maps. We acknowledge that our sample size is relatively small, however it is known that auditory agnosia for ES is a rare neuropsychological disorder. Previously published neuropsychological reports focused mainly on single cases, with few exceptions of group studies. Furthermore, at variance with previous studies, our patient sample had selective and relatively small lesions in comparison to patients with stroke lesions which typically involve large parts of the cortex or to patients with neurodegenerative disorders which affect multiple areas.

Materials and Methods

We performed three consecutive studies. In Study 1 we used an Activation Likelihood Estimation (ALE) Meta-Analysis to identify the areas which are consistently activated in neuroimaging studies of ES processing. In Study 2, we report on neurosurgical patients who had a lesion involving the areas revealed by the ALE Meta-Analysis and were impaired at ES recognition. In Study 3 we compared the fMRI maps of patients and healthy controls to understand how these areas might be functionally deregulated because of the lesion.

Study 1: Activation Likelihood Estimation (ALE) Meta-analysis of Neuroimaging Studies of Environmental Sound Recognition

Data Used for the Meta-analysis

The functional imaging studies included in this meta-analysis were obtained from a comprehensive PubMed, ISI web of knowledge and Cochrane database literature review focusing on ES recognition (search strings: “ES,” “fMRI,” “PET,” “SPECT”). The references of the retrieved articles were screened in order to identify additional articles dealing with the neural correlates of ES recognition. Inclusion criteria were as follows: neurologically healthy adults and experiments requiring participants to process ES during fMRI/PET/SPECT measurements. The foci employed in each study had to be reported in a standard reference space (Talairach/Tournoux, MNI). Differences in coordinate spaces (MNI vs. Talairach space) were accounted for by transforming coordinates reported in Talairach space into MNI coordinates using a linear transformation model (Lancaster et al., 2007). A random-effects analysis was performed, and single-subject reports were excluded. Table 1 reports the significance level of the reported activations. All studies reported activations surviving corrections for multiple comparisons except some (study 2, 16, 19, 20, 22, 27, and 37).

TABLE 1
www.frontiersin.org

Table 1. Publications included in the meta-analysis.

Based on these criteria, data from a total of 25 articles (including 22 fMRI and 3 PET studies) were entered into the study (see Table 1). In total, 37 experiments, i.e., lists of activation foci, were included in the first meta-analysis because 8 studies reported coordinates for more than one contrast. In this case, all contrasts were included in the analysis, since all of them reflected ES-related activations, e.g., in Lewis et al. (2006), coordinates from two contrasts (tools vs. animal and animal vs. tools) were reported, and we included both. Please note that this is a common procedure as can be found in previous ALE meta-analyses (for instance in Caspers et al., 2010; Tomasino et al., 2012, 2014). Taken together, the meta-analysis included data from 263 subjects and 627 activation foci.

Statistical Procedure

A statistical map was generated using lists of x, y, and z coordinates after transferring these foci into MNI space (Lancaster et al., 2007). The meta-analysis was completed using the revised version (Eickhoff et al., 2009, 2012) of the GingerALE 2.1.1 software (brainmap.org) for coordinate-based meta-analysis of neuro-imaging results (Turkeltaub et al., 2002; Laird et al., 2005, 2009). Using the False Discovery Rate (FDR) with q = 0.01, the test was corrected for multiple comparisons (Laird et al., 2005, 2009; Eickhoff et al., 2009, 2012), and a minimum cluster size of 100 mm3 was set. The resulting areas were anatomically labeled by reference to probabilistic cytoarchitectonic maps of the human brain using the SPM Anatomy Toolbox (Eickhoff et al., 2005). Using a Maximum Probability Map (MPM), activations were assigned to the most probable histological area at their respective locations.

In meta-analysis 1, we identified the neural regions that were found to be consistently activated when listening to ES across multiple studies. In meta-analyses 2–5 (see below) we investigated how design choices might influence the activation observed in the list of the fMRI studies we evaluated.

In meta-analysis 2 we investigated first the effect of the type of stimulus. Many of the included studies reported contrasts related to different types of ES mixed together, some reported activations related to a specific category of stimuli [Action-related stimuli (Action, tools, human-related stimuli): Experiments n. 3, 5, 10, 12, 18, 19, 21, 25, 30, 33; Animal-related stimuli: Experiments n. 2, 4, 6, 11, 13, 20, 23, 28, 29, 34]. We directly contrasted the two types of stimuli.

In meta-analysis 3 we addressed the effect of the type of control sound. Some studies compared ES stimuli to silent stimuli, rest or fixation conditions [Experiments n. 1, 16–19, 23–25, 31, 32, 37], others compared ES stimuli to other control auditory sounds [Experiments n. 2–15, 20–22, 26–30, 33–36]. We directly contrasted the two types of control conditions. Using silent events vs. control sounds, we compared studies in which the active task (sound recognition) was compared to silent events (or resting) as control condition vs. studies in which the active task (sound recognition) was compared to control sounds (thus other sounds) as control condition.

In meta-analysis 4 we addressed the weight of the type of task: making a category judgment [Experiments n. 3–15, 18–27, 31–37] compared to passive listening [Experiments n. 1, 2, 16, 17, 28–30]. We directly contrasted the two types of control conditions.

In meta-analysis 5 we addressed the weight of the type of response: button press [Experiments n. 2–9, 15, 18–27, 31, 32, 37] vs. no button press/silent decision [Experiments n. 1, 3, 10–14, 16, 17, 28–30, 33–36]. The condition tasks involving button/no-button presses do not necessarily evidence motor cortex activations as the contrast includes also many studies like for example study 5 from Table 1 in which action-related sounds are compared to non-living-related sounds, but both categories (action and non-living) require a button press, and motor cortex activation has been subtracted out.

Study 2: Neuropsychological Study

Participants

Patients

Inclusion/exclusion criteria. We included 7 neurosurgical patients meeting the following inclusion/exclusion criteria. Inclusion criteria were: a lesion involving areas included in the results of the meta-analysis, i.e., right and left temporo-insular-opercular cortex, being native Italian speakers, normal or corrected-to-normal vision and no history of psychiatric disease or drug abuse. Patients were excluded if they reported a hearing loss (as measured by the audiogram examination routinely performed before surgery), previous history of neurological problems or family history of developmental language problems or learning disabilities as well as inadequate speech comprehension, as they needed to understood the task and the instructions (for the tests included in their neuropsychological screening, see Table 2). Patients should not present with aphasia, as measured with standardized clinical tests (see Table 2). In particular, by excluding patients with aphasia and naming deficits we made sure that any impairment in ES recognition would not be related to language problems but reflect impaired ES processing. Lastly, among patients with lesions involving the right hemisphere we excluded those who had visuo-spatial/attentive deficits to make sure that any impairment in ES recognition would not be related to disorders of spatial attention (i.e., auditory) but reflect impaired ES processing.

TABLE 1
www.frontiersin.org

Table 2. Patients' neuropsychological profile.

Seven right-handed neurosurgical patients (4M, 3F) (mean age 55.57 ± 12.47 years, and mean years of schooling 13.28 ± 4.34 years) were admitted to the local General Hospital some days before the beginning of the study. We tested patients' ES recognition ability before surgery. Each patient received a neuropsychological battery the day before the fMRI. The neuropsychological evaluation included tests assessing non-verbal intelligence, verbal short-term memory, praxis, visuo-spatial ability and planning, constructional apraxia, and language. All the patients performed these tasks successfully (See Table 2). Conventional T2-weighted MR imaging revealed low-grade lesions (82.17 ± 50.77, range: 9.94–161 mean cc in volume). The lesion overlap of all the patients showed that the lesion involved part of the left and right superior and MTG, temporal pole, hippocampus and parahippocampal area, insula, rolandic operculum, IFG (pars opercularis and triangularis), precentral gyrus and basal ganglia (see Figure 2A). The overlay plot of all the patients' lesions indicated the voxels most frequently damaged (see the bar code). The most frequent area (in bright green-yellow) corresponds to the hippocampus/insula/STG (see Figure 2A).

The study was approved by the Ethics Committee of our Institute and performed in accordance with the 1964 Declaration of Helsinki and subsequent amendments. The subjects' consent was obtained.

Environmental Sound Auditory Confrontation Naming Task

Stimulus norming study

The primary goal of our norming study was to create a corpus of stimuli and responses, develop scoring criteria and determine a cut-off for the patients' Z scores. We used the original Marcell et al. (2000)'s corpus of stimuli (N = 120) of everyday, non-verbal digitized sounds belonging to many different categories such as sounds produced by animals, people, musical instruments, tools, signals, and fluids (Marcell et al., 2000) to conduct our own rating study for the Italian population, as there might be population-dependent differences in sound knowledge and frequency. *.waw files were downloaded from their archive (http://marcellm.people.cofc.edu/confrontation%20sound%20naming/confront.htm) as 16-bit *.WAV files with a sampling rate of 22,050 Hz.

The stimulus norming study included 20 monolingual native Italian speakers (8 F, 12 M; mean age 37.15 ± 11.24; age range 21–56; mean handedness 88.88, range 100–50; mean years of education 13.65 ± 3.77 years, range 8–18) with no history of neurological or auditory symptoms, different from the group involved in the fMRI study. We found a significant difference in age between patients and healthy controls [t(25)= 3.63, p < 0.001] but no education or gender effect [t(25) = −0.35, p > 0.05 and t(25) = −0.23, p > 0.05]. All participants gave their informed consent.

Following Marcell et al. (2000), the participants' primary task was to carefully listen to sounds and name the stimuli. Sounds were presented at a comfortable, preset loudness established through pilot testing. Each of the randomly ordered sounds was presented once, and participants were allowed 30 s to complete their identification. The tasks lasted about 45 min. Presentation® software (Version 9.9, Neurobehavioral Systems Inc., CA, USA) was used for auditory stimuli presentation. Answers were recorded by a PC and written down by the experimenter for later analysis. The experimenter then used these responses to establish the patients' scoring accuracy.

In order to determine the mean accuracy and evaluate the subjects' responses, we used the same criteria as used by Marcell et al. (2000). In particular, the following were scored as correct: synonyms, accurate descriptions of the sound, plurals, self-corrections. By contrast, lack of response or a “don't know” type of response and generalized superordinate descriptions of the item were scored as incorrect inaccurate descriptions of the sound. Furthermore, our rating study revealed that there were some sounds that were recognized by healthy controls as different from the responses reported in Marcell et al.'s study (Marcell et al., 2000). For example, our healthy participants recognized sounds like “explosion” as a “shot” (N = 5/20 subjects) or sounds like “frying food” as rain (N = 9 subjects), or sounds like typewriter (manual) as cash register (N = 7/20 subjects). For these items, if patients responded in a similar way as controls, we accepted their responses as correct. Last, there were some sounds that were not identified by healthy controls, such as cutting paper or water dripping (in both instances, 10/20 [50%] subjects did not recognize the sound) or a sonar (6/20 [33%] subjects did not recognize the sound). For these items, if patients responded in a similar way as controls, we accepted their responses as correct. Following these criteria, the participants correctly identified 87.83% ± 4.38 sounds (range 80–94.44). This result is very similar to the mean naming accuracy reported by Marcell et al. (2000) in their rating study (82.18 ± 22.67).

Thus, following the results obtained in our norming study, we selected 90 stimuli from the original corpus of stimuli of Marcell et al. (2000). The list of 90 stimuli used for the experiment had the following characteristics: 6.09 ± 0.90 mean familiarity, 3.19 ± 0.53 mean complexity, 3.92 ± 1.17 mean pleasantness, 2.47 ± 1.31 s mean duration, 82.18 ± 22.67 mean naming accuracy, and 6.06 ± 0.98 mean confidence in naming accuracy.

Task and procedure

Patients were asked to carefully listen to some sounds and name the stimuli (“Identify each sound as quickly and as accurately as you can”). We used the naming task similarly to Marcell et al.'s study. Sounds were presented at a comfortable, preset loudness established through pilot testing. Each of the randomly ordered sounds was presented once, and participants were allowed 30 s to complete their identification. Presentation® software (Version 9.9, Neurobehavioral Systems Inc., CA, USA) was used for stimuli presentation. Answers were recorded by a PC and written down by the experimenter for later analysis. Both the patients and healthy controls performed this task prior to the fMRI session. A typical testing session lasted 45 min.

Data analysis

Responses were analyzed by two independent raters. Accuracy was computed following the guidelines of Marcell et al. (2000) and according to the results of our own stimulus rating study (see below). For each patient we determined the Z score to calculate the number of patients whose performances were below the normal range (the reference group was healthy individuals). In addition, we performed a qualitative analysis of errors and labeled them as: semantically related to the target sound, auditorily related to the target sound (some sounds were both semantically and auditorily related to the target sound, in which case we coded them as semantically and auditorily related), unrelated and “I don't know,” and we expressed the total number of the different types of errors as % of the total errors. Last, we coded the errors according to the sound category by Marcell et al. (2000) (in their paper, a classification of sounds according to 27 categories can be found in Table 10).

Study 3: Functional Magnetic Imaging (fMRI) Study

The same ES auditory confrontation naming task with the same stimulus list described above was used during fMRI measurements involving the patients included in Study 2. Study 2 was a stimulus norming study including the original Marcell et al. (2000)'s corpus of stimuli (N = 120), whereas Study 3 included the final set of 90 stimuli. In the fMRI study, patients and healthy participants (see below) silently named the stimuli. We carefully instructed the subjects on how to perform the task. We asked them to listen carefully and name each stimulus, and at the end of the fMRI acquisition they would be asked some questions about each stimulus. Patients were highly motivated to perform the fMRI task as they knew that the fMRI maps are part of their clinical examination. More importantly, during fMRI acquisition we routinely performed online General Linear Model (GLM) analysis and continuously checked the activation and the BOLD signal correlation with the alternation of task and rest. If the GLM analysis showed that activation correlated significantly with the task, patients were performing the task appropriately. On the contrary, if no correlation emerged, we stopped the acquisition, talked to the patient and started the task again.

Healthy Controls for the fMRI Study

The patients' fMRI maps were compared with those of a control group consisting of 12 monolingual native Italian speakers (7 F, 5 M; mean age 35.75 ± 4.2 years old; age range 29–41; mean handedness 88.88 ± 17.88, range 100–50; mean education 16.5 ± 2.23 years, range 13–18). We found a significant difference age and education [t(17) = 19.82, p < 0.001 and t(17) = −3.2, p < 0.05] between patients and healthy controls, but not a gender effect [t(17) = −0.41, p > 0.05].

All participants had normal or corrected-to-normal vision and no history of neurological illness, psychiatric disease, or drug abuse. Following Marcell et al. (2000), we checked that none of the healthy controls responded affirmatively to the self-report question, “To the best of your knowledge, do you have a hearing loss?” All gave their informed consent to participate in the study.

Task and Procedure

The task started with an instruction (3 s). Subjects were asked to “carefully listen to the sounds and silently name the source of each sound as accurately and as quickly as possible.” During auditory stimulation a fixation cross was present on the screen. Blocks of ES recognition stimuli (N = 18, 15 s each) were alternated with baseline resting periods (N = 17, 15 s each, plus two additional resting blocks, one at the beginning of the run and the other at the end). In the baseline condition, a fixation cross (15 s) was presented between blocks and patients and controls were asked to relax. Each 15-s block included 5 stimuli, for a total of 90 stimuli. The same stimulus list used in the off-line pre-fMRI testing was presented during scanning. Presentation® software (Version 9.9, Neurobehavioral Systems Inc., CA, USA) was used for stimulus presentation and synchronization with the MR scanner. Participants listened to the stimuli via an Audio System (Resonance Technology).

fMRI Data Acquisition

A 3-T Philips Achieva whole-body scanner was used for both the patients and healthy controls to acquire T1-weighted anatomical images and functional images using a SENSE-Head-8 channel head coil and a custom-built head restrainer to minimize head movements. For both the patients and controls, functional images were obtained using a T2*-weighted echo-planar image (N = 222 EPI) sequence of the whole brain. The imaging parameters were as follows: repetition time, TR = 2500 ms; echo time, TE = 35 ms, field of view, FOV = 23 cm, acquisition matrix: 128 × 128, slice thickness: 3 mm with no gaps, 90° flip angle, voxel size: 1.8 × 1.8 × 3 mm; parallel imaging, SENSE = 2), and were preceded by 5 dummy images that allowed the MR scanner to reach a steady state.

For healthy controls, high-resolution anatomical images were acquired using a T1-weighted 3D magnetization-prepared, rapid acquisition gradient fast filed echo (T1W 3D TFE SENSE) pulse sequence (TR = 8.2 ms, TE = 3.76 ms, FOV = 24 cm, 190 transverse axial slices of 1 mm thickness, 8° flip angle, voxel size: 1 × 1 × 1 mm) lasting 8.8 min.

For patients, fMRI scanning was always performed before gadolinium injection and 6–10 days prior to craniotomy. In addition, high-resolution T2-weighted and post-gadolinium contrast T1-weighted anatomical MR images were acquired for use with the stereotactic surgical navigation system by using a T1-weighted 3D magnetization-prepared, rapid acquisition gradient-echo fast field echo (T1W_3D_TFE SENSE) pulse sequence (TR = 8.1007 ms, TE = 3.707 ms, FOV = 240.000 mm, 190 sagittal slices of 1 mm thickness, flip angle = 8°, voxel size: 1 × 1 × 1) and a T3-weighted 3D magnetization-prepared, rapid acquisition gradient-echo fast field echo (T2W_3D_TFE SENSE) pulse sequence (TR = 2500 ms, TE = 368.328 ms, FOV = 240.000 mm, 190 sagittal slices of 1 mm thickness, flip angle = 90°, voxel size: 1 × 1 × 1).

fMRI Data Processing and Whole Brain Analysis

fMRI data pre-processing and statistical analysis were performed on UNIX workstations (Ubuntu 8.04 LTS, i386, http://www.ubuntu.com/) using MATLAB r2007b (The Mathworks Inc., Natick, MA/USA) and SPM5 (Statistical Parametric Mapping software, SPM; Wellcome Department of Imaging Neuroscience, London, UK). Dummy images were discharged before further image processing. Pre-processing included spatial realignment of the images to the reference volume of the time series, segmentation producing the parameter file used for normalization of functional data to a standard EPI template of the Montreal Neurological Institute template provided by SPM5, re-sampling to a voxel size of 2 × 2 × 2 mm, and spatial smoothing with a 6-mm FWHM Gaussian kernel to meet the statistical requirements of the General Linear Model and to compensate for residual macro-anatomical variations across subjects.

We checked that the movement parameters for all the patients and healthy controls were < 3 mm for translation and < 3 for rotation. We used the lesion masking image, i.e., a ROI image drawn on the patient's lesion in which the voxels are coded as 0 (tumor area) and 1 (healthy brain tissue). In the normalization procedure, we included the lesion masking image following Brett et al's technique (2001). This procedure allows to exclude the masked region (i.e., the lesion that would otherwise produce artifacts altering the normalization outcome) from normalization. Then, the normalization outcome was inspected carefully. In particular, three observers (B.T., D.S., and M.M.) independently compared the original and the normalized images and excluded any distortion phenomenon.

To delineate the network related to the ES recognition task, we modeled the alternating epochs by a simple boxcar reference vector. A general linear model for blocked designs was applied to each voxel of the functional data by modeling the activation and the baseline conditions for each subject and their temporal derivatives by means of reference waveforms which correspond to boxcar functions convolved with a homodynamic response function (Friston et al., 1995a,b). Furthermore, we included 6 additional regressors that modeled the head movement parameters obtained from the realignment procedure. Accordingly, a design matrix, which comprised contrast modeling alternating intervals of “activation” and “baseline” (resting), was defined. At a single subject level, specific effects were assessed by applying appropriate linear contrasts to the parameter estimates of the baseline and experimental conditions resulting in t-statistics for each voxel. For the single-subject first-level analysis, low-frequency signal drifts were filtered using a cut-off period of 128 s. These t-statistics were then transformed into Z-statistics constituting statistical parametric maps (SPM{Z}) of differences across experimental conditions and between experimental conditions and the baseline. SPM{Z} statistics were interpreted in light of the theory of probabilistic behavior of Gaussian random fields (Friston et al., 1995a,b).

With regard to second-level random effects analyses for both patients and healthy controls, contrast images obtained from individual participants were entered into a one-sample t-test to generate a SPM{T} indicative of significant activations specific for this contrast at the group level. For both patients' and controls' group we included age and education as covariate. We used a threshold of P < 0.05, corrected for multiple comparisons at the cluster level [using family-wise error (FWE)], with a height threshold at the voxel level of P < 0.001, uncorrected.

The following contrast images were calculated: first, we estimated the main effects of CONDITION (ES listening–baseline for the controls > task ES listening–baseline for the patients), then we performed a conjunction null analysis (and not a global null analysis) (Friston et al., 1999), showing the commonly activated network for both tasks (ES listening–baseline for the patients > ES listening–baseline for the controls) using a threshold of p < 0.05, corrected for multiple comparisons at the cluster level (using FWE), with a height threshold at the voxel level of p < 0.001, uncorrected. The anatomical interpretation of the functional imaging results was performed using the SPM Anatomy toolbox (Eickhoff et al., 2005).

Results

Study 1: Meta-analysis Study of the Reviewed fMRI Studies about Environmental Sound Processing

The activation clusters resulting from meta-analysis 1 of all the reviewed studies comprised: (i) the right STG, extending to the MTG and Heschl's gyrus, the insula and the operculum [clusters 1, 2 and 4]; (ii) the left MTG extending to the STG and the supramarginal gyrus (SMG) and the insula [clusters 13 and 14]; (iii) the right [cluster 3] and the left [cluster 14] parahippocampal gyrus; (iv) the right IFG [clusters 5 and 6] and the left IFG including the pars orbitalis, triangularis and opercularis and the precentral area [clusters 16 and 17]; (v) the SMA [clusters 7 and 8]; (vi) the post-central, the inferior and superior parietal lobule [clusters 18 and 19]; (vi) the right putamen and thalamus [clusters 9 and 10]; and (vii) the right [clusters 11 and 12] and the left [clusters 20 and 21] cerebellum (see Table 3 and Figure 1)1.

TABLE 3
www.frontiersin.org

Table 3. Results of the ALE meta-analysis.

FIGURE 1
www.frontiersin.org

Figure 1. (A) Results of the ALE meta-analysis. The overlap between the general network and the networks found in meta-analyses 2–5 (in green). In red the areas not influenced by the external factors. (B) Influence of different design choices on the ES network.

In meta-analysis 2 we investigated first the effect of the type of stimulus. Action- vs. animal- related stimuli activated the left and right MTG, the left SMG, the left pars triangularis of the IFG and the right pars orbitalis. Animal- vs. action- related stimuli activated the left and right STG (see Table 3 and Figure 1). As indicated in Figure 1, data show that part of the network related to ES processing is influenced by the “type of stimulus” factor.

In meta-analysis 3 we addressed the effect of the type of control sound. Some studies compared ES stimuli to silent stimuli, rest or fixation conditions, and others compared ES stimuli to other control auditory sounds. Control sounds vs. silent events or rest activated the IFG (pars triangularis). Silent events or rest vs. control sounds activated the right STG and the left MTG, the left SMG and the right precentral gyrus, the right insula, the right putamen, the middle cingulate cortex, and the cerebellum bilaterally (see Table 3 and Figure 1).

In meta-analysis 4 we addressed the weight of the type of task: category judgment vs. passive listening. Part of the activation found in the left IFG (pars opercularis) and left rolandic operculum was related to making a category judgment compared to passive listening. This contrast revealed activation in the right STG and the left Heschl's gyrus (see Table 3 and Figure 1).

In meta-analysis 5 we addressed the weight of the type of response: button press vs. no button press/silent decision. Part of the activation found in the right and left STG, the right and left insula, the right and left IFG (pars triangularis and orbitalis), the right putamen and the right cerebellum is related to button press. The contrast revealed activation in the left inferior parietal lobe only (see Table 3 and Figure 1).

To sum up, the networks found in meta-analyses 2–5 do not tap ES-related activations because, according to the logic of cognitive subtraction, this is “subtracted out” and the resultant map reflects the effect of an external variable (i.e., type of stimulus, or type of response) on the network.

Figure 1A shows the overlap between the general network and the networks found in meta-analyses 2–5 (in green). The areas that are not influenced by the effect of any external variable are shown in red. These included: the hippocampal area bilaterally, the right rolandic operculum, part of the STG bilaterally, and the left post-central area and superior parietal lobule.

Study 2: Neuropsychological Study on Environmental Sound Recognition Performance

Table 4 and Figure 2 show the patients' performances on the ES confrontation naming task. Patients scored below the normal range (as measured by Z-scores). Most of the patients' responses were not related to target sounds (35.94 ± 17.21%, see Figure 2 and Table 5) or were semantically related to target sounds (29.72 ± 8.73%). The other types of responses were: auditorily related (12.19 ± 6.66%), semantically and auditorily related (11.01 ± 4.20%), and “I don't know” answers (11.85 ± 20.72%) (unrecognized by patients but correctly identified by controls). Last, we coded the errors according to the sound categories by Marcell et al. (2000) (see their Table 10 for a classification of sounds according to 27 categories). We found that the 15.14% of the patients' errors involved musical instruments, 14.74% involved animal sounds, 37.45% involved other categories (e.g., transportation, nature, signals, accidents, weapons), and 32.67% involved actions/human sounds. Note that stimuli belonging to “musical instruments” and “animal sounds” are less numerous than those belonging to the “other” and “actions/human sounds” categories). For this reason, any further investigation of living vs. non-living related differences was not addressed.

TABLE 4
www.frontiersin.org

Table 4. Environmental sound recognition performance and qualitative error analysis in patients and healthy controls.

FIGURE 2
www.frontiersin.org

Figure 2. (A) Overlap of the patients' lesions in standard space. MRIcron software (http://www.mccauslandcenter.sc.edu/mricro/mricron/index.html) was used to draw the patients' lesions on their T1 and T2 MRI scans, creating the ROIs which were normalized to the MNI space using the “Clinical Tool box” (http://www.mccauslandcenter.sc.edu/CRNL/clinical-toolbox) for SPM8 (http://www.fil.ion.ucl.ac.uk/spm/software/spm8/). The results highlighted the areas of the brain that are related to the deficit (Karnath et al., 2004). The number of overlapping lesions is illustrated by different colors that code for increasing frequencies (as indicated in the bar code). (B) Patients' pathological performance (mean accuracy) and healthy controls' accuracy and patients' qualitative analysis of errors. (C) The most frequently lesioned area (in bright green-yellow) is the hippocampus/insula/superior temporal gyrus, as shown by the Anatomy toolbox. By using Marsbar (http://marsbar.sourceforge.net/), we drew two ROIs on the two clusters found in the left and in the right STG (which were less activated in patients than controls), shown respectively in pink and in red. The density bar shows that at least 4 out of 7 patients' lesions overlapped with the ROIs drawn on the STG. (D) Network of areas commonly activated in patients and controls and areas differentially recruited by controls vs. patients during ES recognition in addition to the network for ES processing in patients and controls. Activations were superimposed on a brain template provided by spm5.

TABLE 5
www.frontiersin.org

Table 5. Brain regions showing a significant increase in BOLD response for environmental sound listening in (i) controls > patients, and (ii) patients > controls, and common to both (∩).

Study 3: fMRI Investigation

The areas showing a different activation in patients vs. controls (controls > patients) were: (i) the right STG, (ii) the left STG, (iii) the right IFG (pars opercularis), and (iv) the left insula extending to the IFG (pars triangularis) (see Figure 2 and Table 5). As to differences in activation across the STG, by using Marsbar (http://marsbar.sourceforge.net/), we drew two ROIs on the two clusters found in the left and the right STG (which were less activated in patients than controls), shown respectively in pink and in red in Figure 2B. We overlapped all the ROIs of the lesions checking the density bar showing how many patients had a lesion overlapping with the two ROIs. The map showed that the red ROI (right STG) overlapped with about 2 of the RH lesions, and that the pink ROI (left STG) overlapped with about 2 of the LH lesions. Taken together, these data suggest that the two ROIs on the two clusters found in the left and the right STG overlapped at least with 4 of the 7 lesions, indicating that the lack of activation in the STG is related to brain damage.

The reverse contrast (patients > controls) revealed a higher activation in patients vs. the control group, in the anterior cingulate, the right thalamus and the left inferior parietal lobule.

The functional areas whose activation was comparable to that of controls as revealed by the conjunction analysis (sound listening > baseline in patients > sound listening > baseline in controls) included: (i) the right Heschl's gyrus, extending to the STG, (ii) the left STG, extending to the MTG, (iii) the right IFG (pars opercularis), (iv) the left IFG (pars opercularis), (v) the SMA bilaterally, and (vi) the right insula (see Figure 2 and Table 5)2.

Overlap of the ALE Map with the fMRI Map of Patients and Controls

In Figure 3 we used the “Logical Overlays” function in Mango (http://ric.uthscsa.edu/mango/). We overlapped the ALE map (in blue) with the fMRI map of our patients (in green) and that of healthy controls (in red). Different combination of overlaps were included, e.g., fMRI control and fMRI patients; ALE map and fMRI controls. As shown in Figure 3, the three maps overlap in the STS. This is consistent with the less activation in the STS found in patients vs. controls (see the two ROIs shown in Figure 2C).

FIGURE 3
www.frontiersin.org

Figure 3. We used the “Logical Overlays” function in Mango (http://ric.uthscsa.edu/mango/). We overlapped the ALE map (in blue) with the fMRI map of our patients (in green) and that of healthy controls (in red). Different combination of overlaps were included, e.g., fMRI control and fMRI patients; ALE map and fMRI controls. In particular, in green the overlap of the three maps in the STS.

Discussion

In the present multimodal study we used a new approach combining different techniques and looking for converging evidence from multiple sources to explore the neuroanatomy of ES recognition.

The ALE meta-analysis delineated the core set of regions involved in ES as evidenced by fMRI literature. This analysis revealed the network of areas supporting ES recognition. We next showed that not all of the clusters truly reflect ES processing and how design choices, e.g., type of stimulus, type of task, type of control condition, might have influenced the activation observed in the fMRI studies we evaluated.

The hippocampal area bilaterally, the right rolandic operculum, part of the STG bilaterally, and the left post-central area and superior parietal lobule were not influenced by any of the factors which might influence the ES network. Part of this region (i.e., hippocampus, rolandic operculum, and STG/MTG bilaterally) was found to be most frequently lesioned in our patient sample.

The pathological performance of neurosurgical patients showed that areas of the ES network have a causal role in ES processing, since a lesion in those areas caused a deficit in ES recognition. Patients had a normal performance on the neuropsychological screening. In particular, by excluding patients with aphasia and naming deficits we made sure that any impairment in ES recognition would not be related to language problems but reflect impaired ES processing.

These results indicate that a deficit of ES recognition does not arise exclusively following lesions to the right hemisphere or left hemisphere. So far, no precise anatomical locations have been correlated with auditory agnosia (Lewis et al., 2004). Our data, thus, add new information to the ES recognition related literature, showing that both the left and the right hemisphere, if damaged, can cause a deficit in ES recognition.

One crucial area involved in ES is the STG/MTG. Although the meta-analysis showed that part of the left and right STG/MTG activation during ES processing might be in part related to design choices, this area was one of the most frequently lesioned areas in our patient sample, thus highlighting its causal role in ES processing. The planum temporale is the auditory association cortex and it represents the first (input and processing) node (or computational hub, Griffiths and Warren, 2002) of the network involved in segregating the components of the acoustic stimulus and matching these components with learned spectrotemporal representations. The information is then gated to higher-order cortical areas for further processing (Griffiths and Warren, 2002). The STG activation has been reported as reflecting the input stages of ES processing (Lewis et al., 2004). In our fMRI study, right STG activation was found in patients and controls. This suggests that these areas were still actively functional in patients, too. However, the controls > patients comparison revealed that a greater activation of a sub-part of the STG in controls vs. patients, meaning that patients lacked activation in a crucial sector of the STG. The two ROIs on the two clusters found in the left and the right STG overlapped at least with 4 out of 7 patients lesions, indicating that the lack of STG activation found for patients is related to brain damage. Several authors found bilateral activations in the STG during ES processing, with a larger region of activation in the right STG (Bergerbest et al., 2004). The right STG posterior to the primary auditory cortex has been proposed as a node of the “neural semantic detector model” describing semantic memory for non-verbal sounds (Kraut et al., 2006). Of course, as evidenced by the ES recognition model (Lewis et al., 2004), too, the information resulting from the described processing steps needs the intervention of the semantic system, which is lateralized to the LH. Accordingly, the meta-analysis showed that studies requiring active judgment or categorization as compared to those requiring passive listening additionally activated the left IFG and the left rolandic operculum.

Our ALE meta-analysis included the right and the left parahippocampal gyrus as it does the temporal part of the lesion map. Its role in sound recognition might be related to the localization of sounds or the mental spatial imagery of source sound localization. Some authors suggested that the activation of this area possibly reflects the “imageability” of sounds (Sharda and Singh, 2012). See also (Engel et al., 2009).

ES processing requires allocating auditory attention/memory to the input sounds. Accordingly, the right IFG has been related to auditory working memory (Zatorre et al., 1994; Zatorre, 2001), or to allocating auditory attention (Lipschutz et al., 2002; Binder et al., 2004). Interestingly, it has been shown that auditory verbal hallucinations predominantly activate the right IFG (Sommer et al., 2008). In our fMRI study, the direct controls > patients comparison revealed that a part of the right IFG was more activated in controls than in patients, meaning that impaired patients lacked activation in a crucial sector of the IFG. It has been suggested that ES is polymodal in nature and the IFG bilaterally is responsible for integrating polymodal object representations with concepts in semantic memory. Interestingly, we found that activation in different sectors of the IFG was related to many design choices. Only the right pars triangularis truly reflected activation related to ES processing. This region, together with the insula, is an area that was frequently lesioned in our patients. It is part of a finely tuned attentional network which selects information from the continuous flow of auditory signals and triggers communication and balance between the RH and LH according to the nature of the stimulus (Habib et al., 1995). The insula has numerous direct connections with the auditory cortex (Adams and Janata, 2002; Bamiou et al., 2003). A comparison of a patient with total agnosia following bilateral insular damage (Habib et al., 1995) with a case with no agnosia following left insula–thalamocortical projection damage (Hyman and Tranel, 1989) is indicative of the essential role of the bilateral insula in auditory stimuli pre-processing.

Since ES recognition is both a top-down and bottom-up driven process, it presupposes an interaction between many areas in the brain. And it also presupposes an involvement of the fiber tracts. In our patients, the lesions might also be interpreted in terms of damage to the fiber tracts. Indeed, many of the clusters discussed here are interconnected with the input nodes of the temporal cortex. Accordingly, in the temporal-parietal lobe area, the activation found in the ALE map, including the left precentral/post-central gyrus/supramarginal gyrus, was related to the type of stimulus. We found that action- vs. animal-related stimuli activated the left supramarginal gyrus. It is known that perceptual processing and semantic processing interact to represent ES. Thus, these activations might be related to the action/human sound category processing. With regard to action-related verbs and phrases, it has been shown that imagery of the verbal context could be responsible for activation in sensorimotor areas (e.g., Tomasino and Rumiati, 2013a,b).

Conclusion

ES recognition is dependent on a bilateral network of areas in the temporal, inferior frontal basal ganglia, and areas of the pre- and post-central gyrus, as shown by the ALE meta-analysis. We showed that some of these clusters of activation truly reflect ES processing, whereas others are related to design choices.

The hippocampal area bilaterally, the right rolandic operculum, part of the STG bilaterally, and the left post-central area and superior parietal lobule were not influenced by any of the factors which might influence the ES network. In addition, the lesion map evidenced areas that are necessary for ES processing, namely the hippocampus, STG/MTS area and the rolandic operculum, which might be deregulated in activation as compared to healthy controls.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We wish to thank the patients and the healthy volunteers and gratefully acknowledge the MRI staff of the Unità Operativa di Neuroradiologia (A.O.S. Maria della Misericordia, Udine, Italy).

Footnotes

1. ^All the studies we included reported activations surviving corrections for multiple comparisons except some (study 2, 16, 19, 20, 22, 27, and 37). We re-run the analysis by excluding these studies. We confirm that the same network of activation was found.

2. ^Age and education did not influence the ES related network, as by including these two variables as covariate in our analysis, the same network was found for both patients' and controls' group.

References

Adams, R. B., and Janata, P. (2002). A comparison of neural circuits underlying auditory and visual object categorization. Neuroimage 16, 361–377. doi: 10.1006/nimg.2002.1088

PubMed Abstract | CrossRef Full Text | Google Scholar

Altmann, C. F., Doehrmann, O., and Kaiser, J. (2007). Selectivity for animal vocalizations in the human auditory cortex. Cereb. Cortex 17, 2601–2608. doi: 10.1093/cercor/bhl167

PubMed Abstract | CrossRef Full Text | Google Scholar

Bamiou, D. E., Musiek, F. E., and Luxon, L. M. (2003). The insula (Island of Reil) and its role in auditory processing. Literature review. Brain Res. Brain Res. Rev. 42, 143–154. doi: 10.1016/S0165-0173(03)00172-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Basso, A., Capitani, E., and Laiacona, M. (1987). Raven's coloured progressive matrices: normative values on 305 adult normal controls. Funct. Neurol. 2, 189–194.

PubMed Abstract | Google Scholar

Bergerbest, D., Ghahremani, D. G., and Gabrieli, J. D. (2004). Neural correlates of auditory repetition priming: reduced fMRI activation in the auditory cortex. J. Cogn. Neurosci. 16, 966–977. doi: 10.1162/0898929041502760

PubMed Abstract | CrossRef Full Text | Google Scholar

Binder, J. R., Liebenthal, E., Possing, E. T., Medler, D. A., and Ward, B. D. (2004). Neural correlates of sensory and decision processes in auditory object identification. Nat. Neurosci. 7, 295–301. doi: 10.1038/nn1198

PubMed Abstract | CrossRef Full Text | Google Scholar

Caspers, S., Zilles, K., Laird, A. R., and Eickhoff, S. (2010). ALE meta-analysis of action observation and imitation in the human brain. Neuroimage 50, 1148–1167. doi: 10.1016/j.neuroimage.2009.12.112

PubMed Abstract | CrossRef Full Text | Google Scholar

Clarke, S., Bellmann, A., De, R. F., and Assal, G. (1996). Non-verbal auditory recognition in normal subjects and brain-damaged patients: evidence for parallel processing. Neuropsychologia 34, 587–603. doi: 10.1016/0028-3932(95)00142-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Clarke, S., Bellmann, A., Meuli, R. A., Assal, G., and Steck, A. J. (2000). Auditory agnosia and auditory spatial deficits following left hemispheric lesions: evidence for distinct processing pathways. Neuropsychologia 38, 797–807. doi: 10.1016/S0028-3932(99)00141-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Clarke, S., Bellmann, T. A., Maeder, P., Adriani, M., Vernet, O., Regli, L., et al. (2002). What and where in human audition: selective deficits following focal hemispheric lesions. Exp. Brain Res. 147, 8–15. doi: 10.1007/s00221-002-1203-9

PubMed Abstract | CrossRef Full Text | Google Scholar

De Renzi, E., and Faglioni, P. (1978). Normative data and screening power of a shortened version of the Token Test. Cortex 14, 41–49. doi: 10.1016/S0010-9452(78)80006-9

PubMed Abstract | CrossRef Full Text | Google Scholar

De Renzi, E., Motti, F., and Nichelli, P. (1980). Imitating gestures. A quantitative approach to ideomotor apraxia. Archiv. Neurol. 37, 6–10. doi: 10.1001/archneur.1980.00500500036003

PubMed Abstract | CrossRef Full Text | Google Scholar

Dick, F., Saygin, A. P., Galati, G., Pitzalis, S., Bentrovato, S., D'Amico, S., et al. (2007). What is involved and what is necessary for complex linguistic and nonlinguistic auditory processing: evidence from functional magnetic resonance imaging and lesion data. J. Cogn. Neurosci. 19, 799–816. doi: 10.1162/jocn.2007.19.5.799

PubMed Abstract | CrossRef Full Text | Google Scholar

Eickhoff, S. B., Bzdok, D., Laird, A. R., Kurth, F., and Fox, P. T. (2012). Activation Likelihood Estimation meta-analysis revisited. Neuroimage 59, 2349–2361. doi: 10.1016/j.neuroimage.2011.09.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Eickhoff, S. B., Laird, A. R., Grefkes, C., Wang, L. E., Zilles, K., and Fox, P. T. (2009). Coordinate-based Activation Likelihood Estimation meta-analysis of neuroimaging data: a random-effects approach based on empirical estimates of spatial uncertainty. Hum. Brain Mapp. 30, 2907–2926. doi: 10.1002/hbm.20718

PubMed Abstract | CrossRef Full Text | Google Scholar

Eickhoff, S., Stephan, K. E., Mohlberg, H., Grefkes, C., Fink, G. R., Amunts, K., et al. (2005). A new SPM toolbox for combining probabilistic cytoarchitectonic maps and functional imaging data. Neuroimage 25, 1325–1335. doi: 10.1016/j.neuroimage.2004.12.034

PubMed Abstract | CrossRef Full Text | Google Scholar

Engel, L. R., Frum, C., Puce, A., Walker, N. A., and Lewis, J. W. (2009). Different categories of living and non-living sound-sources activate distinct cortical networks. Neuroimage 47, 1778–1791. doi: 10.1016/j.neuroimage.2009.05.041

PubMed Abstract | CrossRef Full Text | Google Scholar

Engelien, A., Silbersweig, D., Stern, E., Huber, W., Döring, W., Frith, C., et al. (1995). The functional anatomy of recovery from auditory agnosia. A PET study of sound categorization in a neurological patient and normal controls. Brain 118(Pt 6), 1395–1409. doi: 10.1093/brain/118.6.1395

PubMed Abstract | CrossRef Full Text | Google Scholar

Fecteau, S., Armony, J. L., Joanette, Y., and Belin, P. (2004). Is voice processing species-specific in human auditory cortex? An fMRI study. Neuroimage 23, 840–848. doi: 10.1016/j.neuroimage.2004.09.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Friston, K. J., Frith, C. D., Turner, R., and Frackowiak, R. S. J. (1995a). Characterising evoked hemodynamics with fMRI. Neuroimage 2, 157–165.

PubMed Abstract | Google Scholar

Friston, K. J., Holmes, A. P., Price, C. J., Büchel, C., and Worsley, K. (1999). Multi-subject fMRI studies of conjunction analyses. Neuroimage 10, 385–396. doi: 10.1006/nimg.1999.0484

PubMed Abstract | CrossRef Full Text | Google Scholar

Friston, K. J., Holmes, A. P., Worsley, K. J., Poline, J.-B., Frith, C. D., and Frackowiak, R. S. J. (1995b). Statistical parametric maps in functional imaging: a general linear approach. Hum. Brain Mapp. 2, 189–210.

Google Scholar

Fujii, T., Fukatsu, R., Watabe, S., Ohnuma, A., Teramura, K., Kimura, I., et al. (1990). Auditory sound agnosia without aphasia following a right temporal lobe lesion. Cortex 26, 263–268. doi: 10.1016/S0010-9452(13)80355-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Giovagnoli, A. R., Del Pesce, M., Mascheroni, S., Simoncelli, M., Laiacona, M., and Capitani, E. (1996). Trail making test: normative values from 287 normal adult controls. Ital. J. Neurol. Sci. 17, 305–309. doi: 10.1007/BF01997792

PubMed Abstract | CrossRef Full Text | Google Scholar

Griffiths, T. D., and Warren, J. D. (2002). The planum temporale as a computational hub. Trends Neurosci. 25, 348–353. doi: 10.1016/S0166-2236(02)02191-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Habib, M., Daquin, G., Milandre, L., Royere, M. L., Rey, M., Lanteri, A., et al. (1995). Mutism and auditory agnosia due to bilateral insular damage–role of the insula in human communication. Neuropsychologia 33, 327–339. doi: 10.1016/0028-3932(94)00108-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Hackett, T. A., Stepniewska, I., and Kaas, J. H. (1999). Callosal connections of the parabelt auditory cortex in macaque monkeys. Eur. J. Neurosci. 11, 856–866. doi: 10.1046/j.1460-9568.1999.00492.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Hyman, B. T., and Tranel, D. (1989). Hemianesthesia and aphasia. An anatomical and behavioral study. Arch. Neurol. 46, 816–819. doi: 10.1001/archneur.1989.00520430112026

PubMed Abstract | CrossRef Full Text | Google Scholar

Karnath, H.-O., Berger, M. F., Zopf, R., and Küker, W. (2004). Using SPM normalization for lesion analysis in spatial neglect. Brain 127, e10. doi: 10.1093/brain/awh111

PubMed Abstract | CrossRef Full Text | Google Scholar

Kazui, S., Naritomi, H., Sawada, T., Inoue, N., and Okuda, J. (1990). Subcortical auditory agnosia. Brain Lang. 38, 476–487. doi: 10.1016/0093-934X(90)90132-Z

PubMed Abstract | CrossRef Full Text | Google Scholar

Kraut, M. A., Pitcock, J. A., Calhoun, V., Li, J., Freeman, T., and Hart, J. Jr. (2006). Neuroanatomic organization of sound memory in humans. J. Cogn. Neurosci. 18, 1877–1888. doi: 10.1162/jocn.2006.18.11.1877

PubMed Abstract | CrossRef Full Text | Google Scholar

Laird, A. R., Eickhoff, S. B., Kurth, F., Fox, P. M., Uecker, A. M., Turner, J. A., et al. (2009). ALE meta-analysis workflows via the brainmap database: progress towards a probabilistic functional brain atlas. Front. Neuroinform. 3:23. doi: 10.3389/neuro.11.023.2009

PubMed Abstract | CrossRef Full Text | Google Scholar

Laird, A. R., Fox, P. M., Price, C. J., Glahn, D. C., Uecker, A. M., Lancaster, J. L., et al. (2005). ALE meta-analysis: controlling the false discovery rate and performing statistical contrasts. Hum. Brain Mapp. 25, 155–164. doi: 10.1002/hbm.20136

PubMed Abstract | CrossRef Full Text | Google Scholar

Lancaster, J. L., Tordesillas-Gutiérrez, D., Martinez, M., Salinas, F., Evans, A., Zilles, K., et al. (2007). Bias between MNI and Talairach coordinates analyzed using the ICBM-152 brain template. Hum. Brain Mapp. 28, 1194–1205. doi: 10.1002/hbm.20345

PubMed Abstract | CrossRef Full Text | Google Scholar

Leaver, A. M., and Rauschecker, J. P. (2010). Cortical representation of natural complex sounds: effects of acoustic features and auditory object category. J. Neurosci. 30, 7604–7612. doi: 10.1523/JNEUROSCI.0296-10.2010

PubMed Abstract | CrossRef Full Text | Google Scholar

Lewis, J. W. (2006). Cortical networks related to human use of tools. Neuroscientist. 12, 211–231. doi: 10.1177/1073858406288327

PubMed Abstract | CrossRef Full Text | Google Scholar

Lewis, J. W., Phinney, R. E., Brefczynski-Lewis, J. A., and DeYoe, E. A. (2006). Lefties get it “right” when hearing tool sounds. J. Cogn. Neurosci. 18, 1314–1330. doi: 10.1162/jocn.2006.18.8.1314

PubMed Abstract | CrossRef Full Text | Google Scholar

Lewis, J. W., Talkington, W. J., Walker, N. A., Spirou, G. A., Jajosky, A., Frum, C., et al. (2009). Human cortical organization for processing vocalizations indicates representation of harmonic structure as a signal attribute. J. Neurosci. 29, 2283–2296. doi: 10.1523/JNEUROSCI.4145-08.2009

PubMed Abstract | CrossRef Full Text | Google Scholar

Lewis, J. W., Wightman, F. L., Brefczynski, J. A., Phinney, R. E., Binder, J. R., and DeYoe, E. A. (2004). Human brain regions involved in recognizing environmental sounds. Cereb. Cortex 14, 1008–1021. doi: 10.1093/cercor/bhh061

PubMed Abstract | CrossRef Full Text | Google Scholar

Lipschutz, B., Kolinsky, R., Damhaut, P., Wikler, D., and Goldman, S. (2002). Attention-dependent changes of activation and connectivity in dichotic listening. Neuroimage 17, 643–656. doi: 10.1006/nimg.2002.1184

PubMed Abstract | CrossRef Full Text | Google Scholar

Maeder, P. P., Meuli, R. A., Adriani, M., Bellmann, A., Fornari, E., Thiran, J. P., et al. (2001). Distinct pathways involved in sound recognition and localization: a human fMRI study. Neuroimage 14, 802–816. doi: 10.1006/nimg.2001.0888

PubMed Abstract | CrossRef Full Text | Google Scholar

Marcell, M. M., Borrella, D., Greene, M., Kerr, E., and Rogers, S. (2000). Confrontation naming of environmental sounds. J. Clin. Exp. Neuropsychol. 22, 830–864. doi: 10.1076/jcen.22.6.830.949

PubMed Abstract | CrossRef Full Text | Google Scholar

Mendez, M. F., and Geehan, G. R. Jr. (1988). Cortical auditory disorders: clinical and psychoacoustic features. J. Neurol. Neurosurg. Psychiatr. 51, 1–9. doi: 10.1136/jnnp.51.1.1

PubMed Abstract | CrossRef Full Text | Google Scholar

Miceli, G., Laudanna, A., Burani, C., and Capasso, R. (1994). Batteria per L'analisi dei Deficit Afasici. B. A. D. A. [B. A. D. A.: A Battery for the Assessment of Aphasic Disorders.]. Roma: CEPSAG.

Mondini, S., Mapelli, D., Vestri, A., Arcara, G., and Bisiacchi, P. S. (2011). Esame Neuropsicologico Breve 2. Milano: Raffaello Cortina Editore.

Nové-Josserand, A., Fischer, C., Nighoghossian, N., and Trouillas, P. (1998). [Auditory perception disorders due to bilateral cortical lesions. An electrophysiology study]. Rev. Neurol. 154, 318–321.

PubMed Abstract | Google Scholar

Novelli, G., Papagno, C., Capitani, E., Laiacona, M., Vallar, G., and Cappa, S. F. (1996). Tre test clinici di ricerca e produzione lessicale. Taratura su soggetti normali. Arch. Psicol. Neurol. Psichiatry 47, 477–505.

Oldfield, R. C. (1971). The assessment and analysis of handedness: the edinburgh inventory. Neuropsychologia 9, 97–113. doi: 10.1016/0028-3932(71)90067-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Orsini, A., Grossi, D., Capitani, E., Laiacona, M., Papagno, C., and Vallar, G. (1987). Verbal and spatial immediate memory span: normative data from 1355 adults and 1112 children. Ital. J. Neurol. Sci. 8, 539–548. doi: 10.1007/bf02333660

PubMed Abstract | CrossRef Full Text | Google Scholar

Rauschecker, J. P., and Scott, S. K. (2009). Maps and streams in the auditory cortex: nonhuman primates illuminate human speech processing. Nat. Neurosci. 12, 718–724. doi: 10.1038/nn.2331

PubMed Abstract | CrossRef Full Text | Google Scholar

Romanski, L. M., Bates, J. F., and Goldman-Rakic, P. S. (1999b). Auditory belt and parabelt projections to the prefrontal cortex in the rhesus monkey. J. Comp. Neurol. 403, 141–157.

PubMed Abstract | Google Scholar

Romanski, L. M., and Goldman-Rakic, P. S. (2002). An auditory domain in primate prefrontal cortex. Nat. Neurosci. 5, 15–16. doi: 10.1038/nn781

PubMed Abstract | CrossRef Full Text | Google Scholar

Romanski, L. M., Tian, B., Fritz, J., Mishkin, M., Goldman-Rakic, P. S., and Rauschecker, J. P. (1999a). Dual streams of auditory afferents target multiple domains in the primate prefrontal cortex. Nat. Neurosci. 2, 1131–1136. doi: 10.1038/16056

PubMed Abstract | CrossRef Full Text | Google Scholar

Rosati, G., De Bastiani, P., Paolino, E., Prosser, S., Arslan, E., and Artioli, M. (1982). Clinical and audiological findings in a case of auditory agnosia. J. Neurol. 227, 21–27. doi: 10.1007/BF00313543

PubMed Abstract | CrossRef Full Text | Google Scholar

Saygin, A. P., Dick, F., Wilson, S. M., Dronkers, N. F., and Bates, E. (2003). Neural resources for processing language and environmental sounds: evidence from aphasia. Brain 126, 928–945. doi: 10.1093/brain/awg082

PubMed Abstract | CrossRef Full Text | Google Scholar

Schnider, A., Benson, D. F., Alexander, D. N., and Schnider-Klaus, A. (1994). Non-verbal environmental sound recognition after unilateral hemispheric stroke. Brain 117(Pt 2), 281–287. doi: 10.1093/brain/117.2.281

PubMed Abstract | CrossRef Full Text

Sharda, M., and Singh, N. C. (2012). Auditory perception of natural sound categories—an fMRI study. Neuroscience 214, 49–58. doi: 10.1016/j.neuroscience.2012.03.053

PubMed Abstract | CrossRef Full Text | Google Scholar

Sommer, I. E., Diederen, K. M., Blom, J. D., Willems, A., Kushan, L., Slotema, K., et al. (2008). Auditory verbal hallucinations predominantly activate the right inferior frontal area. Brain 131, 3169–3177. doi: 10.1093/brain/awn251

PubMed Abstract | CrossRef Full Text | Google Scholar

Specht, K., and Reul, J. (2003). Functional segregation of the temporal lobes into highly differentiated subsystems for auditory perception:an auditory rapid event-related fMRI-task. Neuroimage 20, 1944–1954. doi: 10.1016/j.neuroimage.2003.07.034

PubMed Abstract | CrossRef Full Text | Google Scholar

Spinnler, M., and Tognoni, G. (1987). Standardizzazione e taratura italiana di test neuropsicologici. Ital. J. Neurol. Sci. 6(Suppl. 8), 1–120.

Staeren, N., Renvall, H., De, M. F., Goebel, R., and Formisano, E. (2009). Sound categories are represented as distributed patterns in the human auditory cortex. Curr. Biol. 19, 498–502. doi: 10.1016/j.cub.2009.01.066

PubMed Abstract | CrossRef Full Text | Google Scholar

Taniwaki, T., Tagawa, K., Sato, F., and Iino, K. (2000). Auditory agnosia restricted to environmental sounds following cortical deafness and generalized auditory agnosia. Clin. Neurol. Neurosurg. 102, 156–162. doi: 10.1016/S0303-8467(00)00090-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Thierry, G., Giraud, A. L., and Price, C. (2003). Hemispheric dissociation in access to the human semantic system. Neuron 38, 499–506. doi: 10.1016/S0896-6273(03)00199-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Tomasino, B., Chiesa, A., and Fabbro, F. (2014). Disentangling the neural mechanisms involved in Hinduism- and Buddhism-related meditations. Brain Cogn. 90, 32–40. doi: 10.1016/j.bandc.2014.03.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Tomasino, B., Fregona, S., Skrap, M., and Fabbro, F. (2012). Meditation-related activations are modulated by the practices needed to obtain it and by the expertise: an ALE meta-analysis study. Front. Hum. Neurosci. 6:346. doi: 10.3389/fnhum.2012.00346

PubMed Abstract | CrossRef Full Text | Google Scholar

Tomasino, B., and Rumiati, R. I. (2013a). At the mercy of strategies: the role of motor representations in language understanding. Front. Psychol. 4:27. doi: 10.3389/fpsyg.2013.00027

PubMed Abstract | CrossRef Full Text | Google Scholar

Tomasino, B., and Rumiati, R. I. (2013b). Introducing the special topic “The when and why of sensorimotor processes in conceptual knowledge and abstract concepts”. Front. Hum. Neurosci. 7:498. doi: 10.3389/fnhum.2013.00498

CrossRef Full Text | Google Scholar

Turkeltaub, P. E., Eden, G. F., Jones, K. M., and Zeffiro, T. A. (2002). Meta-analysis of the functional neuroanatomy of single-word reading: method and validation. Neuroimage 16, 765–780. doi: 10.1006/nimg.2002.1131

PubMed Abstract | CrossRef Full Text | Google Scholar

Vignolo, L. A. (1982). Auditory agnosia. Philos. Trans. R. Soc. Lond. B Biol. Sci. 298, 49–57. doi: 10.1098/rstb.1982.0071

PubMed Abstract | CrossRef Full Text | Google Scholar

Wilson, B. A., Cockburn, J., and Halligan, P. W. (1987). Behavioral Inattention Test (BIT). Bury St. Edmunds: Thames Valley Test Company.

Zatorre, R. J. (2001). Neural specializations for tonal processing. Ann. N.Y. Acad. Sci. 930, 193–210. doi: 10.1111/j.1749-6632.2001.tb05734.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Zatorre, R. J., Evans, A. C., and Meyer, E. (1994). Neural mechanisms underlying melodic perception and memory for pitch. J. Neurosci. 14, 1908–1919.

PubMed Abstract | Google Scholar

Zatorre, R. J., Evans, A. C., Meyer, E., and Gjedde, A. (1992). Lateralization of phonetic and pitch discrimination in speech processing. Science 256, 846–849. doi: 10.1126/science.1589767

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: environmental sounds, fMRI, neurosurgical patients, Activation Likelihood Estimation (ALE) meta-analysis, lesion mapping

Citation: Tomasino B, Canderan C, Marin D, Maieron M, Gremese M, D'Agostini S, Fabbro F and Skrap M (2015) Identifying environmental sounds: a multimodal mapping study. Front. Hum. Neurosci. 9:567. doi: 10.3389/fnhum.2015.00567

Received: 01 June 2015; Accepted: 28 September 2015;
Published: 21 October 2015.

Edited by:

Yong He, Beijing Normal University, China

Reviewed by:

Christiane Thiel, Carl von Ossietzky Universität Oldenburg, Germany
Mingrui Xia, Beijing Normal University, China

Copyright © 2015 Tomasino, Canderan, Marin, Maieron, Gremese, D'Agostini, Fabbro and Skrap. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Barbara Tomasino, YnRvbWFzaW5vQHVkLmxuZi5pdA==

co-first author.

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.