Skip to main content

REVIEW article

Front. Neurosci., 18 October 2022
Sec. Perception Science
This article is part of the Research Topic Response to an Object Near the Head/Body: Multisensory Coding and Motor Processing Guided by Sensory Systems View all 9 articles

Integration of somatosensory and motor-related information in the auditory system

  • Department of Physiology, Anatomy & Genetics, University of Oxford, Oxford, United Kingdom

An ability to integrate information provided by different sensory modalities is a fundamental feature of neurons in many brain areas. Because visual and auditory inputs often originate from the same external object, which may be located some distance away from the observer, the synthesis of these cues can improve localization accuracy and speed up behavioral responses. By contrast, multisensory interactions occurring close to the body typically involve a combination of tactile stimuli with other sensory modalities. Moreover, most activities involving active touch generate sound, indicating that stimuli in these modalities are frequently experienced together. In this review, we examine the basis for determining sound-source distance and the contribution of auditory inputs to the neural encoding of space around the body. We then consider the perceptual consequences of combining auditory and tactile inputs in humans and discuss recent evidence from animal studies demonstrating how cortical and subcortical areas work together to mediate communication between these senses. This research has shown that somatosensory inputs interface with and modulate sound processing at multiple levels of the auditory pathway, from the cochlear nucleus in the brainstem to the cortex. Circuits involving inputs from the primary somatosensory cortex to the auditory midbrain have been identified that mediate suppressive effects of whisker stimulation on auditory thalamocortical processing, providing a possible basis for prioritizing the processing of tactile cues from nearby objects. Close links also exist between audition and movement, and auditory responses are typically suppressed by locomotion and other actions. These movement-related signals are thought to cancel out self-generated sounds, but they may also affect auditory responses via the associated somatosensory stimulation or as a result of changes in brain state. Together, these studies highlight the importance of considering both multisensory context and movement-related activity in order to understand how the auditory cortex operates during natural behaviors, paving the way for future work to investigate auditory-somatosensory interactions in more ecological situations.

Introduction

Vision and hearing provide the primary source of information about distant objects and events that may be too far away to engage the other senses. Furthermore, visual and auditory cues arising from the same source are often complementary and integrated by the brain in ways that shape our perception of the information provided by the source, including its location, timing and identity (see Opoku-Baah et al., 2021 for a recent review). The visual and auditory systems are equally important for processing inputs from nearby sources, such as when we are reading text on or listening to our mobile phones. Within this proximal region of space, other sensory modalities come more into play, which therefore need to be considered too if we are to understand the neural basis for perception and behaviors that are guided by sensory cues close to the head and body.

In this review, we examine how the processing of auditory inputs is influenced by other sensory modalities near the head, focusing primarily on interactions with the somatosensory system. We start by considering how the sound field changes with distance from the source and the implications of this for perceiving nearby auditory objects. We examine the evidence for sensitivity to auditory target distance, including the contribution of auditory inputs to the encoding of space close to the body, before looking at the way auditory and somatosensory signals interact to influence perception and neuronal responses, particularly within the auditory regions of the brain. Most studies have examined these crossmodal interactions by presenting sounds whilst applying vibrotactile stimuli to the skin, but proprioceptive signals arising from changes in body orientation also need to be considered to understand the full extent to which somatosensory signals are integrated with auditory processing. Finally, because a range of actions—including exploratory movements, reaching and grasping, as well as avoidance responses—can be elicited by nearby sensory stimuli, we examine the impact of motor-related activity on auditory processing, highlighting some of the similarities between the effects of movement, brain state and somatosensory cues on the auditory system.

Sound propagation and distance cues

Sound waves behave differently as they propagate away from the source (Figure 1; Moore and King, 1999). Within the near field, the region of the sound field closest to the source, a complex relationship exists between sound pressure level and distance as some of the energy circulates without propagating. At a distance corresponding to roughly one wavelength, the far field begins, which can be subdivided into a free-field region where the sound pressure level follows the inverse square law, decreasing by 6 dB with each doubling of the distance from the source, and a reverberant field, where reflections from the walls and other surfaces within the room result in delayed and distorted versions of the direct sound. These properties give rise to several cues that help the brain to encode sound sources as they vary with distance from the listener. They include changes in level, frequency composition (with high frequencies scattered more at longer distances), the ratio of direct-to-reverberant sound energy (which declines with increasing distance), and, for sounds that are sufficiently close to fall within the acoustic near field, variations in interaural level differences on each side of the listener (Brungart and Rabinowitz, 1999; Kolarik et al., 2016).

FIGURE 1
www.frontiersin.org

Figure 1. At locations near a sound source (left), or near the walls of a reverberant room (right), sound levels are variable and do not follow the normal, free-field rule of decrease with the inverse square of distance. The blue areas on the graph represent the variable influences of the near field and reverberation. Adapted from Moore and King (1999).

Sensitivity to sound-source distance

Most studies of auditory distance perception in humans have focused on the far-field region of space (sometimes using virtual acoustic space stimuli to simulate distances of around a meter or more from the listener’s ears), where this relies principally on estimates of the level of the sound and the difference in level between the direct and reverberant sound (Bronkhorst and Houtgast, 1999; Zahorik, 2002; Kolarik et al., 2013). The relative contribution of these two cues depends on the type of sound used and its direction (Zahorik, 2002), and changes with distance and room reverberation time (Kolarik et al., 2013). Of course, sound level provides a useful absolute distance cue only if the source has a fixed level and a number of other factors, including other sensory cues and body orientation, have been shown to affect auditory distance perception (Harris et al., 2015; Kolarik et al., 2016). In particular, auditory distance judgments can be influenced by the more accurate and reliable spatial information provided by visual stimuli (Loomis et al., 1998; Zahorik, 2001; Anderson and Zahorik, 2014; Kolarik et al., 2016).

Behavioral studies in humans (Brungart, 1999; Zahorik, 2002; Kopčo and Shinn-Cunningham, 2011) and rabbits (Kuwada et al., 2015) have also demonstrated distance sensitivity for sound sources in close proximity to the head. In the absence of sound-level cues, a comparison of the levels of direct and reverberant sound energy appears to play the dominant role (Kopčo and Shinn-Cunningham, 2011). However, if there are no room reflections, as in anechoic conditions, it is likely that interaural level differences in low-frequency sounds (< 3 kHz) are used to judge the distance of nearby lateral sources (Brungart, 1999). Neurophysiological evidence for level-independent sensitivity to the distance of nearby sound sources has been obtained using fMRI in human non-primary posterior auditory cortex, with these findings again suggesting that this is based primarily on a comparison of the levels of direct and reverberant sounds (Kopěo et al., 2012; Kopco et al., 2020). Since amplitude modulation depth changes with the ratio of direct-to-reverberant sound energy—and therefore the source distance—it has been suggested that the sensitivity of neurons in the inferior colliculus (IC) to the depth of amplitude modulation might provide a basis for using this cue to represent sound-source distance (Kim et al., 2015).

Auditory contributions to the representation of peripersonal space

Sensitivity to sounds close to the body, and particularly for looming sounds that are likely to be indicative of an approaching object (Seifritz et al., 2002; Maier et al., 2008), contributes to the representation in the brain of peripersonal space, the region of space within our immediate reach, where exteroceptive information from the eyes and ears interacts with somatosensory inputs. Multisensory processing in peripersonal space provides information about the position of the body in the environment and helps to localize nearby objects (Serino, 2019).

The processing of multisensory signals in peripersonal space has been studied extensively within a frontoparietal network of the primate brain (reviewed by Serino, 2019). Neurons in these areas respond both to tactile stimuli on specific parts of the body and to visual and/or auditory stimuli in close proximity to the body. For example, acoustically-responsive neurons recorded in the ventral premotor cortex of awake monkeys were found to respond more strongly when broadband sounds were placed closer to the head, often independently of stimulus level (Graziano et al., 1999). The majority of these neurons also responded to visual and tactile stimulation, and are thought to represent sensory information that is used to guide reaching or avoidance responses (Graziano and Gross, 1998). Furthermore, behavioral (Occelli et al., 2011; Teneggi et al., 2013) and electrophysiological (Bernasconi et al., 2018) studies in humans have reported that sounds are more likely to influence tactile processing when they are presented in close proximity to the head than further away.

While this work highlights the importance of detecting and responding to nearby sounds, particularly those looming toward the body, the methods used for investigating sensitivity to sound-source distance in many studies of peripersonal space are limited in terms of the range of sound-source locations tested and by the lack of attention to the way sound properties change with distance within the near field. Indeed, whether auditory-tactile peripersonal space exists at all has been questioned (Holmes et al., 2020), and the frontoparietal cortical areas in macaques and humans that have been the focus of peripersonal space research are dominated by visual and somatosensory, rather than auditory, inputs (Macaluso and Maravita, 2010).

External objects generating auditory (or visual) signals that move closer to the head and body can signify approaching danger, which therefore requires a rapid response (Graziano and Cooke, 2006). However, integration of tactile information with other sensory modalities is not limited to those situations. Palpation and manipulation of objects in peripersonal space, using the hands, feet, mouth or—especially in rodents (Sofroniew and Svoboda, 2015)—via exploratory movements of mechanosensitive whiskers, will often generate sound. This may occur as a direct result of the haptic interactions with the object, but sounds can also be produced by locomotory or other accompanying movements required to execute these actions. There are therefore many situations where auditory and somatosensory signals are present at the same time and it is important to understand how they are combined and integrated in the brain.

Perceptual consequences of interactions between auditory and somatosensory inputs

Although most perceptual studies of multisensory integration have focused on vision and audition (Opoku-Baah et al., 2021), interactions can occur across all sensory modalities and there is growing evidence for close links between auditory and tactile processing. Over the past 20 years, numerous studies have explored the effects of auditory-tactile interactions on human perception. In fact, similarities between audition and touch were recognized much earlier (Von Békésy, 1959), including the fact that both senses are based on the detection of frequency-dependent mechanical displacements. Sensitivity to the frequency of mechanical vibrations in each modality, which overlap in a low-frequency range that extends up to about 1,000 Hz, is an important factor in determining how they interact (Soto-Faraco and Deco, 2009; Figure 2).

FIGURE 2
www.frontiersin.org

Figure 2. Comparison of human sensitivity to vibrotactile and acoustic stimuli as a function of the frequency of stimulation. The vibrotactile sensitivity curve is based on Gescheider et al. (2002) and the hearing threshold curve is based on International Standard ISO 389-7: 2003(E) (Reference threshold of hearing under free-field and diffuse-field listening conditions).

Simultaneous audio-tactile stimulation can improve reaction times (Sperdin et al., 2009; Godenzini et al., 2021) and stimulus detection (Schürmann et al., 2004; Gillmeister and Eimer, 2007; Ro et al., 2009), and increase perceived loudness (Gillmeister and Eimer, 2007; Yarrow et al., 2008). These interactions have been shown to occur in a frequency-dependent manner (Ro et al., 2009; Tajadura-Jiménez et al., 2009; Wilson et al., 2010a,b), with the largest perceptual improvements reported when the frequencies of the auditory and vibrotactile stimuli are closely matched (Ro et al., 2009; Wilson et al., 2010b). More generally, auditory-tactile interactions are particularly prominent in the temporal domain (Occelli et al., 2011) and therefore play a fundamental role in the way individuals interact with their environments, as well as in the production and perception of music and speech (Ito et al., 2009; Occelli et al., 2011; Keough et al., 2019).

There is a growing body of evidence from studies in human participants that the auditory and somatosensory systems can reciprocally bias each other (Soto-Faraco and Deco, 2009; Occelli et al., 2011; Villanueva and Zampini, 2018). Perceptual interactions between these modalities are, however, to some extent task dependent and asymmetric. This is often explored by asking participants to attend to one modality while ignoring distractor stimuli in the other modality. For example, performance on a tactile frequency-discrimination task is impaired by task-irrelevant auditory distractors, so long as the auditory stimulus is of a similar frequency to the attended tactile stimulus, whereas judgments of the intensity of the tactile stimulus are unaffected (Yau et al., 2009). By contrast, the presence of a tactile distractor can bias both the perceived frequency and intensity of an auditory stimulus (Yau et al., 2010). Crossmodal adaptation paradigms, in which the different stimuli are separated in time, have shown that an auditory adaptor can also influence the perception of vibrotactile frequency but not intensity, with this interaction again limited to overlapping frequencies in each modality (Crommett et al., 2017). The impact of stimulus frequency in integrating audition and touch is highlighted by the parchment-skin illusion. Increasing over headphones the high-frequency components of sound accompanied by the motion of rubbing the hands together leads to subjects reporting a drier or rougher sensation of their skin (Jousmäki and Hari, 1998).

Psychophysical evidence for frequency-specific audio-tactile interactions has been observed for a wide range of repetition rates, extending down into the flutter range (< 50 Hz) (Convento et al., 2019), where the stimuli are perceived as a series of individual pulses rather than a vibratory hum, and in the perception of frequency sweeps (Crommett et al., 2019). This is likely to be important for sensing objects that generate correlated tactile and auditory signals, as in the aforementioned example of texture perception (Lederman, 1979). These findings indicate that auditory and tactile frequency representations are integrated in the brain. In humans, cortical regions exhibiting sensitivity to both tactile and auditory frequency information appear to be sparsely distributed within modality-specific areas in the parietal and temporal lobes (Rahman et al., 2020), while interactions between these modalities in mouse somatosensory cortical areas have been found to depend on the frequency of tactile stimulation (Zhang et al., 2020).

In the spatial domain, where vision is normally the dominant sense, behavioral and electrophysiological measurements in humans have shown that similar crossmodal interactions take place between auditory and somatosensory stimuli irrespective of their locations (Murray et al., 2005). Nevertheless, several studies have demonstrated that auditory spatial responses can be biased by tactile or proprioceptive inputs (Pick et al., 1969; Caclin et al., 2002; Sanabria et al., 2005; Bruns and Röder, 2010). Just as exposure to a consistent spatial mismatch between auditory and visual stimuli shifts auditory localization judgments that persist for a few minutes when sounds are subsequently presented alone (Radeau and Bertelson, 1974; Recanzone, 1998), an audio-tactile ventriloquism aftereffect exists, suggesting that auditory space is continually recalibrated to compensate for spatial disparities between these modalities too (Bruns et al., 2011). Somatosensory capture of auditory motion perception has also been demonstrated (Soto-Faraco et al., 2004; Figure 3). Tactile distractors moving in a conflicting direction disrupt auditory motion perception (Figure 3C), whereas auditory motion distractors have a smaller effect on the apparent direction of motion of tactile stimuli (Figure 3D).

FIGURE 3
www.frontiersin.org

Figure 3. Crossmodal capture of the perception of apparent motion direction with auditory and tactile stimuli in humans. (A) Schematic of the experimental setup. Tactile stimuli were delivered by hand-held vibrating foam cubes, and auditory stimuli were presented from two loudspeakers. Participants were asked to evaluate whether auditory or tactile target stimuli were moving to the right or left while ignoring stimulation in the other distractor modality. (B) The direction of apparent motion for the two modalities was either congruent or conflicting, with the stimuli presented either synchronously or asynchronously, with a 500 ms delay between the target and distractor. (C) Effects of a tactile motion distractor on the accuracy of auditory motion direction judgments. Note that task-irrelevant tactile motion in the opposing direction impaired the ability of participants to detect the correct sound movement direction, but only when the stimuli were presented synchronously. (D) Effects of an auditory motion distractor on the accuracy of tactile motion direction judgments. A smaller crossmodal modulatory effect was observed when the tactile stimulus was the target and the auditory stimulus the distractor. Adapted with permission from Soto-Faraco et al. (2004).

Different sensory cues appear to be combined and integrated in a statistically optimal way according to their relative reliability (Ernst and Bülthoff, 2004). While this can explain why certain sensory modalities appear to dominate and capture specific aspects of perception in other modalities, it is also important to take differences between species and their habitats into account. Thus, in mice, which are nocturnal animals, the weighting between vision and audition is reversed, with audition dominating vision when perceptual decisions are made in the presence of conflicting visual and auditory information (Song et al., 2017), suggesting that “sensory hierarchies” can be species as well as task specific. By the same token, given their reliance on whisker movements for exploratory behavior and tactile discrimination (Sofroniew and Svoboda, 2015), there is evidence that rodents may place even more weight on the somatosensory system when combining inputs across different sensory modalities (Rao et al., 2014). This highlights the importance of considering the natural ecology of the species in question when considering the functional consequences and neural underpinnings of multisensory integration in the brain.

Interactions between auditory and somatosensory inputs in the brain

Given the extensive evidence for the effects of combining auditory and tactile cues on human perception, it is not surprising that the majority of studies exploring the neural substrates for these multisensory interactions in both humans and other species have focused on the cerebral cortex. In addition to the higher-level cortical association areas implicated in the representation of peripersonal space, multisensory signals have been found to converge throughout the cortex, including in auditory, somatosensory and visual early cortical areas (reviewed in Ghazanfar and Schroeder, 2006; Meijer et al., 2019). Thus, responses to auditory stimulation in somatosensory cortical areas have been reported in both human fMRI (Liang et al., 2013; Pérez-Bellido et al., 2018; Rahman et al., 2020) and magnetoencephalography studies (Lütkenhöner et al., 2002), as well as in single-neuron measurements in monkeys (Zhou and Fuster, 2004) and mice (Carvell and Simons, 1986; Zhang et al., 2020; Godenzini et al., 2021). More attention has been paid, however, to the way that other sensory modalities affect activity in the auditory cortex, with somatosensory influences demonstrated using a range of methods and in several different species (Foxe et al., 2002; Fu et al., 2003; Brosch et al., 2005; Kayser et al., 2005; Schürmann et al., 2006; Lakatos et al., 2007; Lemus et al., 2010; Iurilli et al., 2012; Nordmark et al., 2012; Hoefer et al., 2013; Rao et al., 2014; Lohse et al., 2021).

While most studies of multisensory processing have shown that sound-evoked responses in the auditory cortex are modulated by somatosensory or visual stimuli, there are several reports in different animal species that some neurons in the auditory cortex can be driven by other sensory modalities (Fu et al., 2003; Brosch et al., 2005; Bizley et al., 2007; Lemus et al., 2010; Meredith and Allman, 2015; Morrill and Hasenstaub, 2018). The functional significance of responses evoked by the “non-principal” modality in auditory or other sensory cortical areas is unknown. Lemus et al. (2010) reported that neurons exist in macaque monkey primary auditory cortex (A1) that respond to somatosensory stimulation, but were unable to decode the identity of the tactile stimulus from the activity of these neurons. There is some indication that responses in auditory cortex to visual and somatosensory stimuli can arise as a result of the behavioral procedure used to train monkeys over long periods in auditory tasks (Brosch et al., 2005). However, this does not explain the presence of these neurons in many other studies, where they have been most commonly reported under anesthesia or in awake animals that were not engaged in a sensory task.

On the basis of the changes induced in sound-evoked response properties, the modulatory influence of vision on the auditory cortex has been implicated in various perceptual phenomena, including visual enhancement of speech perception (Schroeder et al., 2008), sound localization (Bizley and King, 2008) and auditory scene analysis (Atilgan et al., 2018), as well as the ventriloquist illusion (Bonath et al., 2014; Zierul et al., 2017). Tactile (and visual) inputs have been shown to reset the phase of ongoing neuronal oscillations in the auditory cortex, effectively amplifying the response to auditory events that are aligned with the peaks in the oscillations (Lakatos et al., 2007; Kayser et al., 2008; Kayser, 2009; Figure 4). This in turn offers a potential explanation for the effects of somatosensory stimulation on sound intensity perception that were described in the previous section.

FIGURE 4
www.frontiersin.org

Figure 4. Modulatory influence of somatosensory stimuli on the auditory cortex in non-human primates. A somatosensory stimulus (hand, bottom row) resets the phase of the local field potential (LFP) oscillations in auditory cortex (top traces). As a result, auditory stimuli (ears, bottom row) are more likely to elicit spikes (blue dots, middle row) when they arrive during peaks than during troughs in the ongoing LFP. Based on Lakatos et al. (2007) and Kayser (2009).

A close interaction between what is heard and felt is an essential aspect of playing a musical instrument, and studies of musical training have provided unique insights into the plasticity of multisensory integration at a perceptual level and demonstrated that physiological and anatomical changes take place in the relevant cortical areas (Münte et al., 2002; Herholz and Zatorre, 2012). For example, compared to non-musicians, trumpet players display significantly increased cortical signal strength for combined auditory-somatosensory stimuli exclusively for tactile stimulation of their lips (Schulz et al., 2003). Musicians have also been found to react faster to tactile and non-musical auditory stimuli presented either separately or together (Landry and Champoux, 2017), are less affected by the audio-tactile flash illusion—where presentation of one tactile stimulus with multiple task-irrelevant tones normally leads to the perception of more than one touch (Landry et al., 2017), and are more sensitive to audio-tactile incongruencies than non-musicians (Kuchenbuch et al., 2014). In fact, musical training also appears to narrow the temporal integration window for binding auditory and visual signals (Petrini et al., 2009), which is seen for music, but not speech (Lee and Noppeney, 2011). Changes in multisensory processing in the auditory cortex are not only limited to professional musicians (Pantev et al., 2015), since auditory-somatosensory training in non-musicians can result in greater cortical plasticity than auditory-only training (Pantev et al., 2009).

Crossmodal suppression of cortical activity

The modulatory influence of somatosensory inputs on auditory cortical responses ranges from mostly facilitatory effects in monkeys (Fu et al., 2003; Kayser et al., 2005; Lakatos et al., 2007) to predominantly suppressive interactions in ferrets (Meredith and Allman, 2015), cats (Khorevin and Shelest, 1998) and rodents (Iurilli et al., 2012; Rao et al., 2014; Lohse et al., 2021), at least in the primary or core auditory areas. This may indicate that somatosensory stimuli that are more salient behaviorally can downregulate responses to potentially distracting sounds that are experienced at the same time. The finding by Rao et al. (2014) that facial touch during social interactions with conspecifics inhibits activity in A1 of rats supports this possibility, and provides a rare example of multisensory processing in a natural ecological context. Similar conclusions about the interplay between these modalities have also been drawn in studies demonstrating suppression of auditory responses in somatosensory cortical areas (Gobbelé et al., 2003; Zhang et al., 2020).

A recent study by Lohse et al. (2021) showed that these crossmodal interactions in mouse A1 are divisive in nature with passive whisker stimulation suppressing auditory responses primarily around the preferred sound frequency of the neurons, as shown by the changes in the frequency response profiles recorded at different levels of the auditory pathway (Figure 5). Divisive normalization is regarded as one of the hallmarks of multisensory integration, and has also been implicated in audio-tactile interactions in the mouse somatosensory cortex (Zhang et al., 2020) and in the way neuronal responses in other brain areas are determined by the efficacy and spatial relationship of the individual stimuli (Ohshiro et al., 2011). This is likely to represent a “canonical operation” performed by sensory neurons, which could be implemented by various biophysical mechanisms, many of which involve some form of inhibition (Carandini and Heeger, 2011). In this regard, it is interesting that crossmodal divisive scaling by non-driving somatosensory inputs is also found in mouse auditory thalamocortical neurons (Lohse et al., 2021), which, at least in rats, lack local recurrent connectivity (Bartlett and Smith, 1999).

FIGURE 5
www.frontiersin.org

Figure 5. Somatosensory influence on spectral tuning in the auditory midbrain, thalamus, and cortex of mice. (A) Schematic of the auditory pathway from midbrain to cortex. (B) Frequency response profiles from representative neurons (electrophysiologically recorded units) with and without concurrent whisker stimulation. Note the lack of effect of whisker stimulation in the CNIC, whereas an additive facilitatory effect was observed in the IC shell and in MGBm, and divisive suppression was found in MGBv, MGBd and in A1. Recordings in the IC were obtained in both anesthetized and awake mice; no differences were observed in the effects of whisker stimulation on auditory responses according to whether the mice were anesthetized or not. A1, primary auditory cortex; IC, inferior colliculus; CNIC, central nucleus of the IC; MGB, medial geniculate body; MGBd, MGBm, MGBv, dorsal, medial and ventral divisions of the MGB. Adapted from Lohse et al. (2021).

From a functional perspective, divisive scaling of A1 responses has also been linked to behavioral improvements in frequency discrimination at the expense of sound detection performance (Guo et al., 2017). Related to this is the finding that suppression of auditory cortical responses by visual inputs is associated with an increase in response reliability and in the amount of stimulus-related information transmitted (Bizley et al., 2007; Kayser et al., 2010). Although more research is needed in behaving animals, it is possible that while tactile inputs reduce auditory cortical activity, potentially prioritizing stimuli that touch the whiskers or other parts of the body, they may actually serve to enhance auditory acuity.

Neural circuits underlying auditory-somatosensory interactions

An important question that has implications for both the nature of the neural circuitry involved and the role of multisensory interactions in goal-directed behavior is where they take place in the brain. Convergence of inputs from different modalities in early cortical areas could arise from direct connections between those areas, feedback from higher-level associations areas, or be inherited from subcortical inputs (Driver and Noesselt, 2008; Cappe et al., 2012). In the case of audio-tactile interactions, connections between somatosensory cortex and auditory cortex have been described in several species (Cappe and Barone, 2005; Budinger et al., 2006; Hackett et al., 2007a; Ro et al., 2013; Meredith and Allman, 2015; Godenzini et al., 2021). This is illustrated in Figure 6 for studies carried out in humans (Figure 6A) and in gerbils (Figure 6B). It remains to be determined whether connections from somatosensory cortex to auditory cortical areas, including A1, are causally involved in the multisensory interactions observed physiologically. However, Godenzini et al. (2021) recently reported that photoinhibition of the auditory cortical projection to the forepaw region of the primary somatosensory cortex (S1) eliminated the sound-induced reduction in reaction time on a tactile detection task. This finding, along with other studies of cortical circuitry (reviewed by Meijer et al., 2019), indicates that corticocortical connections are likely to play an important role in mediating multisensory influences on perception and behavior.

FIGURE 6
www.frontiersin.org

Figure 6. Direct connectivity between the primary somatosensory cortex (S1) and primary auditory cortex (A1) in humans and rodents. (A) Fiber tracts connecting S1 and A1 in human cortex demonstrated using diffusion tensor imaging with deterministic tractography. (B) Direct projections from S1 to A1 in Mongolian gerbils were demonstrated by injecting fluorescein-labeled dextran into A1. Arrowheads indicate retrogradely-labeled cells in S1. HL/Tr, hindlimb/trunk regions; D, dorsal; C, caudal. Panel (A) is adapted with permission from Ro et al. (2013). Panel (B) is adapted with permission from Budinger et al. (2006).

Convergence of inputs from different sensory modalities is not, however, restricted to the cerebral cortex. Ascending inputs to the auditory cortex from multisensory regions of the thalamus have been described (Budinger et al., 2006; Hackett et al., 2007b; Cappe et al., 2009; Lohse et al., 2021), indicating that the multisensory properties of cortical neurons may, at least in part, simply reflect their thalamic inputs. In the following section, we briefly review the evidence for somatosensory inputs at different subcortical levels of the central auditory pathway, and outline the circuitry by which tactile stimulation can influence auditory processing.

Somatosensory influences on subcortical auditory processing

Compared to other sensory systems, the auditory pathway includes a large number of subcortical nuclei, most of which also receive non-auditory inputs (Wu et al., 2015). The first relay in the auditory pathway is the cochlear nucleus. Neurons in the dorsal cochlear nucleus (DCN) are thought to represent spectral localization cues (Yu and Young, 2000), which are primarily responsible for vertical localization and for distinguishing between sound directions in front of and behind the head (Kumpik and King, 2019). This role in sound localization is aided by proprioceptive inputs to the DCN, which in cats provide information about the orientation of the mobile external ears (Kanold and Young, 2001) and in rats may help to distinguish moving sound sources from the apparent movement produced by motion of the head (Wigderson et al., 2016). Somatosensory inputs to the DCN have also been implicated in suppressing the effects of self-generated noises on the central auditory system, such as those produced by vocalizing, licking and masticating (Shore and Zhou, 2006; Singla et al., 2017).

Recently, Ansorge et al. (2021) reported that whisker stimulation can modify sound-evoked activity at this level of the auditory system, enhancing the activity of DCN neurons and producing more diverse and cell-type specific effects in the ventral cochlear nucleus (VCN). Although the functional consequences of these multisensory facilitatory interactions are unclear, electrical stimulation of the spinal trigeminal nucleus has been found to enhance temporal coding by bushy cells, the principal output neurons of the VCN, which has implications for the subsequent processing of vocal and spatial information (Heeringa et al., 2018).

Changes in neuronal activity resulting from the integration of auditory and somatosensory signals should also be observed in the downstream targets of the cochlear nucleus. However, Lohse et al. (2021) found no effects of whisker stimulation in mice on the auditory responses of neurons in the central nucleus of the IC (CNIC) (Figure 5), which receives both direct and indirect inputs from the cochlear nucleus. It is possible that some of the somatosensory influences reported in the cochlear nucleus might have subthreshold effects on the activity of CNIC neurons, but this remains to be explored. The CNIC is the core or lemniscal part of the auditory midbrain and is surrounded by a shell comprising the dorsal cortex, a lateral (or external) cortex and a rostral cortex, which can be distinguished by their connections and response properties (Liu et al., 2022). Inputs from multiple subcortical and cortical levels of the somatosensory system have been identified in the IC shell, particularly its lateral cortex (Aitkin et al., 1978; Jain and Shore, 2006; Lesicko et al., 2016; Olthof et al., 2019; Lohse et al., 2021; Figures 7A,B), supporting the notion that the IC should be regarded as a hub for processing auditory signals in the context of other sensory, motor and cognitive information (Gruters and Groh, 2012).

FIGURE 7
www.frontiersin.org

Figure 7. Somatosensory inputs to inhibitory cells in the lateral shell of IC and their influence on auditory responses in the MGBv. (A) Axons from the dorsal column nuclei (DCoN) labeled with biotinylated dextran amine (BDA) terminate in inhibitory sectors labeled with GAD67 in the shell of the IC. Labeled axons and terminals are shown in teal, while the outlines of the GAD67 modules are shown in purple. (B) BDA-labeled axons from the trunk area of the primary somatosensory cortex (S1) also project primarily to these inhibitory sectors in the shell of the IC. (C) Left, anterograde transsynaptic labeling of neurons (red) in the IC shell from S1 (whisker area). These neurons are double labeled with VGAT + (YFP green) and project to the MGBv (right, red axons and terminals). (D) Optogenetically activating S1-recipient IC neurons expressing channelrhodopsin inhibits responses to tones in MGBv. LP, lateral posterior thalamus. Panels (A,B) adapted from Lesicko et al. (2016). Panels (C,D) adapted from Lohse et al. (2021).

The IC provides most of the auditory input to the superior colliculus, a major site for the integration of multisensory spatial information (Meredith and Stein, 1986; King, 2004). Because spatial information in each sensory modality is encoded using different reference frames, eye position signals have to be incorporated in order to align the different maps of space in the superior colliculus (Jay and Sparks, 1984; Hartline et al., 1995; Populin et al., 2004). The activity of some neurons in both the IC (Groh et al., 2001; Zwiers et al., 2004) and A1 (Werner-Reiss et al., 2003) of monkeys is also affected by eye position. While these signals could arise from proprioceptive feedback from the extraocular muscles, it is more likely that they are conveyed by corollary discharge from brain regions involved in controlling eye movements (Gruters and Groh, 2012).

The IC also projects to the medial geniculate body (MGB) in the thalamus, which provides the gateway to the auditory cortex. Neurons in the medial division of the MGB (MGBm), the suprageniculate nucleus (SGN) and the posterior intralaminar nucleus (PIN) respond to auditory and somatosensory stimuli (Bordi and LeDoux, 1994; Lohse et al., 2021), and, through their projections to the amygdala, are thought to be involved in fear conditioning (Cruikshank et al., 1992; Bordi and LeDoux, 1994). Most of the ascending input from the auditory thalamus to A1 comes from the ventral division of the MGB (MGBv). Although traditionally considered to be a purely auditory structure, somatosensory inputs have been shown to inhibit the activity of neurons in the MGBv (Khorevin, 1980; Kimura and Imbe, 2018; Lohse et al., 2021). Indeed, in mice, concurrent whisker stimulation has been found to suppress the sound-evoked responses of neurons recorded in the MGBv and the adjacent dorsal division of the MGB (MGBd) (Lohse et al., 2021). As with A1, whisker stimulation resulted in divisive scaling of the auditory responses of these neurons and their axon terminals, indicating that these multisensory integrative properties are most likely fed forward to the cortex (Figure 5).

Whisker stimulation has widespread effects on the MGB, suppressing auditory thalamocortical responses and enhancing the auditory responses of neurons in the more medial higher-order thalamic nuclei that principally project to other brain areas. Corresponding effects of audio-tactile stimulation have, however, not been found in the somatosensory thalamus (Lohse et al., 2021). Interactions between other sensory modalities in the rodent thalamus are also asymmetric, since visual stimuli can facilitate the processing of whisker deflection in the ventral posteromedial nucleus, whereas tactile modulation of visual responses has not been observed in the dorsal lateral geniculate nucleus (Allen et al., 2017; Bieler et al., 2018). Together, these findings suggest that multisensory convergence at the level of the thalamus may serve to make objects that touch the whiskers more salient, while reducing the impact of concurrent sounds. Whether this is also the case in primates and other species in which vision dominates multisensory interactions in peripersonal space remains to be seen.

Circuits underlying somatosensory suppression of auditory cortical activity

Using a combination of anterograde transsynaptic viral tagging and optogenetic manipulations, Lohse et al. (2021) demonstrated that whisker-stimulation-induced suppression in the auditory thalamocortical pathway is mediated by a corticocolliculo-thalamocortical loop (Figures 7C,D, 8). More specifically, layer 5 neurons in S1 project to a subset of inhibitory neurons in the lateral shell of the IC (Figures 7B,C), which, in turn, project to and suppress the sound-evoked responses of neurons in MGBv and MGBd (Figures 7C,D). The auditory cortex then inherits these suppressed responses from the auditory thalamus. These findings add to a growing body of evidence that communication between different cortical areas can be mediated by transthalamic circuits, as well as by cortico-cortical connections (Lohse et al., 2019; Mo and Sherman, 2019; Figure 8). They also show that the auditory midbrain is a part of the circuitry involved in integrating multisensory signals across the cerebral cortex.

FIGURE 8
www.frontiersin.org

Figure 8. Circuits for somatosensory-auditory interactions in the auditory system. SOC, superior olivary complex; AudTRN, auditory sector of the TRN; SomTRN, somatosensory sector of the TRN; PO, posterior medial nucleus; VP, ventroposterior nucleus. Other abbreviations in main text.

Another structure that could be involved in the modulatory effects of somatosensory stimuli on auditory processing at the level of the thalamus is the thalamic reticular nucleus (TRN), which contains networks of GABAergic neurons that are organized into sensory and motor subdivisions with inhibitory projections to specific thalamic nuclei (Crabtree, 2018; Figure 8). Although previously thought to be modality specific, multisensory convergence has now been demonstrated within the sensory sectors of the TRN, which includes the presence of predominantly suppressive interactions between auditory and somatosensory inputs (Kimura, 2017). However, currently available evidence suggests that the TRN is not part of the circuit implementing the suppressive effects of whisker stimulation on the auditory thalamus (Lohse et al., 2021), though the possibility that this intrathalamic pathway contributes under particular behavioral demands cannot yet be ruled out.

Together, these studies have demonstrated that several neural substrates, involving both ascending and descending projections, exist for combining somatosensory and auditory (as well as visual) inputs in the brain. The somatosensory system can therefore interact with multiple levels of abstraction of the auditory world, from influencing simple frequency representations in the cochlear nucleus to learned relevant auditory categories in the cortex, allowing enormous flexibility in how these circuits operate under different conditions.

Movement and the auditory system

There is more to understanding the causes, neural underpinnings and functional consequences of multisensory interactions than simply looking at the way particular combinations of sensory stimuli interact to alter the activity of neurons. As we have already stated, sounds are often generated by movement of the body, most obviously in the case of vocalizing, but also during a range of other activities. Indeed, there is growing evidence for a functional role for activation of motor areas in the brain in many aspects of auditory perception, not only for sounds associated with movement—speech, music and other action-related sounds—but even when listening passively (Aglioti and Pazzaglia, 2010; Lima et al., 2016; Froese and González-Grandón, 2020). Furthermore, it has been proposed that unsupervised sensorimotor learning based on the dynamic acoustic inputs resulting from an animal’s own movements can help to establish a stable representation of auditory space in the brain without the need for visual feedback (Aytekin et al., 2008).

Despite the close relationship between the auditory and motor systems, it is important that the sounds generated by an animal’s own actions are filtered out so that they do not interfere with the processing of auditory stimuli coming from other sources. Vocalizing suppresses auditory cortical activity both in humans (Paus et al., 1996) and animals (Eliades and Wang, 2003). Furthermore, some neurons in auditory cortex respond to perturbed vocal feedback signals (Eliades and Wang, 2008), suggesting the existence of circuits that predict the acoustic consequences of particular movements in order to suppress self-generated auditory inputs. While much of this work has focused on the auditory cortex, as we saw in the section on subcortical pathways, these effects first arise at the level of the cochlear nucleus. Thus, non-auditory inputs allow DCN neurons to cancel out responses to sounds generated by licking movements while retaining their sensitivity to external sounds (Singla et al., 2017; Figure 9). This depends on mossy fiber inputs from the spinal trigeminal nucleus, which presumably convey somatosensory information related to licking behavior.

FIGURE 9
www.frontiersin.org

Figure 9. Neurons in the mouse dorsal cochlear nucleus, but not the ventral cochlear nucleus, cancel out the effects of self-generated sounds that result from licking a waterspout. (A) Dextran-conjugated Alexa 594 labeling (green) at recording sites in DCN and VCN (arrowheads). DAPI, red. Right, higher magnification of dashed white box on left showing a labeled fusiform cell (arrowhead) in the DCN. (B) Example VCN unit exhibiting responses time locked to the licks. Arrows and dashed lines indicate times of tongue contact with the lick spout. Traces represent the microphone recording (top), smoothed firing rate (middle) and VCN unit recording (bottom; scale: 30 μV). (C) Example DCN unit recorded during licking. Note the lack of lick-related responses. (D) Example VCN unit responses to an externally generated sound with temporal and spectral properties that roughly matched the licking sounds (“lick mimic”). Same unit as in panel (A). Traces represent a schematic of the r.m.s. of the lick mimic (top), smoothed firing rate (middle) and the VCN unit recording (bottom; scale bar: 30 μV). (E) Example DCN unit responses to the lick mimic. Same unit as in panel (C). The unit responded to the lick mimic, but not when this sound was generated as a result of the animal’s own licking behavior. Adapted with permission from Singla et al. (2017).

Modulation of sound-evoked activity is also seen with other movements, including locomotion. In contrast to the facilitatory effect of locomotion on visual cortical responses (Niell and Stryker, 2010), the activity of auditory cortical neurons is typically suppressed, most likely via corollary discharge signals conveyed by the secondary motor cortex to auditory cortical inhibitory neurons (Schneider et al., 2014). This circuit may therefore serve to cancel the predictable acoustic consequences of locomotion and other types of movement (Schneider et al., 2018). The influence of locomotion is not restricted to the auditory cortex, since sound-evoked activity in both the MGB (McGinley et al., 2015; Williamson et al., 2015) and the IC (Yang et al., 2020) is also suppressed when a mouse runs and, at least in the IC, this is accompanied by a sharpening in frequency selectivity (Yang et al., 2020). The source of these suppressive effects on subcortical auditory responses is unclear, but could reflect somatosensory inputs that are activated during locomotion or descending signals from motor cortical areas (Olthof et al., 2019).

On the other hand, movements tend to co-occur with changes in other physiological parameters (Reimer et al., 2014). Periods of running in rodents usually occur together with eye movements, increasing pupil dilation and rhythmic whisking as the mouse scans the path ahead (Figure 10), an observation consistent with the notion that locomotion is one of several manifestations of a particular brain state and behavioral pattern seen during active exploration. Furthermore, both active whisking (Fanselow and Nicolelis, 1999; Petersen, 2019) and behavioral state (Lee et al., 2020) have been found to affect stimulus processing in the somatosensory cortex, potentially altering its influence on the auditory system. In these circumstances at least, the suppressive effect on auditory responses may not be the result of movement as such, but a consequence of the altered brain state (McGinley et al., 2015) and the purpose of the suppression may not primarily be the cancellation of responses to specific movement-related sounds. Instead, it may reflect a re-allocation of processing resources away from acoustic input and the suppressed auditory cortex toward somatosensory (or visual) cues that provide more useful information about nearby objects while the animal actively explores.

FIGURE 10
www.frontiersin.org

Figure 10. Locomotion and whisking are correlated with brain state. Changes in pupil diameter (in pixels) of a mouse over time with periods of running and whisking indicated by the colors. Increased pupil dilation occurred during running and whisking. Spontaneous whisking often preceded the onset of running and always accompanied locomotion. Whisker velocity was calculated as the maximum movement of a whisker tip in a 30-frame window and was measured as the whisker-tip position change in pixels per frame recorded at 200 fps.

Conclusion

Although most multisensory research involving audition has focused on its links with vision, extensive interactions also take place between auditory and tactile inputs. This is not restricted to situations where sound-emitting objects approach and eventually touch the body, as in a whining mosquito landing on the back of your neck, but also reflects the fact that sounds are frequently produced during haptic interactions with objects in the environment and as a result of other actions, such as vocalizing. The importance of combining and integrating somatosensory and auditory inputs for perception and the regulation of behavior is indicated by the extensive interactions that have been shown to take place between these two sensory modalities, which have now been demonstrated at almost every level of the auditory system.

We are nonetheless still some way from having a circuit-level understanding of how these modalities interact to influence perception. This largely reflects the few studies that have so far explored audio-tactile interactions at the cellular level in animals trained to perform sensory detection or discrimination tasks. Consequently, relatively little work in non-human species has addressed how attention influences the way these stimuli are integrated and used to guide behavior. In humans, most research on the relationship between attention and multisensory processing has focused on interactions between visual and auditory inputs (Talsma et al., 2010). Nevertheless, selective attention has also been shown to affect the functional coupling between the auditory and somatosensory systems. For example, application of transcranial magnetic stimulation over S1 impaired sound frequency discrimination when human participants simultaneously attended to auditory and tactile frequency, but not when attention was directed to the auditory stimulus alone (Convento et al., 2018). This study also illustrates the value of circuit manipulation methods, which are being used increasingly in humans and animals, for demonstrating a causal contribution of specific brain areas to audio-tactile processing and multisensory integration more generally (Yau et al., 2015).

Details of the neural circuitry involved in merging tactile and auditory signals are still being worked out and it remains unclear why some areas are dominated by facilitatory interactions and others by crossmodal suppression, as is the case, for example, in different divisions of the MGB (Figure 8), or why somatosensory inputs should interface with the auditory system at so many processing levels. One possibility is that tighter regulation of auditory processing may be provided by somatosensory inputs to subcortical structures, whereas cortical involvement may enable greater flexibility crucial for adaptive behavior. The role of some of these circuits may therefore become apparent only under certain conditions (when an animal is engaged in a task, rather than passively listening or anesthetized). Indeed, while the somatosensory system can make us aware of events on the body surface and thus inform us about our immediate surroundings in a passive way, its most important function is to process tactile stimulation during natural behaviors. This means taking brain state into account, utilizing tools for real-time tracking of different parts of the body (Mathis et al., 2018), including individual whisker movements (Sehara et al., 2021), and by considering how different behavioral situations or environmental affordances affect processing (Gibson, 1986), when investigating how somatosensory and auditory inputs are integrated in the brain.

Author contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work, and approved it for publication.

Funding

This research was supported by Wellcome through a Principal Research Fellowship (WT108369/Z/2015/Z) to AK and by a Sir Henry Wellcome Postdoctoral Fellowship (224121/Z/21/Z) to ML.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Aglioti, S. M., and Pazzaglia, M. (2010). Representing actions through their sound. Exp. Brain Res. 206, 141–151. doi: 10.1007/s00221-010-2344-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Aitkin, L. M., Dickhaus, H., Schult, W., and Zimmermann, M. (1978). External nucleus of inferior colliculus: Auditory and spinal somatosensory afferents and their interactions. J. Neurophysiol. 41, 837–847. doi: 10.1152/jn.1978.41.4.837

PubMed Abstract | CrossRef Full Text | Google Scholar

Allen, A. E., Procyk, C. A., Brown, T. M., and Lucas, R. J. (2017). Convergence of visual and whisker responses in the primary somatosensory thalamus (ventral posterior medial region) of the mouse. J. Physiol. 595, 865–881. doi: 10.1113/JP272791

PubMed Abstract | CrossRef Full Text | Google Scholar

Anderson, P. W., and Zahorik, P. (2014). Auditory/visual distance estimation: Accuracy and variability. Front. Psychol. 5:1097. doi: 10.3389/fpsyg.2014.01097

PubMed Abstract | CrossRef Full Text | Google Scholar

Ansorge, J., Wu, C., Shore, S. E., and Krieger, P. (2021). Audiotactile interactions in the mouse cochlear nucleus. Sci. Rep. 11:6887. doi: 10.1038/s41598-021-86236-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Atilgan, H., Town, S. M., Wood, K. C., Jones, G. P., Maddox, R. K., Lee, A., et al. (2018). Integration of visual information in auditory cortex promotes auditory scene analysis through multisensory binding. Neuron 97, 640–655.e4. doi: 10.1016/j.neuron.2017.12.034

PubMed Abstract | CrossRef Full Text | Google Scholar

Aytekin, M., Moss, C. F., and Simon, J. Z. (2008). A sensorimotor approach to sound localization. Neural Comput. 20, 603–635. doi: 10.1162/neco.2007.12-05-094

PubMed Abstract | CrossRef Full Text | Google Scholar

Bartlett, E. L., and Smith, P. H. (1999). Anatomic, intrinsic, and synaptic properties of dorsal and ventral division neurons in rat medial geniculate body. J. Neurophysiol. 81, 1999–2016. doi: 10.1152/jn.1999.81.5.1999

PubMed Abstract | CrossRef Full Text | Google Scholar

Bernasconi, F., Noel, J. P., Park, H. D., Faivre, N., Seeck, M., Spinelli, L., et al. (2018). Audio-tactile and peripersonal space processing around the trunk in human parietal and temporal cortex: An intracranial EEG Study. Cereb. Cortex 28, 3385–3397. doi: 10.1093/cercor/bhy156

PubMed Abstract | CrossRef Full Text | Google Scholar

Bieler, M., Xu, X., Marquardt, A., and Hanganu-Opatz, I. L. (2018). Multisensory integration in rodent tactile but not visual thalamus. Sci. Rep. 8:15684. doi: 10.1038/s41598-018-33815-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Bizley, J. K., and King, A. J. (2008). Visual-auditory spatial processing in auditory cortical neurons. Brain Res. 1242, 24–36. doi: 10.1016/j.brainres.2008.02.087

PubMed Abstract | CrossRef Full Text | Google Scholar

Bizley, J. K., Nodal, F. R., Bajo, V. M., Nelken, I., and King, A. J. (2007). Physiological and anatomical evidence for multisensory interactions in auditory cortex. Cereb. Cortex 17, 2172–2189. doi: 10.1093/cercor/bhl128

PubMed Abstract | CrossRef Full Text | Google Scholar

Bonath, B., Noesselt, T., Krauel, K., Tyll, S., Tempelmann, C., and Hillyard, S. A. (2014). Audio-visual synchrony modulates the ventriloquist illusion and its neural/spatial representation in the auditory cortex. NeuroImage 98, 425–434. doi: 10.1016/j.neuroimage.2014.04.077

PubMed Abstract | CrossRef Full Text | Google Scholar

Bordi, F., and LeDoux, J. E. (1994). Response properties of single units in areas of rat auditory thalamus that project to the amygdala. II. Cells receiving convergent auditory and somatosensory inputs and cells antidromically activated by amygdala stimulation. Exp. Brain Res. 98, 275–286. doi: 10.1007/BF00228415

PubMed Abstract | CrossRef Full Text | Google Scholar

Bronkhorst, A. W., and Houtgast, T. (1999). Auditory distance perception in rooms. Nature 397, 517–520. doi: 10.1038/17374

PubMed Abstract | CrossRef Full Text | Google Scholar

Brosch, M., Selezneva, E., and Scheich, H. (2005). Nonauditory events of a behavioral procedure activate auditory cortex of highly trained monkeys. J. Neurosci. 25, 6797–6806. doi: 10.1523/JNEUROSCI.1571-05.2005

PubMed Abstract | CrossRef Full Text | Google Scholar

Brungart, D. S. (1999). Auditory localization of nearby sources. III. Stimulus effects. J. Acoust. Soc. Am. 106, 3589–3602. doi: 10.1121/1.428212

CrossRef Full Text | Google Scholar

Brungart, D. S., and Rabinowitz, W. M. (1999). Auditory localization of nearby sources. Head-related transfer functions. J. Acoust. Soc. Am. 106, 1465–1479. doi: 10.1121/1.427180

CrossRef Full Text | Google Scholar

Bruns, P., and Röder, B. (2010). Tactile capture of auditory localization is modulated by hand posture. Exp. Psychol. 57, 267–274. doi: 10.1027/1618-3169/a000032

PubMed Abstract | CrossRef Full Text | Google Scholar

Bruns, P., Spence, C., and Röder, B. (2011). Tactile recalibration of auditory spatial representations. Exp. Brain Res. 209, 333–344. doi: 10.1007/s00221-011-2543-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Budinger, E., Heil, P., Hess, A., and Scheich, H. (2006). Multisensory processing via early cortical stages: Connections of the primary auditory cortical field with other sensory systems. Neuroscience 143, 1065–1083. doi: 10.1016/j.neuroscience.2006.08.035

PubMed Abstract | CrossRef Full Text | Google Scholar

Caclin, A., Soto-Faraco, S., Kingstone, A., and Spence, C. (2002). Tactile “capture” of audition. Percept. Psychophys. 64, 616–630. doi: 10.3758/bf03194730

PubMed Abstract | CrossRef Full Text | Google Scholar

Cappe, C., and Barone, P. (2005). Heteromodal connections supporting multisensory integration at low levels of cortical processing in the monkey. Eur. J. Neurosci. 22, 2886–2902. doi: 10.1111/j.1460-9568.2005.04462.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Cappe, C., Morel, A., Barone, P., and Rouiller, E. M. (2009). The thalamocortical projection systems in primate: An anatomical support for multisensory and sensorimotor interplay. Cereb. Cortex 19, 2025–2037. doi: 10.1093/cercor/bhn228

PubMed Abstract | CrossRef Full Text | Google Scholar

Cappe, C., Rouiller, E. M., and Barone, P. (2012). “Cortical and thalamic pathways for multisensory and sensorimotor interplay,” in The Neural Bases of Multisensory Processes, eds M. M. Murray and M. T. Wallace (Boca Raton: CRC Press).

Google Scholar

Carandini, M., and Heeger, D. J. (2011). Normalization as a canonical neural computation. Nat. Rev. Neurosci. 13, 51–62. doi: 10.1038/nrn3136

PubMed Abstract | CrossRef Full Text | Google Scholar

Carvell, G. E., and Simons, D. J. (1986). Somatotopic organization of the second somatosensory area (SII) in the cerebral cortex of the mouse. Somatosens. Res. 3, 213–237. doi: 10.3109/07367228609144585

PubMed Abstract | CrossRef Full Text | Google Scholar

Convento, S., Rahman, M. S., and Yau, J. M. (2018). Selective attention gates the interactive crossmodal coupling between perceptual systems. Curr. Biol. 28, 746–752.e5. doi: 10.1016/j.cub.2018.01.021

PubMed Abstract | CrossRef Full Text | Google Scholar

Convento, S., Wegner-Clemens, K. A., and Yau, J. M. (2019). Reciprocal interactions between audition and touch in flutter frequency perception. Multisens. Res. 32, 67–85. doi: 10.1163/22134808-20181334

PubMed Abstract | CrossRef Full Text | Google Scholar

Crabtree, J. W. (2018). Functional diversity of thalamic reticular subnetworks. Front. Syst. Neurosci. 12:41. doi: 10.3389/fnsys.2018.00041

PubMed Abstract | CrossRef Full Text | Google Scholar

Crommett, L. E., Madala, D., and Yau, J. M. (2019). Multisensory perceptual interactions between higher-order temporal frequency signals. J. Exp. Psychol. Gen. 148, 1124–1137. doi: 10.1037/xge0000513

PubMed Abstract | CrossRef Full Text | Google Scholar

Crommett, L. E., Pérez-Bellido, A., and Yau, J. M. (2017). Auditory adaptation improves tactile frequency perception. J. Neurophysiol. 117, 1352–1362. doi: 10.1152/jn.00783.2016

PubMed Abstract | CrossRef Full Text | Google Scholar

Cruikshank, S. J., Edeline, J. M., and Weinberger, N. M. (1992). Stimulation at a site of auditory-somatosensory convergence in the medial geniculate nucleus is an effective unconditioned stimulus for fear conditioning. Behav. Neurosci. 106, 471–483. doi: 10.1037//0735-7044.106.3.471

CrossRef Full Text | Google Scholar

Driver, J., and Noesselt, T. (2008). Multisensory interplay reveals crossmodal influences on ‘sensory-specific’ brain regions, neural responses, and judgments. Neuron 57, 11–23. doi: 10.1016/j.neuron.2007.12.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Eliades, S. J., and Wang, X. (2003). Sensory-motor interaction in the primate auditory cortex during self-initiated vocalizations. J. Neurophysiol. 89, 2194–2207. doi: 10.1152/jn.00627.2002

PubMed Abstract | CrossRef Full Text | Google Scholar

Eliades, S. J., and Wang, X. (2008). Neural substrates of vocalization feedback monitoring in primate auditory cortex. Nature 453, 1102–1106. doi: 10.1038/nature06910

PubMed Abstract | CrossRef Full Text | Google Scholar

Ernst, M. O., and Bülthoff, H. H. (2004). Merging the senses into a robust percept. Trends Cogn. Sci. 8, 162–169. doi: 10.1016/j.tics.2004.02.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Fanselow, E. E., and Nicolelis, M. A. (1999). Behavioral modulation of tactile responses in the rat somatosensory system. J. Neurosci. 19, 7603–7616. doi: 10.1523/JNEUROSCI.19-17-07603.1999

PubMed Abstract | CrossRef Full Text | Google Scholar

Foxe, J. J., Wylie, G. R., Martinez, A., Schroeder, C. E., Javitt, D. C., Guilfoyle, D., et al. (2002). Auditory-somatosensory multisensory processing in auditory association cortex: An fMRI study. J. Neurophysiol. 88, 540–543. doi: 10.1152/jn.2002.88.1.540

PubMed Abstract | CrossRef Full Text | Google Scholar

Froese, T., and González-Grandón, X. (2020). How passive is passive listening? Toward a sensorimotor theory of auditory perception. Phenom. Cogn. Sci. 19, 619–651. doi: 10.1007/s11097-019-09641-6

CrossRef Full Text | Google Scholar

Fu, K. M., Johnston, T. A., Shah, A. S., Arnold, L., Smiley, J., Hackett, T. A., et al. (2003). Auditory cortical neurons respond to somatosensory stimulation. J. Neurosci. 23, 7510–7515. doi: 10.1523/JNEUROSCI.23-20-07510.2003

PubMed Abstract | CrossRef Full Text | Google Scholar

Gescheider, G. A., Bolanowski, S. J., Pope, J. V., and Verrillo, R. T. (2002). A four-channel analysis of the tactile sensitivity of the fingertip: Frequency selectivity, spatial summation, and temporal summation. Somatosens. Mot. Res. 19, 114–124. doi: 10.1080/08990220220131505

PubMed Abstract | CrossRef Full Text | Google Scholar

Ghazanfar, A. A., and Schroeder, C. E. (2006). Is neocortex essentially multisensory? Trends Cogn. Sci. 10, 278–285. doi: 10.1016/j.tics.2006.04.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Gibson, J. J. (1986). The Ecological Approach to Visual Perception. New York, NY: Psychology Press, doi: 10.4324/9780203767764

PubMed Abstract | CrossRef Full Text | Google Scholar

Gillmeister, H., and Eimer, M. (2007). Tactile enhancement of auditory detection and perceived loudness. Brain Res. 1160, 58–68. doi: 10.1016/j.brainres.2007.03.041

PubMed Abstract | CrossRef Full Text | Google Scholar

Gobbelé, R., Schürmann, M., Forss, N., Juottonen, K., Buchner, H., and Hari, R. (2003). Activation of the human posterior parietal and temporoparietal cortices during audiotactile interaction. NeuroImage 20, 503–511. doi: 10.1016/s1053-8119(03)00312-4

CrossRef Full Text | Google Scholar

Godenzini, L., Alwis, D., Guzulaitis, R., Honnuraiah, S., Stuart, G. J., and Palmer, L. M. (2021). Auditory input enhances somatosensory encoding and tactile goal-directed behavior. Nat. Commun. 12:4509. doi: 10.1038/s41467-021-24754-w

PubMed Abstract | CrossRef Full Text | Google Scholar

Graziano, M. S., and Cooke, D. F. (2006). Parieto-frontal interactions, personal space, and defensive behavior. Neuropsychologia 44, 2621–2635. doi: 10.1016/j.neuropsychologia.2005.09.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Graziano, M. S., and Gross, C. G. (1998). Spatial maps for the control of movement. Curr. Opin. Neurobiol. 8, 195–201. doi: 10.1016/s0959-4388(98)80140-2

CrossRef Full Text | Google Scholar

Graziano, M. S., Reiss, L. A., and Gross, C. G. (1999). A neuronal representation of the location of nearby sounds. Nature 397, 428–430. doi: 10.1038/17115

PubMed Abstract | CrossRef Full Text | Google Scholar

Groh, J. M., Trause, A. S., Underhill, A. M., Clark, K. R., and Inati, S. (2001). Eye position influences auditory responses in primate inferior colliculus. Neuron 29, 509–518. doi: 10.1016/s0896-6273(01)00222-7

CrossRef Full Text | Google Scholar

Gruters, K. G., and Groh, J. M. (2012). Sounds and beyond: Multisensory and other non-auditory signals in the inferior colliculus. Front. Neural Circuits 6:96. doi: 10.3389/fncir.2012.00096

PubMed Abstract | CrossRef Full Text | Google Scholar

Guo, W., Clause, A. R., Barth-Maron, A., and Polley, D. B. (2017). A corticothalamic circuit for dynamic switching between feature detection and discrimination. Neuron 95, 180.e–194.e. doi: 10.1016/j.neuron.2017.05.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Hackett, T. A., Smiley, J. F., Ulbert, I., Karmos, G., Lakatos, P., de la Mothe, L. A., et al. (2007a). Sources of somatosensory input to the caudal belt areas of auditory cortex. Perception 36, 1419–1430. doi: 10.1068/p5841

PubMed Abstract | CrossRef Full Text | Google Scholar

Hackett, T. A., De La Mothe, L. A., Ulbert, I., Karmos, G., Smiley, J., and Schroeder, C. E. (2007b). Multisensory convergence in auditory cortex, II. Thalamocortical connections of the caudal superior temporal plane. J. Comp. Neurol. 502, 924–952. doi: 10.1002/cne.21326

PubMed Abstract | CrossRef Full Text | Google Scholar

Harris, L. R., Carnevale, M. J., D’Amour, S., Fraser, L. E., Harrar, V., Hoover, A. E., et al. (2015). How our body influences our perception of the world. Front. Psychol. 6:819. doi: 10.3389/fpsyg.2015.00819

PubMed Abstract | CrossRef Full Text | Google Scholar

Hartline, P. H., Vimal, R. L., King, A. J., Kurylo, D. D., and Northmore, D. P. (1995). Effects of eye position on auditory localization and neural representation of space in superior colliculus of cats. Exp. Brain Res. 104, 402–408. doi: 10.1007/BF00231975

PubMed Abstract | CrossRef Full Text | Google Scholar

Heeringa, A. N., Wu, C., and Shore, S. E. (2018). Multisensory integration enhances temporal coding in ventral cochlear nucleus bushy cells. J. Neurosci. 38, 2832–2843. doi: 10.1523/JNEUROSCI.2244-17.2018

PubMed Abstract | CrossRef Full Text | Google Scholar

Herholz, S. C., and Zatorre, R. J. (2012). Musical training as a framework for brain plasticity: Behavior, function, and structure. Neuron 76, 486–502. doi: 10.1016/j.neuron.2012.10.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Hoefer, M., Tyll, S., Kanowski, M., Brosch, M., Schoenfeld, M. A., Heinze, H. J., et al. (2013). Tactile stimulation and hemispheric asymmetries modulate auditory perception and neural responses in primary auditory cortex. NeuroImage 79, 371–382. doi: 10.1016/j.neuroimage.2013.04.119

PubMed Abstract | CrossRef Full Text | Google Scholar

Holmes, N. P., Martin, D., Mitchell, W., Noorani, Z., and Thorne, A. (2020). Do sounds near the hand facilitate tactile reaction times? Four experiments and a meta-analysis provide mixed support and suggest a small effect size. Exp. Brain Res. 238, 995–1009. doi: 10.1007/s00221-020-05771-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Ito, T., Tiede, M., and Ostry, D. J. (2009). Somatosensory function in speech perception. Proc. Natl. Acad. Sci. U S A. 106, 1245–1248. doi: 10.1073/pnas.0810063106

PubMed Abstract | CrossRef Full Text | Google Scholar

Iurilli, G., Ghezzi, D., Olcese, U., Lassi, G., Nazzaro, C., Tonini, R., et al. (2012). Sound-driven synaptic inhibition in primary visual cortex. Neuron 73, 814–828. doi: 10.1016/j.neuron.2011.12.026

PubMed Abstract | CrossRef Full Text | Google Scholar

Jain, R., and Shore, S. (2006). External inferior colliculus integrates trigeminal and acoustic information: Unit responses to trigeminal nucleus and acoustic stimulation in the guinea pig. Neurosci. Lett. 395, 71–75. doi: 10.1016/j.neulet.2005.10.077

PubMed Abstract | CrossRef Full Text | Google Scholar

Jay, M. F., and Sparks, D. L. (1984). Auditory receptive fields in primate superior colliculus shift with changes in eye position. Nature 309, 345–347. doi: 10.1038/309345a0

PubMed Abstract | CrossRef Full Text | Google Scholar

Jousmäki, V., and Hari, R. (1998). Parchment-skin illusion: Sound-biased touch. Curr. Biol. 8:R190. doi: 10.1016/s0960-9822(98)70120-4

CrossRef Full Text | Google Scholar

Kanold, P. O., and Young, E. D. (2001). Proprioceptive information from the pinna provides somatosensory input to cat dorsal cochlear nucleus. J. Neurosci. 21, 7848–7858. doi: 10.1523/JNEUROSCI.21-19-07848.2001

PubMed Abstract | CrossRef Full Text | Google Scholar

Kayser, C. (2009). Phase resetting as a mechanism for supramodal attentional control. Neuron 64, 300–302. doi: 10.1016/j.neuron.2009.10.022

PubMed Abstract | CrossRef Full Text | Google Scholar

Kayser, C., Logothetis, N. K., and Panzeri, S. (2010). Visual enhancement of the information representation in auditory cortex. Curr. Biol. 20, 19–24. doi: 10.1016/j.cub.2009.10.068

PubMed Abstract | CrossRef Full Text | Google Scholar

Kayser, C., Petkov, C. I., Augath, M., and Logothetis, N. K. (2005). Integration of touch and sound in auditory cortex. Neuron 48, 373–384. doi: 10.1016/j.neuron.2005.09.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Kayser, C., Petkov, C. I., and Logothetis, N. K. (2008). Visual modulation of neurons in auditory cortex. Cereb. Cortex 18, 1560–1574. doi: 10.1093/cercor/bhm187

PubMed Abstract | CrossRef Full Text | Google Scholar

Keough, M., Derrick, D., and Gick, B. (2019). Cross-modal effects in speech perception. Annu. Rev. Linguist. 5, 49–66. doi: 10.1146/annurev-linguistics-011718-012353

PubMed Abstract | CrossRef Full Text | Google Scholar

Khorevin, V. I. (1980). Effect of electrodermal stimulation on single unit responses to acoustic stimulation in the parvocellular part of the medial geniculate body. Neurophysiology 12, 129–134. doi: 10.1007/BF01065307

CrossRef Full Text | Google Scholar

Khorevin, V. I., and Shelest, I. I. (1998). Interaction of influences from the auditory and somatosensory afferent systems on neurons of the primary auditory cortex. Neurophysiology 30, 383–385. doi: 10.1007/BF03027684

CrossRef Full Text | Google Scholar

Kim, D. O., Zahorik, P., Carney, L. H., Bishop, B. B., and Kuwada, S. (2015). Auditory distance coding in rabbit midbrain neurons and human perception: Monaural amplitude modulation depth as a cue. J. Neurosci. 35, 5360–5372. doi: 10.1523/JNEUROSCI.3798-14.2015

PubMed Abstract | CrossRef Full Text | Google Scholar

Kimura, A. (2017). Robust interactions between the effects of auditory and cutaneous electrical stimulations on cell activities in the thalamic reticular nucleus. Brain Res. 1661, 49–66. doi: 10.1016/j.brainres.2017.02.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Kimura, A., and Imbe, H. (2018). Robust subthreshold cross-modal modulation of auditory response by cutaneous electrical stimulation in first- and higher-order auditory thalamic nuclei. Neuroscience 372, 161–180. doi: 10.1016/j.neuroscience.2017.12.051

PubMed Abstract | CrossRef Full Text | Google Scholar

King, A. J. (2004). The superior colliculus. Curr. Biol. 14, R335–R338. doi: 10.1016/j.cub.2004.04.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Kolarik, A., Cirstea, S., and Pardhan, S. (2013). Discrimination of virtual auditory distance using level and direct-to-reverberant ratio cues. J. Acoust. Soc. Am. 134, 3395–3398. doi: 10.1121/1.4824395

CrossRef Full Text | Google Scholar

Kolarik, A. J., Moore, B. C., Zahorik, P., Cirstea, S., and Pardhan, S. (2016). Auditory distance perception in humans: A review of cues, development, neuronal bases, and effects of sensory loss. Atten. Percept. Psychophys. 78, 373–395. doi: 10.3758/s13414-015-1015-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Kopco, N., Doreswamy, K. K., Huang, S., Rossi, S., and Ahveninen, J. (2020). Cortical auditory distance representation based on direct-to-reverberant energy ratio. NeuroImage 208:116436. doi: 10.1016/j.neuroimage.2019.116436

PubMed Abstract | CrossRef Full Text | Google Scholar

Kopěo, N., Huang, S., Belliveau, J. W., Raij, T., Tengshe, C., and Ahveninen, J. (2012). Neuronal representations of distance in human auditory cortex. Proc. Natl. Acad. Sci. U S A. 109, 11019–11024. doi: 10.1073/pnas.1119496109

PubMed Abstract | CrossRef Full Text | Google Scholar

Kopčo, N., and Shinn-Cunningham, B. G. (2011). Effect of stimulus spectrum on distance perception for nearby sources. J. Acoust. Soc. Am. 130, 1530–1541. doi: 10.1121/1.3613705

CrossRef Full Text | Google Scholar

Kuchenbuch, A., Paraskevopoulos, E., Herholz, S. C., and Pantev, C. (2014). Audio-tactile integration and the influence of musical training. PLoS One 9:e85743. doi: 10.1371/journal.pone.0085743

PubMed Abstract | CrossRef Full Text | Google Scholar

Kumpik, D. P., and King, A. J. (2019). A review of the effects of unilateral hearing loss on spatial hearing. Hear. Res. 372, 17–28. doi: 10.1016/j.heares.2018.08.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Kuwada, S., Kim, D. O., Koch, K. J., Abrams, K. S., Idrobo, F., Zahorik, P., et al. (2015). Near-field discrimination of sound source distance in the rabbit. J. Assoc. Res. Otolaryngol. 16, 255–262. doi: 10.1007/s10162-014-0505-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Lakatos, P., Chen, C. M., O’Connell, M. N., Mills, A., and Schroeder, C. E. (2007). Neuronal oscillations and multisensory interaction in primary auditory cortex. Neuron 53, 279–292. doi: 10.1016/j.neuron.2006.12.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Landry, S. P., and Champoux, F. (2017). Musicians react faster and are better multisensory integrators. Brain Cogn. 111, 156–162. doi: 10.1016/j.bandc.2016.12.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Landry, S. P., Sharp, A., Pagé, S., and Champoux, F. (2017). Temporal and spectral audiotactile interactions in musicians. Exp. Brain Res. 235, 525–532. doi: 10.1007/s00221-016-4813-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Lederman, S. J. (1979). Auditory texture perception. Perception 8, 93–103. doi: 10.1068/p080093

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, C., Kheradpezhouh, E., Diamond, M. E., and Arabzadeh, E. (2020). State-dependent changes in perception and coding in the mouse somatosensory cortex. Cell Rep. 32:108197. doi: 10.1016/j.celrep.2020.108197

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, H., and Noppeney, U. (2011). Long-term music training tunes how the brain temporally binds signals from multiple senses. Proc. Natl. Acad. Sci. U S A. 108, E1441–E1450. doi: 10.1073/pnas.1115267108

PubMed Abstract | CrossRef Full Text | Google Scholar

Lemus, L., Hernández, A., Luna, R., Zainos, A., and Romo, R. (2010). Do sensory cortices process more than one sensory modality during perceptual judgments? Neuron 67, 335–348. doi: 10.1016/j.neuron.2010.06.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Lesicko, A. M., Hristova, T. S., Maigler, K. C., and Llano, D. A. (2016). Connectional modularity of top-down and bottom-up multimodal inputs to the lateral cortex of the mouse inferior colliculus. J. Neurosci. 36, 11037–11050. doi: 10.1523/JNEUROSCI.4134-15.2016

PubMed Abstract | CrossRef Full Text | Google Scholar

Liang, M., Mouraux, A., Hu, L., and Iannetti, G. D. (2013). Primary sensory cortices contain distinguishable spatial patterns of activity for each sense. Nat. Commun. 4:1979. doi: 10.1038/ncomms2979

PubMed Abstract | CrossRef Full Text | Google Scholar

Lima, C. F., Krishnan, S., and Scott, S. K. (2016). Roles of supplementary motor areas in auditory processing and auditory imagery. Trends Neurosci. 39, 527–542. doi: 10.1016/j.tins.2016.06.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, M., Dai, J., Zhou, M., Liu, J., Ge, X., Wang, N., et al. (2022). Mini-review: The neural circuits of the non-lemniscal inferior colliculus. Neurosci. Lett. 776:136567. doi: 10.1016/j.neulet.2022.136567

PubMed Abstract | CrossRef Full Text | Google Scholar

Lohse, M., Bajo, V. M., and King, A. J. (2019). Development, organization and plasticity of auditory circuits: Lessons from a cherished colleague. Eur. J. Neurosci. 49, 990–1004. doi: 10.1111/ejn.13979

PubMed Abstract | CrossRef Full Text | Google Scholar

Lohse, M., Dahmen, J. C., Bajo, V. M., and King, A. J. (2021). Subcortical circuits mediate communication between primary sensory cortical areas in mice. Nat. Commun. 12:3916. doi: 10.1038/s41467-021-24200-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Loomis, J. M., Klatzky, R. L., Philbeck, J. W., and Golledge, R. G. (1998). Assessing auditory distance perception using perceptually directed action. Percept. Psychophys. 60, 966–980. doi: 10.3758/bf03211932

PubMed Abstract | CrossRef Full Text | Google Scholar

Lütkenhöner, B., Lammertmann, C., Simões, C., and Hari, R. (2002). Magnetoencephalographic correlates of audiotactile interaction. NeuroImage 15, 509–522. doi: 10.1006/nimg.2001.0991

PubMed Abstract | CrossRef Full Text | Google Scholar

Macaluso, E., and Maravita, A. (2010). The representation of space near the body through touch and vision. Neuropsychologia 48, 782–795. doi: 10.1016/j.neuropsychologia.2009.10.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Maier, J. X., Chandrasekaran, C., and Ghazanfar, A. A. (2008). Integration of bimodal looming signals through neuronal coherence in the temporal lobe. Curr. Biol. 18, 963–968. doi: 10.1016/j.cub.2008.05.043

PubMed Abstract | CrossRef Full Text | Google Scholar

Mathis, A., Mamidanna, P., Cury, K. M., Abe, T., Murthy, V. N., Mathis, M. W., et al. (2018). DeepLabCut: Markerless pose estimation of user-defined body parts with deep learning. Nat. Neurosci. 21, 1281–1289. doi: 10.1038/s41593-018-0209-y

PubMed Abstract | CrossRef Full Text | Google Scholar

McGinley, M. J., David, S. V., and McCormick, D. A. (2015). Cortical membrane potential signature of optimal states for sensory signal detection. Neuron 87, 179–192. doi: 10.1016/j.neuron.2015.05.038

PubMed Abstract | CrossRef Full Text | Google Scholar

Meijer, G. T., Mertens, P., Pennartz, C., Olcese, U., and Lansink, C. S. (2019). The circuit architecture of cortical multisensory processing: Distinct functions jointly operating within a common anatomical network. Prog. Neurobiol. 174, 1–15. doi: 10.1016/j.pneurobio.2019.01.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Meredith, M. A., and Allman, B. L. (2015). Single-unit analysis of somatosensory processing in the core auditory cortex of hearing ferrets. Eur. J. Neurosci. 41, 686–698. doi: 10.1111/ejn.12828

PubMed Abstract | CrossRef Full Text | Google Scholar

Meredith, M. A., and Stein, B. E. (1986). Visual, auditory, and somatosensory convergence on cells in superior colliculus results in multisensory integration. J. Neurophysiol. 56, 640–662. doi: 10.1152/jn.1986.56.3.640

PubMed Abstract | CrossRef Full Text | Google Scholar

Mo, C., and Sherman, S. M. (2019). A sensorimotor pathway via higher-order thalamus. J. Neurosci. 39, 692–704. doi: 10.1523/JNEUROSCI.1467-18.2018

PubMed Abstract | CrossRef Full Text | Google Scholar

Moore, D. R., and King, A. J. (1999). Auditory perception: The near and far of sound localization. Curr. Biol. 9, R361–R363. doi: 10.1016/s0960-9822(99)80227-9

CrossRef Full Text | Google Scholar

Morrill, R. J., and Hasenstaub, A. R. (2018). Visual information present in infragranular layers of mouse auditory cortex. J. Neurosci. 38, 2854–2862. doi: 10.1523/JNEUROSCI.3102-17.2018

PubMed Abstract | CrossRef Full Text | Google Scholar

Münte, T. F., Altenmüller, E., and Jäncke, L. (2002). The musician’s brain as a model of neuroplasticity. Nat. Rev. Neurosci. 3, 473–478. doi: 10.1038/nrn843

PubMed Abstract | CrossRef Full Text | Google Scholar

Murray, M. M., Molholm, S., Michel, C. M., Heslenfeld, D. J., Ritter, W., Javitt, D. C., et al. (2005). Grabbing your ear: Rapid auditory-somatosensory multisensory interactions in low-level sensory cortices are not constrained by stimulus alignment. Cereb. Cortex 15, 963–974. doi: 10.1093/cercor/bhh197

PubMed Abstract | CrossRef Full Text | Google Scholar

Niell, C. M., and Stryker, M. P. (2010). Modulation of visual responses by behavioral state in mouse visual cortex. Neuron 65, 472–479. doi: 10.1016/j.neuron.2010.01.033

PubMed Abstract | CrossRef Full Text | Google Scholar

Nordmark, P. F., Pruszynski, J. A., and Johansson, R. S. (2012). BOLD responses to tactile stimuli in visual and auditory cortex depend on the frequency content of stimulation. J. Cogn. Neurosci. 24, 2120–2134. doi: 10.1162/jocn_a_00261

CrossRef Full Text | Google Scholar

Occelli, V., Spence, C., and Zampini, M. (2011). Audiotactile interactions in temporal perception. Psychon. Bull. Rev. 18, 429–454. doi: 10.3758/s13423-011-0070-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Ohshiro, T., Angelaki, D. E., and DeAngelis, G. C. (2011). A normalization model of multisensory integration. Nat. Neurosci. 14, 775–782. doi: 10.1038/nn.2815

PubMed Abstract | CrossRef Full Text | Google Scholar

Olthof, B., Rees, A., and Gartside, S. E. (2019). Multiple nonauditory cortical regions innervate the auditory midbrain. J. Neurosci. 39, 8916–8928. doi: 10.1523/JNEUROSCI.1436-19.2019

PubMed Abstract | CrossRef Full Text | Google Scholar

Opoku-Baah, C., Schoenhaut, A. M., Vassall, S. G., Tovar, D. A., Ramachandran, R., and Wallace, M. T. (2021). Visual influences on auditory behavioral, neural, and perceptual processes: A review. J. Assoc. Res. Otolaryngol. 22, 365–386. doi: 10.1007/s10162-021-00789-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Pantev, C., Lappe, C., Herholz, S. C., and Trainor, L. (2009). Auditory-somatosensory integration and cortical plasticity in musical training. Annals N. Y. Acad. Sci. 1169, 143–150. doi: 10.1111/j.1749-6632.2009.04588.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Pantev, C., Paraskevopoulos, E., Kuchenbuch, A., Lu, Y., and Herholz, S. C. (2015). Musical expertise is related to neuroplastic changes of multisensory nature within the auditory cortex. Eur. J. Neurosci. 41, 709–717. doi: 10.1111/ejn.12788

PubMed Abstract | CrossRef Full Text | Google Scholar

Paus, T., Perry, D. W., Zatorre, R. J., Worsley, K. J., and Evans, A. C. (1996). Modulation of cerebral blood flow in the human auditory cortex during speech: Role of motor-to-sensory discharges. Eur. J. Neurosci. 8, 2236–2246. doi: 10.1111/j.1460-9568.1996.tb01187.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Pérez-Bellido, A., Anne Barnes, K., Crommett, L. E., and Yau, J. M. (2018). Auditory frequency representations in human somatosensory cortex. Cereb. Cortex 28, 3908–3921. doi: 10.1093/cercor/bhx255

PubMed Abstract | CrossRef Full Text | Google Scholar

Petersen, C. C. H. (2019). Sensorimotor processing in the rodent barrel cortex. Nat. Rev. Neurosci. 20, 533–546. doi: 10.1038/s41583-019-0200-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Petrini, K., Dahl, S., Rocchesso, D., Waadeland, C. H., Avanzini, F., Puce, A., et al. (2009). Multisensory integration of drumming actions: Musical expertise affects perceived audiovisual asynchrony. Exp. Brain Res. 198, 339–352. doi: 10.1007/s00221-009-1817-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Pick, H. L., Warren, D. H., and Hay, J. C. (1969). Sensory conflict in judgments of spatial direction. Percept. Psychophys. 6, 203–205. doi: 10.3758/BF03207017

CrossRef Full Text | Google Scholar

Populin, L. C., Tollin, D. J., and Yin, T. C. (2004). Effect of eye position on saccades and neuronal responses to acoustic stimuli in the superior colliculus of the behaving cat. J. Neurophysiol. 92, 2151–2167. doi: 10.1152/jn.00453.2004

PubMed Abstract | CrossRef Full Text | Google Scholar

Radeau, M., and Bertelson, P. (1974). The after-effects of ventriloquism. Q. J. Exp. Psychol. 26, 63–71. doi: 10.1080/14640747408400388

PubMed Abstract | CrossRef Full Text | Google Scholar

Rahman, M. S., Barnes, K. A., Crommett, L. E., Tommerdahl, M., and Yau, J. M. (2020). Auditory and tactile frequency representations are co-embedded in modality-defined cortical sensory systems. NeuroImage 215:116837. doi: 10.1016/j.neuroimage.2020.116837

PubMed Abstract | CrossRef Full Text | Google Scholar

Rao, R. P., Mielke, F., Bobrov, E., and Brecht, M. (2014). Vocalization-whisking coordination and multisensory integration of social signals in rat auditory cortex. eLife 3:e03185. doi: 10.7554/eLife.03185

PubMed Abstract | CrossRef Full Text | Google Scholar

Recanzone, G. H. (1998). Rapidly induced auditory plasticity: The ventriloquism aftereffect. Proc. Natl. Acad. Sci. U S A. 95, 869–875. doi: 10.1073/pnas.95.3.869

PubMed Abstract | CrossRef Full Text | Google Scholar

Reimer, J., Froudarakis, E., Cadwell, C. R., Yatsenko, D., Denfield, G. H., and Tolias, A. S. (2014). Pupil fluctuations track fast switching of cortical states during quiet wakefulness. Neuron 84, 355–362. doi: 10.1016/j.neuron.2014.09.033

PubMed Abstract | CrossRef Full Text | Google Scholar

Ro, T., Ellmore, T. M., and Beauchamp, M. S. (2013). A neural link between feeling and hearing. Cereb. Cortex 23, 1724–1730. doi: 10.1093/cercor/bhs166

PubMed Abstract | CrossRef Full Text | Google Scholar

Ro, T., Hsu, J., Yasar, N. E., Elmore, L. C., and Beauchamp, M. S. (2009). Sound enhances touch perception. Exp. Brain Res. 195, 135–143. doi: 10.1007/s00221-009-1759-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Sanabria, D., Soto-Faraco, S., and Spence, C. (2005). Assessing the effect of visual and tactile distractors on the perception of auditory apparent motion. Exp. Brain Res. 166, 548–558. doi: 10.1007/s00221-005-2395-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Schneider, D. M., Nelson, A., and Mooney, R. (2014). A synaptic and circuit basis for corollary discharge in the auditory cortex. Nature 513, 189–194. doi: 10.1038/nature13724

PubMed Abstract | CrossRef Full Text | Google Scholar

Schneider, D. M., Sundararajan, J., and Mooney, R. (2018). A cortical filter that learns to suppress the acoustic consequences of movement. Nature 561, 391–395. doi: 10.1038/s41586-018-0520-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Schroeder, C. E., Lakatos, P., Kajikawa, Y., Partan, S., and Puce, A. (2008). Neuronal oscillations and visual amplification of speech. Trends Cogn. Sci. 12, 106–113. doi: 10.1016/j.tics.2008.01.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Schulz, M., Ross, B., and Pantev, C. (2003). Evidence for training-induced crossmodal reorganization of cortical functions in trumpet players. Neuroreport 14, 157–161. doi: 10.1097/00001756-200301200-00029

PubMed Abstract | CrossRef Full Text | Google Scholar

Schürmann, M., Caetano, G., Hlushchuk, Y., Jousmäki, V., and Hari, R. (2006). Touch activates human auditory cortex. NeuroImage 30, 1325–1331. doi: 10.1016/j.neuroimage.2005.11.020

PubMed Abstract | CrossRef Full Text | Google Scholar

Schürmann, M., Caetano, G., Jousmäki, V., and Hari, R. (2004). Hands help hearing: Facilitatory audiotactile interaction at low sound-intensity levels. J. Acoust. Soc. Am. 115, 830–832. doi: 10.1121/1.1639909

CrossRef Full Text | Google Scholar

Sehara, K., Zimmer-Harwood, P., Larkum, M. E., and Sachdev, R. (2021). Real-time closed-loop feedback in behavioral time scales using DeepLabCut. eNeuro 8, ENEURO.415–ENEURO.420. doi: 10.1523/ENEURO.0415-20.2021

PubMed Abstract | CrossRef Full Text | Google Scholar

Seifritz, E., Neuhoff, J. G., Bilecen, D., Scheffler, K., Mustovic, H., Schächinger, H., et al. (2002). Neural processing of auditory looming in the human brain. Curr. Biol. 12, 2147–2151. doi: 10.1016/s0960-9822(02)01356-8

CrossRef Full Text | Google Scholar

Serino, A. (2019). Peripersonal space (PPS) as a multisensory interface between the individual and the environment, defining the space of the self. Neurosci. Biobehav. Rev. 99, 138–159. doi: 10.1016/j.neubiorev.2019.01.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Shore, S. E., and Zhou, J. (2006). Somatosensory influence on the cochlear nucleus and beyond. Hear. Res. 21, 90–99. doi: 10.1016/j.heares.2006.01.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Singla, S., Dempsey, C., Warren, R., Enikolopov, A. G., and Sawtell, N. B. (2017). A cerebellum-like circuit in the auditory system cancels responses to self-generated sounds. Nat. Neurosci. 20, 943–950. doi: 10.1038/nn.4567

PubMed Abstract | CrossRef Full Text | Google Scholar

Sofroniew, N. J., and Svoboda, K. (2015). Whisking. Curr. Biol. 25, R137–R140. doi: 10.1016/j.cub.2015.01.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Song, Y. H., Kim, J. H., Jeong, H. W., Choi, I., Jeong, D., Kim, K., et al. (2017). A neural circuit for auditory dominance over visual perception. Neuron 93, 940–954.e6. doi: 10.1016/j.neuron.2017.01.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Soto-Faraco, S., and Deco, G. (2009). Multisensory contributions to the perception of vibrotactile events. Behav. Brain Res. 196, 145–154. doi: 10.1016/j.bbr.2008.09.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Soto-Faraco, S., Spence, C., and Kingstone, A. (2004). Congruency effects between auditory and tactile motion: Extending the phenomenon of cross-modal dynamic capture. Cogn. Affect. Behav. Neurosci. 4, 208–217. doi: 10.3758/cabn.4.2.208

PubMed Abstract | CrossRef Full Text | Google Scholar

Sperdin, H. F., Cappe, C., Foxe, J. J., and Murray, M. M. (2009). Early, low-level auditory-somatosensory multisensory interactions impact reaction time speed. Front. Integr. Neurosci. 3:2. doi: 10.3389/neuro.07.002.2009

PubMed Abstract | CrossRef Full Text | Google Scholar

Tajadura-Jiménez, A., Kitagawa, N., Väljamäe, A., Zampini, M., Murray, M. M., and Spence, C. (2009). Auditory-somatosensory multisensory interactions are spatially modulated by stimulated body surface and acoustic spectra. Neuropsychologia 47, 195–203. doi: 10.1016/j.neuropsychologia.2008.07.025

PubMed Abstract | CrossRef Full Text | Google Scholar

Talsma, D., Senkowski, D., Soto-Faraco, S., and Woldorff, M. G. (2010). The multifaceted interplay between attention and multisensory integration. Trends Cogn. Sci. 14, 400–410. doi: 10.1016/j.tics.2010.06.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Teneggi, C., Canzoneri, E., di Pellegrino, G., and Serino, A. (2013). Social modulation of peripersonal space boundaries. Curr. Biol. 23, 406–411. doi: 10.1016/j.cub.2013.01.043

PubMed Abstract | CrossRef Full Text | Google Scholar

Villanueva, L., and Zampini, M. (2018). Reciprocal interference between audition and touch in the perception of duration. Multisens. Res. 31, 351–371. doi: 10.1163/22134808-00002583

PubMed Abstract | CrossRef Full Text | Google Scholar

Von Békésy, G. (1959). Similarities between hearing and skin sensations. Psychol. Rev. 66, 1–22. doi: 10.1037/h0046967

PubMed Abstract | CrossRef Full Text | Google Scholar

Werner-Reiss, U., Kelly, K. A., Trause, A. S., Underhill, A. M., and Groh, J. M. (2003). Eye position affects activity in primary auditory cortex of primates. Curr. Biol. 13, 554–562. doi: 10.1016/s0960-9822(03)00168-4

CrossRef Full Text | Google Scholar

Wigderson, E., Nelken, I., and Yarom, Y. (2016). Early multisensory integration of self and source motion in the auditory system. Proc. Natl. Acad. Sci. U S A. 113, 8308–8313. doi: 10.1073/pnas.1522615113

PubMed Abstract | CrossRef Full Text | Google Scholar

Williamson, R. S., Hancock, K. E., Shinn-Cunningham, B. G., and Polley, D. B. (2015). Locomotion and task demands differentially modulate thalamic audiovisual processing during active search. Curr. Biol. 25, 1885–1891. doi: 10.1016/j.cub.2015.05.045

PubMed Abstract | CrossRef Full Text | Google Scholar

Wilson, E. C., Braida, L. D., and Reed, C. M. (2010a). Perceptual interactions in the loudness of combined auditory and vibrotactile stimuli. J. Acoust. Soc. Am. 127, 3038–3043. doi: 10.1121/1.3377116

CrossRef Full Text | Google Scholar

Wilson, E. C., Reed, C. M., and Braida, L. D. (2010b). Integration of auditory and vibrotactile stimuli: Effects of frequency. J. Acoust. Soc. Am. 127, 3044–3059. doi: 10.1121/1.3365318

CrossRef Full Text | Google Scholar

Wu, C., Stefanescu, R. A., Martel, D. T., and Shore, S. E. (2015). Listening to another sense: Somatosensory integration in the auditory system. Cell Tissue Res. 361, 233–250. doi: 10.1007/s00441-014-2074-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, Y., Lee, J., and Kim, G. (2020). Integration of locomotion and auditory signals in the mouse inferior colliculus. eLife 9:e52228. doi: 10.7554/eLife.52228

PubMed Abstract | CrossRef Full Text | Google Scholar

Yarrow, K., Haggard, P., and Rothwell, J. C. (2008). Vibrotactile–auditory interactions are post-perceptual. Perception 37, 1114–1130. doi: 10.1068/p5824

PubMed Abstract | CrossRef Full Text | Google Scholar

Yau, J. M., DeAngelis, G. C., and Angelaki, D. E. (2015). Dissecting neural circuits for multisensory integration and crossmodal processing. Phil. Trans. R. Soc. Lond. B Biol. Sci. 370:20140203. doi: 10.1098/rstb.2014.0203

PubMed Abstract | CrossRef Full Text | Google Scholar

Yau, J. M., Olenczak, J. B., Dammann, J. F., and Bensmaia, S. J. (2009). Temporal frequency channels are linked across audition and touch. Curr. Biol. 19, 561–566. doi: 10.1016/j.cub.2009.02.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Yau, J. M., Weber, A. I., and Bensmaia, S. J. (2010). Separate mechanisms for audio-tactile pitch and loudness interactions. Front. Psychol. 1:160. doi: 10.3389/fpsyg.2010.00160

PubMed Abstract | CrossRef Full Text | Google Scholar

Yu, J. J., and Young, E. D. (2000). Linear and nonlinear pathways of spectral information transmission in the cochlear nucleus. Proc. Natl. Acad. Sci. U S A. 97, 11780–11786. doi: 10.1073/pnas.97.22.11780

PubMed Abstract | CrossRef Full Text | Google Scholar

Zahorik, P. (2001). Estimating sound source distance with and without vision. Optom. Vis. Sci. 78, 270–275. doi: 10.1097/00006324-200105000-00009

PubMed Abstract | CrossRef Full Text | Google Scholar

Zahorik, P. (2002). Assessing auditory distance perception using virtual acoustics. J. Acoust. Soc. Am. 111, 1832–1846. doi: 10.1121/1.1458027

CrossRef Full Text | Google Scholar

Zhang, M., Kwon, S. E., Ben-Johny, M., O’Connor, D. H., and Issa, J. B. (2020). Spectral hallmark of auditory-tactile interactions in the mouse somatosensory cortex. Commun. Biol. 3:64. doi: 10.1038/s42003-020-0788-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, Y. D., and Fuster, J. M. (2004). Somatosensory cell response to an auditory cue in a haptic memory task. Behav. Brain Res. 153, 573–578. doi: 10.1016/j.bbr.2003.12.024

PubMed Abstract | CrossRef Full Text | Google Scholar

Zierul, B., Röder, B., Tempelmann, C., Bruns, P., and Noesselt, T. (2017). The role of auditory cortex in the spatial ventriloquism aftereffect. NeuroImage 162, 257–268. doi: 10.1016/j.neuroimage.2017.09.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Zwiers, M. P., Versnel, H., and Van Opstal, A. J. (2004). Involvement of monkey inferior colliculus in spatial hearing. J. Neurosci. 24, 4145–4156. doi: 10.1523/JNEUROSCI.0199-04.2004

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: auditory, somatosensory, multisensory integration, movement, cortex, thalamus, midbrain, perception

Citation: Lohse M, Zimmer-Harwood P, Dahmen JC and King AJ (2022) Integration of somatosensory and motor-related information in the auditory system. Front. Neurosci. 16:1010211. doi: 10.3389/fnins.2022.1010211

Received: 02 August 2022; Accepted: 28 September 2022;
Published: 18 October 2022.

Edited by:

Duck O. Kim, University of Connecticut Health Center, United States

Reviewed by:

Melville Wohlgemuth, University of Arizona, United States
Jeffrey M. Yau, Baylor College of Medicine, United States
Ximena González-Grandon, Ibero-American University, Mexico

Copyright © 2022 Lohse, Zimmer-Harwood, Dahmen and King. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Andrew J. King, andrew.king@dpag.ox.ac.uk

Present Address: Michael Lohse, Sainsbury Wellcome Centre, London, United Kingdom

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.