Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 29 September 2021
Sec. Perception Science

Visual Perceptual Load Attenuates Age-Related Audiovisual Integration in an Audiovisual Discrimination Task

\r\nYanna Ren&#x;Yanna Ren1†Hannan Li&#x;Hannan Li2†Yan LiYan Li1Tao WangTao Wang3Weiping Yang*Weiping Yang4*
  • 1Department of Psychology, College of Humanities and Management, Guizhou University of Traditional Chinese Medicine, Guiyang, China
  • 2University Science Park Management Center, Guiyang University, Guiyang, China
  • 3Department of Light and Chemical Engineering, Guizhou Light Industry Technical College, Guiyang, China
  • 4Department of Psychology, Faculty of Education, Hubei University, Wuhan, China

Previous studies confirmed that the cognitive resources are limited for each person, and perceptual load affects the detection of stimulus greatly; however, how the visual perceptual load influences audiovisual integration (AVI) is still unclear. Here, 20 older and 20 younger adults were recruited to perform an auditory/visual discrimination task under various visual perceptual-load conditions. The analysis for the response times revealed a significantly faster response to the audiovisual stimulus than to the visual stimulus or auditory stimulus (all p < 0.001), and a significantly slower response by the older adults than by the younger adults to all targets (all p ≤ 0.024). The race-model analysis revealed a higher AV facilitation effect for older (12.54%) than for younger (7.08%) adults under low visual perceptual-load conditions; however, no obvious difference was found between younger (2.92%) and older (3.06%) adults under medium visual perceptual-load conditions. Only the AV depression effect was found for both younger and older adults under high visual perceptual-load conditions. Additionally, the peak latencies of AVI were significantly delayed in older adults under all visual perceptual-load conditions. These results suggested that visual perceptual load decreased AVI (i.e., depression effects), and the AVI effect was increased but delayed for older adults.

Introduction

Individuals simultaneously receive many types of sensory information from the outside world, including visual, auditory, tactile, olfactory, and gustatory information; however, our brain can select and merge the available information to facilitate the perception of the outside world. For example, when driving a car, it is necessary to watch the road condition (visual information), listen to the voice broadcast (auditory information), and operate the steering wheel (tactile information) simultaneously. The interactive processing of information from multiple sensory modalities is called multisensory integration. Individuals acquire most of the information from the environment through visual and auditory modalities, and numerous studies have found that responses to audiovisual stimulus were faster and more accurate than those to auditory-only or visual-only stimulus (Meredith et al., 1987; Stein and Meredith, 1993; Stein, 2012). The procedure that merges visual and auditory information is defined as audiovisual integration (AVI), which assists individuals in identifying objects much more easily and is the topical issue of multisensory integration (Raab, 1962; Miller, 1982). Ordinary life is complex and AVI might be disturbed by various distractors from the surrounding environment. Furthermore, for each person, the cognitive resources are limited, that is, if the cognitive demand is higher for one task, less will be left to process other tasks, known as “perceptual load theory” (Kahneman, 1973; Lavie and Tsal, 1994; Lavie, 1995). Under different perceptual loads, how individuals efficiently integrate visual and auditory information has been a hot multisensory research topic in recent years (Alsius et al., 2005, 2007; Baseler et al., 2014; Wahn and König, 2015; Ren et al., 2020a).

Alsius et al. (2005, 2014) and Ren et al. (2020a) investigated the additional perceptual load of visual transients on AVI. In studies by Alsius et al. (2005, 2014) McGurk stimuli were used in audiovisual redundant tasks to assess AVI, and meaningless checkerboard images and white noise were used in the study by Ren et al. (2020a). In the low perceptual-load condition, participants were instructed to respond only to the audiovisual redundant task but to respond to the visual distractor at the same time in the high perceptual-load condition. Consistent results showed that the AVI was lower in the condition with high perceptual load of visual transients (dual task) than in the condition with a low perceptual load of visual transients (single task). In the studies by Alsius et al. (2005, 2014) and Ren et al. (2020a), the visual distractors were transitorily presented and accompanied by stimuli in an audiovisual redundant task. To investigate the influence of sustained visual distractors on AVI, Wahn and König (2015) instructed participants to continuously track visual moving balls when performing the audiovisual redundancy task, and their results revealed that AVI was comparable under high sustained visual perceptual-load conditions and low sustained visual perceptual-load conditions, indicating that sustained visual perceptual load did not disrupt AVI. Dual tasks were employed in the abovementioned studies, in which one task was used to assess AVI and the other was used to manipulate visual perceptual load, and the visual perceptual load was from the additional distractor, during which the low perceptual-load condition involved focused attention but high perceptual-load condition involved divided attention. Therefore, it is difficult to evidence whether the perceptual load influences AVI or attention clearly. To clarify this question, the current study aimed to investigate how the visual perceptual load (demand) from the visual stimulus itself influences AVI in a single task.

Macdonald and Lavie (2011) found that visual perceptual load severely affected the auditory perception and even induced “inattentional deafness,” showing that the participant failed to notice the tone (∼79%) during the visual detection task under high visual perceptual load (Macdonald and Lavie, 2011). Diederich et al. (2008) proposed that AVI occurred only when the processing of auditory and visual information was completed in a specific time frame. Considering the limitation of nerve-center energy proposed by Kahneman (1973) in “Attention and Effort,” with increasing visual perceptual load, more perceptual resources occupied by visual stimuli and the processing of auditory stimuli will undoubtedly be limited. Colonius and Diederich (2004) reported that the AVI effect was greater when the intensities of visual and auditory stimuli were equalized, and it is relatively lower if they were not coordinated (Colonius and Diederich, 2004). Therefore, we hypothesized that the AVI effect is higher for equalized pairs (low visual perceptual-load conditions) than for incoordinate pairs (high visual perceptual-load conditions). The hypothesis was tested using a visual/auditory discrimination task with different visual perceptual-load stimuli. If the AVI was lower under the higher visual perceptual-load conditions than under the low or medium visual perceptual-load conditions, it would be concluded that the visual perceptual load greatly affected the AVI and that the visual perceptual load reduced the AVI effect.

Additionally, aging is an important factor that influences AVI, and some researchers reported that the AVI was higher for older adults than for younger adults (Laurienti et al., 2006; Peiffer et al., 2007; Stephen et al., 2010; Ren et al., 2020b), but how AVI changed with the alteration of visual perceptual load for older adults is also still unclear. Recently, some of them proposed that higher AVI in older adults is an adaptive mechanism to compensate for unimodal sensory decline (Laurienti et al., 2006; Peiffer et al., 2007; Freiherr et al., 2013; Ren et al., 2020c), and the audiovisual perceptual training could improve cognitive function of healthy older adults (Anguera et al., 2013; Yang et al., 2018; O’Brien et al., 2020) and patients with mild cognitive impairment (Lee et al., 2020). Therefore, it is necessary to investigate the effect of the perceptual load of visual stimuli on AVI to clarify the interaction of perceptual load and AVI, which might further offer a reference for the selection of experimental materials in cognitive training tasks. Therefore, another aim of the present study was to investigate the aging effect on the interaction between AVI and visual perceptual load. Considering that auditory and visual stimulus processing is generally slower in older adults (Grady, 2008; Anderson, 2019), but they are still able to correctly perceive the outside world, we hypothesized that a compensatory mechanism also existed for older adults during AVI even under visual perceptual-load conditions. However, one of the well-documented effects of aging is that there is a reduction in the total amount of resources available for information processing (Cabeza et al., 2006; Paige and Gutchess, 2015); thus, performance might be even worse in high visual perceptual-load condition. In the current study, the hypothesis was tested by comparing AVI between older and younger adults under all visual perceptual-load conditions. If the AVI effect of the older adults is higher than that of younger adults under any visual perceptual-load conditions, it would be concluded that the older adults might establish a compensatory mechanism.

Materials and Methods

Subjects

The sample size was calculated using the GPower 3.1.9.2 program.1 The total sample size was 36, with an effect size f of 0.4 and a power (1-β err prob) of 0.8 (Lakens, 2021). Therefore, 20 healthy older adults (59–73 years, mean age ± SD, 63 ± 4) and 20 younger college students (20–23 years, mean age ± SD, 22 ± 1) were recruited to participate in this study. All older adults were recruited from Guiyang City, and all younger adults were undergraduates. All participants had normal hearing, had a normal or corrected-to-normal vision, had no color blindness or color weakness, and were not informed of the aim of this study. Those participants whose mini-mental state examination (MMSE) scores were out of the range of 25–30 were excluded from the experiment (Bleecker et al., 1988). All the participants signed the informed consent form approved by the Ethics Committee of the Second Affiliated Hospital of Guizhou University of Traditional Chinese Medicine, and obtained the remuneration for their time.

Stimuli

The visual stimuli included a red/blue circle and square with a 2.6-cm diameter single- or dual-border (Bruner et al., 1956; Figure 1B). The auditory stimuli contained a 60-dB white noise for auditory target stimuli and 60-dB 1,000-Hz sinusoidal tone for auditory non-target stimuli, which were edited using Audacity 2.3.02 for a duration of 150 ms (10 ms of rise or/fall cosine gate) (Ren et al., 2016, 2018). The experiment contained three sessions, namely, a low visual perceptual-load session, a medium visual perceptual-load session, and a high visual perceptual-load session. For low visual perceptual-load session, the visual target stimuli were red circles and squares with single- or dual-border, and the visual non-target stimuli were blue circles and squares with single- or dual-border (Figure 1B). For medium visual perceptual-load session, the visual target stimuli were red circles with single or dual borders, and the visual non-target stimuli included red squares with single or dual borders and blue circles and squares with single or dual borders (Figure 1B). For high visual perceptual-load session, the visual target stimuli were red circles with dual borders, and the visual non-target stimuli included red circles with single borders, red squares with single or dual borders, and blue circles and squares with single or dual borders (Figure 1B). Under all of the visual perceptual-load conditions, the audiovisual target stimuli were the simultaneous presentation of the auditory target stimulus and visual target stimulus, and the audiovisual non-target stimulus was the simultaneous presentation of the auditory non-target stimulus and visual non-target stimulus. The following two conditions: the visual non-target stimulus accompanied by auditory target stimulus and the visual target stimulus accompanied by auditory non-target stimulus were not included.

FIGURE 1
www.frontiersin.org

Figure 1. Experimental design. (A) A possible sequence for the auditory non-target, audiovisual target, and visual non-target stimuli in the experiment. (B) Stimuli types. Low load, low visual perceptual-load condition; Medium load, medium visual perceptual-load condition; High load, high visual perceptual-load condition.

Procedure

The experiment was conducted in a dimly lit and sound-attenuated room (Cognitive Psychology Laboratory, Guizhou University of Traditional Chinese Medicine, China). The presentation for all stimuli and collection for behavioral response were controlled by E-prime 2.0 (Psychology Software Tools, Inc., Pittsburgh, PA, United States). The visual stimulus (V) was presented on the screen of the computer in front of the participant with 60-cm distance, and the background of the monitor (Dell, E2213c) was gray during the experiment. The auditory stimulus (A) was presented through two speakers (Edifier, R19U) located on the right and the left of the screen. Each session started with a 3,000-ms fixation, and then, all stimuli (A,V,AV) were presented for 150 ms randomly with a randomized interstimulus interval (ISI) of 1,500–2,200 ms (Figure 1A). The participant was instructed only to respond to target stimuli by pressing the right button of the mouse as accurately and rapidly as possible, but withhold responses for non-target stimuli. In total, three sessions were conducted randomly, and each session lasted for 10 min with suitable rest according to the physical condition of the participant individually. Each session contained 240 trials, including 120 target and 120 non-target stimuli, with 40 trials for each stimulus type.

Data Analysis

The accuracy of responding is the percentage of correct responses (the response time falling within the average time duration ± 2.5 SD) relative to the total number of target stimuli. The response times (RTs) and accuracy for each participant under each condition were calculated separately, and then, submitted to a 2group (older, younger) × 3load (low, medium, high) × 3stimulus (A,V,AV) ANOVA with Greenhouse–Geisser corrections followed by a post hoc analysis.

To interpret the phenomenon that the response to AV stimulus was obviously faster than to A or V stimulus, the separate-activation model and coactivation model were raised (Meredith et al., 1987; Stein and Meredith, 1993; Stein, 2012). The separate-activation model is that the processing of auditory and visual information never combined, and the response was induced by the winner of the race, and so it was also called “race model.” However, the coactivation model is that the activation of auditory and visual was combined and this induced the response cooperatively. The race model was calculated based on the cumulative distribution functions (CDFs) of visual-only and auditory-only response in 10-ms time bins, P(RM) = [P(V) + P(A)]-P(V) × P(A) (Miller, 1982, 1986; Laurienti et al., 2006). P(A) and P(V) were the response probability to a visual-only or auditory-only trial within a given timeframe, respectively, and P(RM) is the predicted response probability for AV trial basing on P(A) and P(V) (Figures 2A,B). If P(AV) is significantly different from P(RM), which suggests the AV violated the race model and the co-activation model was applied, then the AVI was assumed to have occurred (t-test, p ≤ 0.05). If P(AV) was significantly greater than P(RM), it was defined as AV facilitation, otherwise, AV depression (Meredith et al., 1987; Stein and Meredith, 1993). As previously studied, the AVI effect was assessed using race model by analyzing the RTs data (Laurienti et al., 2006; Peiffer et al., 2007; Stevenson et al., 2014; Ren et al., 2016, 2018). A probability-difference curve was generated by subtracting the P(RM) of an individual from his/her audiovisual CDFs [P(AV)] in each 10-ms time bin, and the peak value (peak benefit) of the curve was used to assess AVI ability (Xu et al., 2020). The time frame from the stimulus onset to peak benefit was peak latency, which is an important index to evaluate when AVI occurred (Laurienti et al., 2006; Peiffer et al., 2007; Ren et al., 2016, 2018). SPSS version 19.0 software (SPSS, Tokyo, Japan) was used for all statistical analyses.

FIGURE 2
www.frontiersin.org

Figure 2. Cumulative distribution functions of auditory stimuli, visual stimuli, race model, and audiovisual stimuli for older (A) and younger (B) adults under low visual perceptual-load conditions. Low_older, low visual perceptual-load condition for older adults; Low_younger, low visual perceptual-load condition for younger adults.

Results

Accuracy

As shown in Figure 3B, the accuracy for each participant and each condition was higher than 80%. The 2group (older, younger) × 3load (low, medium, high) × 3stimulus (A,V,AV) ANOVA analysis revealed a significant group main effect [F(1, 38) = 9.289, p = 0.004, ηp2 = 0.196], showing that the accuracy was lower for older adults than that for younger adults. There was a significant perceptual load main effect [F(2, 76) = 7.026, p = 0.003, ηp2 = 0.156], showing that the accuracy under low and medium visual perceptual-load conditions was higher than that under high visual perceptual-load conditions. Additionally, a significant stimulus type main effect was also found [F(2, 76) = 13.038, p < 0.001, ηp2 = 0.255], showing that the accuracy was higher when responding to bimodal AV stimulus than unimodal V or A stimulus (AV > V > A). In addition, the interaction of perceptual load × group was marginally significant [F(2, 76) = 3.368, p = 0.049, ηp2 = 0.081]. The post hoc analysis using pairwise comparison with Bonferroni correction for perceptual load revealed that the accuracy was lower under high visual perceptual-load conditions than that under low (p = 0.004) and medium (p = 0.018) visual perceptual-load conditions, but there was no significant difference between the low and medium visual perceptual-load conditions (p = 1.000) for older adults. The post hoc analysis using pairwise comparison with Bonferroni correction for the group revealed that the accuracy for older adults was lower than that for younger adults under both low (p = 0.003) and medium (p = 0.009) visual perceptual-load conditions, but no significant difference was found between older and younger adults (p = 0.168) under high visual perceptual-load conditions.

FIGURE 3
www.frontiersin.org

Figure 3. Higher response time (A) and lower accuracy (B) for older adults than for younger adults under all visual perceptual-load conditions. *p < 0.05, **p < 0.01, ***p < 0.001. Low load, low visual perceptual-load condition; Medium load, medium visual perceptual-load condition; High load, high visual perceptual-load condition.

Response Times

The 2group (older, younger) × 3load (low, medium, high) × 3stimulus (A,V,AV) ANOVA analysis for RTs (Figure 3A) revealed significant group main effect [F(1, 38) = 5.528, p = 0.024, ηp2 = 0.127], showing that the response was slower by older adults than that by younger adults. There were significant perceptual load main effects [F(2, 76) = 166.072, p < 0.001, ηp2 = 0.814], showing that the response under the low visual perceptual-load conditions was fastest (low > medium > high). Additionally, significant stimulus type main effect was also found [F(2, 76) = 105.521, p < 0.001, ηp2 = 0.156], showing that the response to AV stimulus was faster than that to A or V stimulus (AV > V > A). The interaction of perceptual load × group was marginally significant [F(2, 76) = 3.324, p = 0.048, ηp2 = 0.080]. The post hoc analysis using pairwise comparison with Bonferroni correction revealed that the response by older adults was obviously slower than that by younger adults under low (p = 0.028) and high (p = 0.009) visual perceptual-load conditions, but no significant difference was found between younger and older adults under medium visual perceptual-load conditions (p = 0.125). Additionally, under all perceptual-load conditions, the response was slower by older adults than that by younger adults (all p ≥ 0.002). The interaction of perceptual load × stimulus was also significant [F(4, 152) = 18.927, p < 0.001, ηp2 = 0.332]. The post hoc analysis using pairwise comparison with Bonferroni correction for perceptual load revealed that under low and medium visual perceptual-load conditions, the response to the three stimuli were significantly different (AV > V > A, all p ≤ 0.004). However, under the high visual perceptual-load conditions, the response to A stimulus was slower than that to AV (p < 0.001) or V (p < 0.001) stimulus, but no significant difference was found between AV and V stimuli (p = 1.000). Additionally, the response among all visual perceptual-load conditions was significantly different for all stimuli (low > medium > high, all p < 0.001).

Race Model Analysis

The AVI was assessed using race model under all visual perceptual-load conditions, as Figure 2A for older adults and Figure 2B for younger adults under low perceptual-load conditions. Significant AVI was found under all visual perceptual-load conditions for both older (two-tailed t-test, all p ≤ 0.047) and younger (two-tailed t-test, all p ≤ 0.05) adults. The AV facilitation effect for older adults (12.54%) was higher than that for younger adults (7.08%) under low visual perceptual-load conditions (Figure 4A), but under medium visual perceptual-load conditions no significant difference was found between older (3.06%) and younger adults (2.92%) (Figure 4B). Under high visual perceptual-load conditions, only the AV depression effect was found in both older (−11.47%) and younger adults (−16.89%) (Figure 4C). These results indicated the AVI was higher in older adults, and the facilitation effect was absent and even the depression effect occurred with an increase in the visual perceptual load for both older (Figure 5A) and younger (Figure 5B) adults. Additionally, the peak latencies were delayed for older adults under all visual perceptual-load conditions than that for younger adults, exhibiting 420 vs. 300 ms, 430 vs. 350 ms, and 850 vs. 600 ms for low, medium, and high visual perceptual-load conditions, respectively (Figure 4D), indicating delayed AVI for older adults than that for younger adults.

FIGURE 4
www.frontiersin.org

Figure 4. Significant audiovisual facilitation effect was mainly found in the low visual perceptual-load conditions (A), but depression effects were found in the medium (B) and high (C) visual perceptual-load conditions. The AVI was delayed for older adults than that for younger adults under all visual perceptual-load conditions. (D) Low, low visual perceptual-load condition; Medium, medium visual perceptual-load condition; High, high visual perceptual-load condition. ***p < 0.001.

FIGURE 5
www.frontiersin.org

Figure 5. The AVI effect changed from facilitation to depression for both older (A) and younger (B) adults with increasing of visual perceptual load. Low, low visual perceptual-load condition; Medium, medium visual perceptual-load condition; High, high visual perceptual-load condition.

Discussion

The current study aims to investigate the influence of visual perceptual load on AVI and its aging effect. The results showed that the AV facilitation effect was higher for older adults than that for younger adults under low visual perceptual-load conditions, and no obvious difference was found between younger and older adults under medium visual perceptual-load conditions; however, only the AV depression effect was found for both younger and older adults under high visual perceptual-load conditions. Additionally, the AVI was delayed in older adults compared with younger adults under all visual perceptual-load conditions.

Unexpectedly, the AVI effect was altered from facilitation to depression with the increasing visual perceptual load. According to the nerve-center energy theory proposed by Kahneman (1973)in “Attention and Effort” (1973), the complex stimulus can expend more energy and the simple stimulus can expend less energy. The visual task is a simple “feature identification” (single feature) under low visual perceptual-load conditions, but it is complex “object identification” (conjunction of features) under medium (two features) and high (three features) visual perceptual-load conditions. Therefore, more energy was allocated to the visual stimulus under medium visual perceptual-load conditions than that under low visual perceptual-load conditions, and more under high visual perceptual-load conditions than that under medium visual perceptual-load conditions. Additionally, a recent study reported that AVI was sensitive to the amount of available cognitive resources (Michail et al., 2021). In the current study, the memory loads were different among perceptual-load conditions, showing the highest memory load under high perceptual-load conditions (three attributes) and the lowest memory load under low perceptual-load conditions (one attribute). Therefore, more cognitive resources are allocated to the visual stimulus under medium visual perceptual-load conditions than that under low visual perceptual-load conditions, and more under high visual perceptual-load conditions than that under medium visual perceptual-load conditions. The multisensory integration occurred in the human primary visual cortex (Murray et al., 2016). The increased visual demands could, theoretically, limit resources for integrative processing. With increasing perceptual load, more resources were shifted to visual stimuli, which reduced auditory weighting, and therefore limited the benefits of integration (Deloss et al., 2013). Therefore, it is reasonable that the AV facilitation effect was mainly found under low visual perceptual-load conditions and it decreased with the increasing visual perceptual load. However, there is a limitation that it was impossible to rule out of the role of memory load, and precise experimental designs are necessary for future studies. The biased competition model assumes that nerve-center energy enhances the selected sensory neural response and suppresses the irrelevant sensory neural response. With the additional visual perceptual load, more energy was diverted to the visual stimulus, which also led to AV suppression in addition to facilitation (Kamijo et al., 2007). Therefore, the AV depression effect under the medium and high visual perceptual-load conditions might be mainly attributed to the biased competition.

In addition, the response was slower with increasing visual perceptual load. Only one attribute (red color) was identified under low visual perceptual-load conditions; however, two attributes (red color and circle) were identified under medium visual perceptual-load conditions, and three attributes (red color, circle, double frame) were identified under high visual perceptual-load conditions. The capacity of the visual system for processing information is limited, and the performance was worse if the subject was instructed to identify two attributes than that to one attribute (Duncan, 1984; Kastner and Ungerleider, 2001). Therefore, the reduced response speed might be mainly attributed to the limited processing resources. Additionally, under the low visual perceptual-load conditions, the red color attribute is quickly and effortlessly detected, resulting from its salience, but it cannot be completed through the simple feature property under the medium and high visual perceptual-load conditions, which leads to a slower response. Therefore, another possible reason for the slower response with increasing visual perceptual load is the weakened salience for object identification than that for the simple feature.

Additionally, consistent with our original hypothesis, the AVI effect was higher in older adults than that in younger adults under low visual perceptual load conditions, which was also consistent with some of the previous studies reporting a higher AVI effect in older adults (Laurienti et al., 2006; Peiffer et al., 2007; Diederich et al., 2008; Deloss et al., 2013; Sekiyama et al., 2014). Although the primary brain regions for AVI are the same in older and younger adults, Grady (2009) reported that older adults recruited additional brain areas to participate in AVI (Grady, 2009). Diaconescu et al. (2013) used MEG to record the responses of subjects (15 younger adults and 16 elderly adults) to semantically related bimodal (V + A) and unimodal (V or A) sensory stimuli to capture the different brain areas involved in AVI. They reported that older individuals activated a specific brain network, the medial prefrontal cortex or posterior parietal cortex, when responding to audiovisual information. A comparative study between older and younger adults was also conducted by Ren et al. (2018) using an auditory/visual discrimination task, and they found activity in the visual cortex during AV stimuli processing in older but not in younger adults, which also indicated that older adults recruited the primary visual cortex involved in AVI. Additionally, reductions in hemispheric asymmetry in older adults have been found in many cognitive tasks, including episodic memory retrieval (Grady et al., 2002), perception (Grady et al., 1994, 2000), and working memory (Cabeza, 2002; Methqal et al., 2017). Therefore, the higher AVI effect of older adults, as a possible adaptive mechanism, might have mainly resulted from decreased brain specialization and activation of a distinct brain network to compensate for dysfunction in visual-only or auditory-only stimulus processing compared with younger adults (Gazzaniga et al., 2008; Koelewijn et al., 2010; Diaconescu et al., 2013; Baseler et al., 2014; Gutchess, 2014; Setti et al., 2014; Ren et al., 2018). However, the current studies could not identify whether the higher AVI is an adaptive mechanism or an indicator for unisensory decline, and future neuroimaging studies are necessary to clarify this matter. In addition, the AVI was significantly delayed in older adults compared with younger adults under all visual perceptual-load conditions, which was consistent with previous studies (Ren et al., 2018, 2020b). According to the “time-window-of-integration model” proposed by Diederich et al. (2008) before the integration of auditory and visual information (second stage), it is necessary to complete early auditory and visual information processing, which was assumed to be independent (first stage). There were significant sensory declines in older adults, showing slowed auditory and visual information processing in the first stage (Salthouse, 1996; Hirst et al., 2019); therefore, the delayed AVI with aging might be mainly attributed to age-related sensory decline.

Conclusion

In conclusion, resulting from biased competition, the AVI effect was reduced with increasing visual perceptual load for both older and younger adults. It is possible that as a compensatory mechanism of the unimodal perceptual decline, the AVI of older adults was higher than that of younger adults, but their AVI effect was delayed resulting from age-related sensory decline.

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author/s.

Ethics Statement

The studies involving human participants were reviewed and approved by the Ethics Committee of the Second Affiliated Hospital of Guizhou University of Traditional Chinese Medicine. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

YR and WY conceived and designed the experiments. YL and TW collected the data. YR analyzed the data and wrote the draft manuscript. HL revised the manuscript and responsed to the reviewers by receiving comments for WY. All authors contributed to the article and approved the submitted version.

Funding

This study was partially supported by the Science and Technology Planning Project of Guizhou Province (QianKeHeJiChu-ZK [2021] General 120), the National Natural Science Foundation of China (Nos. 31800932 and 31700973), the Innovation and Entrepreneurship Project for High-level Overseas Talent of Guizhou Province [(2019)04], and the Humanity and Social Science Youth Foundation of the Ministry of Education of China (Nos. 18XJC190003 and 16YJC190025).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We would like to thank the individuals who participated in this study and as also the reviewers for their helpful comments and suggestions on revised versions of this manuscript.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2021.740221/full#supplementary-material

Footnotes

  1. ^ http://stats.idre.ucla.edu/other/gpower/
  2. ^ https://audacity.onl/

References

Anderson, N. D. (2019). Cognitive neuroscience of aging. J. Gerontol. Ser. B 74, 1083–1085.

Google Scholar

Alsius, A., Möttönen, R., Sams, M. E., Soto-Faraco, S., and Tiippana, K. (2014). Effect of attentional load on audiovisual speech perception: evidence from ERPs. Front. Psychol. 5:727.

Google Scholar

Alsius, A., Navarra, J., Campbell, R., and Soto-Faraco, S. (2005). Audiovisual integration of speech falters under high attention demands. Curr. Biol. 15, 839–843. doi: 10.1016/j.cub.2005.03.046

PubMed Abstract | CrossRef Full Text | Google Scholar

Alsius, A., Navarra, J., and Soto-Faraco, S. (2007). Attention to touch weakens audiovisual speech integration. Exp. Brain Res. 183, 399–404. doi: 10.1007/s00221-007-1110-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Anguera, J. A., Boccanfuso, J., Rintoul, J. L., Al-Hashimi, O., Faraji, F., Janowich, J., et al. (2013). Video game training enhances cognitive control in older adults. Nature 501, 97–101. doi: 10.1038/nature12486

PubMed Abstract | CrossRef Full Text | Google Scholar

Baseler, H. A., Harris, R. J., Young, A. W., and Andrews, T. J. (2014). Neural responses to expression and gaze in the posterior superior temporal sulcus interact with facial identity. Cereb. Cortex 24, 737–744. doi: 10.1093/cercor/bhs360

PubMed Abstract | CrossRef Full Text | Google Scholar

Bleecker, M. L., Bolla-Wilson, K., Kawas, C., and Agnew, J. (1988). Age-specific norms for the mini-mental state exam. Neurology 38, 1565–1568. doi: 10.1212/wnl.38.10.1565

PubMed Abstract | CrossRef Full Text | Google Scholar

Bruner, J. S., Goodnow, J. J., and Austin, G. A. (1956). A study of thinking. Philos. Phenomenol. Res. 7, 215–221.

Google Scholar

Cabeza, R. (2002). Hemispheric asymmetry reduction in older adults: the HAROLD model. Psychol. Aging 17, 85–100. doi: 10.1037/0882-7974.17.1.85

PubMed Abstract | CrossRef Full Text | Google Scholar

Cabeza, R., Nyberg, L., and Park, D. (2006). Cognitive neuroscience of aging: linking cognitive and cerebral aging. Psychologist 163, 560–561.

Google Scholar

Colonius, H., and Diederich, A. (2004). Multisensory interaction in saccadic reaction time: a time-window-of-integration model. J. Cogn. Neurosci. 16, 1000–1009. doi: 10.1162/0898929041502733

PubMed Abstract | CrossRef Full Text | Google Scholar

Deloss, D. J., Pierce, R. S., and Andersen, G. J. (2013). Multisensory integration, aging, and the sound-induced flash illusion. Psychol. Aging 28, 802–812. doi: 10.1037/a0033289

PubMed Abstract | CrossRef Full Text | Google Scholar

Diaconescu, A. O., Hasher, L., and McIntosh, A. R. (2013). Visual dominance and multisensory integration changes with age. Neuroimage 65, 152–166. doi: 10.1016/j.neuroimage.2012.09.057

PubMed Abstract | CrossRef Full Text | Google Scholar

Diederich, A., Colonius, H., and Schomburg, A. (2008). Assessing age-related multisensory enhancement with the time-window-of-integration model. Neuropsychologia 46, 2556–2562. doi: 10.1016/j.neuropsychologia.2008.03.026

PubMed Abstract | CrossRef Full Text | Google Scholar

Duncan, J. (1984). Selective attention and the organization of visual information. J. Exp. Psychol. Gen. 113, 501–517. doi: 10.1037/0096-3445.113.4.501

PubMed Abstract | CrossRef Full Text | Google Scholar

Freiherr, J., Lundström, J. N., Habel, U., and Reetz, K. (2013). Multisensory integration mechanisms during aging. Front. Hum. Neurosci. 7:863.

Google Scholar

Gazzaniga, M. S., Ivry, R. B., and Mangun, G. R. (2008). Cognitive Neuroscience: the Biology of the Mind. California: MIT Press.

Google Scholar

Grady, C. L. (2008). Cognitive neuroscience of aging. Ann. N. Y. Acad. Sci. 1124, 127–144.

Google Scholar

Grady, C. L. (2009). Functional Neuroimaging Studies of Aging. Encyclopedia of Neuroscience. Pittsburgh: Academic Press.

Google Scholar

Grady, C. L., Bernstein, L. J., Beig, S., and Siegenthaler, A. L. (2002). The effects of encoding task on age-related differences in the functional neuroanatomy of face memory. Psychol. Aging 17, 7–23. doi: 10.1037/0882-7974.17.1.7

PubMed Abstract | CrossRef Full Text | Google Scholar

Grady, C. L., Maisog, J. M., Horwitz, B., Ungerleider, L. G., Mentis, M. J., Salerno, J. A., et al. (1994). Age-related changes in cortical blood flow activation during visual processing of faces and location. J. Neurosci. 14(3 Pt 2), 1450–1462. doi: 10.1523/jneurosci.14-03-01450.1994

PubMed Abstract | CrossRef Full Text | Google Scholar

Grady, C. L., McIntosh, A. R., Horwitz, B., and Rapoport, S. I. (2000). Age-related changes in the neural correlates of degraded and nondegraded face processing. Cogn. Neuropsychol. 17, 165–186. doi: 10.1080/026432900380553

PubMed Abstract | CrossRef Full Text | Google Scholar

Gutchess, A. (2014). Plasticity of the aging brain: new directions in cognitive neuroscience. Science 346, 579–582. doi: 10.1126/science.1254604

PubMed Abstract | CrossRef Full Text | Google Scholar

Hirst, R. J., Setti, A., Kenny, R. A., and Newell, F. N. (2019). Age-related sensory decline mediates the Sound-Induced flash illusion: evidence for reliability weighting models of multisensory perception. Sci. Rep. 9:19347.

Google Scholar

Kahneman, D. (1973). Attention and Effort. Englewood Cliffs, NJ: Prentice-Hall.

Google Scholar

Kamijo, K., Nishihira, Y., Higashiura, T., and Kuroiwa, K. (2007). The interactive effect of exercise intensity and task difficulty on human cognitive processing. Int. J. Psychophysiol. 65, 114–121. doi: 10.1016/j.ijpsycho.2007.04.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Kastner, S., and Ungerleider, L. G. (2001). The neural basis of biased competition in human visual cortex. Neuropsychologia 39, 1263–1276. doi: 10.1016/s0028-3932(01)00116-6

CrossRef Full Text | Google Scholar

Koelewijn, T., Bronkhorst, A., and Theeuwes, J. (2010). Attention and the multiple stages of multisensory integration: a review of audiovisual studies. Acta Psychol. 134, 372–384. doi: 10.1016/j.actpsy.2010.03.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Lakens, D. (2021). Sample Size Justification. Available Online at: https://psyarxiv.com/9d3yf/ (accessed January 4, 2021).

Google Scholar

Laurienti, P. J., Burdette, J. H., Maldjian, J. A., and Wallace, M. T. (2006). Enhanced multisensory integration in older adults. Neurobiol. Aging 27, 1155–1163. doi: 10.1016/j.neurobiolaging.2005.05.024

PubMed Abstract | CrossRef Full Text | Google Scholar

Lavie, N. (1995). Perceptual load as a necessary condition for selective attention. J. Exp. Psychol. Hum. Percept. Perform. 21, 451–468. doi: 10.1037/0096-1523.21.3.451

PubMed Abstract | CrossRef Full Text | Google Scholar

Lavie, N., and Tsal, Y. (1994). Perceptual load as a major determinant of the locus of selection in visual attention. Percept. Psychophys. 56, 183–197. doi: 10.3758/bf03213897

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, L. P., Har, A. W., Ngai, C. H., Lai, D. W. L., Lam, B. Y., and Chan, C. C. (2020). Audiovisual integrative training for augmenting cognitive- motor functions in older adults with mild cognitive impairment. BMC Geriatr. 20:64.

Google Scholar

Macdonald, J. S. P., and Lavie, N. (2011). Visual perceptual load induces inattentional deafness. Atten. Percept. Psychophys. 73, 1780–1789. doi: 10.3758/s13414-011-0144-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Meredith, M. A., Nemitz, J. W., and Stein, B. E. (1987). Determinants of multisensory integration in superior colliculus neurons. I. temporal factors. J. Neurosci. 7, 3215–3229. doi: 10.1523/jneurosci.07-10-03215.1987

PubMed Abstract | CrossRef Full Text | Google Scholar

Methqal, I., Provost, J.-S., Wilson, M. A., Monchi, O., Amiri, M., Pinsard, B., et al. (2017). Age-related shift in neuro-activation during a word-matching task. Front. Aging Neurosci. 9:265.

Google Scholar

Murray, M. M., Thelen, A., Thut, G., Romei, V., Martuzzi, R., and Matusz, P. J. (2016). The multisensory function of the human primary visual cortex. Neuropsychologia 83, 161–169. doi: 10.1016/j.neuropsychologia.2015.08.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Michail, G., Senkowski, D., Niedeggen, M., and Keil, J. (2021). Memory load alters perception-related neural oscillations during multisensory integration. J. Neurosci. 41, 1505–1515. doi: 10.1523/jneurosci.1397-20.2020

PubMed Abstract | CrossRef Full Text | Google Scholar

Miller, J. (1982). Divided attention: evidence for coactivation with redundant signals. Cogn. Psychol. 14, 247–279. doi: 10.1016/0010-0285(82)90010-x

CrossRef Full Text | Google Scholar

Miller, J. (1986). Timecourse of coactivation in bimodal divided attention. Percept. Psychophys. 40, 331–343. doi: 10.3758/bf03203025

PubMed Abstract | CrossRef Full Text | Google Scholar

O’Brien, J. M., Chan, J. S., and Setti, A. (2020). Audio-visual training in older adults: 2-interval-forced choice task improves performance. Front. Neurosci. 14:569212.

Google Scholar

Paige, L. E., and Gutchess, A. H. (2015). “Cognitive neuroscience of aging,” in Encyclopedia of Geropsychology, ed. N. A. Pachana (Berlin: Springer).

Google Scholar

Peiffer, A. M., Mozolic, J. L., Hugenschmidt, C. E., and Laurienti, P. J. (2007). Age-related multisensory enhancement in a simple audiovisual detection task. Neuroreport 18, 1077–1081. doi: 10.1097/wnr.0b013e3281e72ae7

PubMed Abstract | CrossRef Full Text | Google Scholar

Raab, D. H. (1962). Statistical facilitation of simple reaction times. Trans. N. Y. Acad. Sci. 24, 574–590. doi: 10.1111/j.2164-0947.1962.tb01433.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Ren, Y., Li, S., Wang, T., and Yang, W. (2020a). Age-related shifts in theta oscillatory activity during audio-visual integration regardless of visual attentional load. Front. Aging Neurosci. 12:571950.

Google Scholar

Ren, Y., Ren, Y., Yang, W., Tang, X., Wu, F., Wu, Q., et al. (2018). Comparison for younger and older adults: stimulus temporal asynchrony modulates audiovisual integration. Int. J. Psychophysiol. 124, 1–11. doi: 10.1016/j.ijpsycho.2017.12.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Ren, Y., Xu, Z., Lu, S., Wang, T., and Yang, W. (2020b). Stimulus specific to age-related audio-visual integration in discrimination tasks. i-Perception 11, 1–14.

Google Scholar

Ren, Y., Xu, Z., Wang, T., and Yang, W. (2020c). Age-related alterations in audiovisual integration: a brief overview. Psychologia 62, 233–252. doi: 10.2117/psysoc.2020-a002

CrossRef Full Text | Google Scholar

Ren, Y., Yang, W., Nakahashi, K., Takahashi, S., and Wu, J. (2016). Audiovisual integration delayed by stimulus onset asynchrony between auditory and visual stimuli in older adults. Perception 46, 205–218. doi: 10.1177/0301006616673850

PubMed Abstract | CrossRef Full Text | Google Scholar

Salthouse, T. A. (1996). The processing-speed theory of adult age differences in cognition. Psychol. Rev. 103, 403–428. doi: 10.1037/0033-295x.103.3.403

PubMed Abstract | CrossRef Full Text | Google Scholar

Sekiyama, K., Takahiro, S., and Shinichi, S. (2014). Enhanced audiovisual integration with aging in speech perception: a heightened McGurk effect in older adults. Front. Psychol. 5:323.

Google Scholar

Setti, A., Stapleton, J., Leahy, D., Walsh, C., Kenny, R. A., and Newell, F. N. (2014). Improving the efficiency of multisensory integration in older adults: audio-visual temporal discrimination training reduces susceptibility to the sound-induced flash illusion. Neuropsychologia 61, 259–268. doi: 10.1016/j.neuropsychologia.2014.06.027

PubMed Abstract | CrossRef Full Text | Google Scholar

Stein, B. E. (2012). The New Handbook of Multisensory Processing. Cambridge: MIT Press.

Google Scholar

Stein, B. E., and Meredith, M. A. (1993). The Merging of the Senses. Cambridge: MIT Press.

Google Scholar

Stephen, J. M., Knoefel, J. E., Adair, J., Hart, B., and Aine, C. J. (2010). Aging-related changes in auditory and visual integration measured with MEG. Neurosci. Lett. 484, 76–80. doi: 10.1016/j.neulet.2010.08.023

PubMed Abstract | CrossRef Full Text | Google Scholar

Stevenson, R. A., Ghose, D., Fister, J. K., Sarko, D. K., Altieri, N. A., Nidiffer, A. R., et al. (2014). Identifying and quantifying multisensory integration: a tutorial review. Brain Topogr. 27, 707–730. doi: 10.1007/s10548-014-0365-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Wahn, B., and König, P. (2015). Audition and vision share spatial attentional resources, yet attentional load does not disrupt audiovisual integration. Front. Psychol. 6:1084.

Google Scholar

Xu, Z., Yang, W., Zhou, Z., and Ren, Y. (2020). Cue–target onset asynchrony modulates interaction between exogenous attention and audiovisual integration. Cogn. Process. 21, 261–270. doi: 10.1007/s10339-020-00950-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, W., Guo, A., Li, Y., Qiu, J., Li, S., Yin, S., et al. (2018). Audio-visual spatiotemporal perceptual training enhances the P300 component in healthy older adults. Front. Psychol. 9:2537.

Google Scholar

Keywords: audiovisual integration, perceptual load, older adults, race model, discrimination task

Citation: Ren Y, Li H, Li Y, Wang T and Yang W (2021) Visual Perceptual Load Attenuates Age-Related Audiovisual Integration in an Audiovisual Discrimination Task. Front. Psychol. 12:740221. doi: 10.3389/fpsyg.2021.740221

Received: 13 July 2021; Accepted: 23 August 2021;
Published: 29 September 2021.

Edited by:

Julian Keil, University of Kiel, Germany

Reviewed by:

Rebecca Hirst, Trinity College Dublin, Ireland
Jonathan Wilbiks, University of New Brunswick Fredericton, Canada

Copyright © 2021 Ren, Li, Li, Wang and Yang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Weiping Yang, swywp@163.com

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.