Skip to main content

ORIGINAL RESEARCH article

Front. Hum. Neurosci.

Sec. Speech and Language

Volume 19 - 2025 | doi: 10.3389/fnhum.2025.1560558

Neural Speech Tracking in a Virtual Acoustic Environment: Audio-Visual Benefit for Unscripted Continuous Speech

Provisionally accepted
  • 1 Division of Neuropsychology, Department of Psychology, Faculty of Medicine and Health Sciences, University of Oldenburg, Oldenburg, Germany
  • 2 Department of Medical Physics and Acoustics, Faculty of Medicine and Health Sciences, University of Oldenburg, Oldenburg, Lower Saxony, Germany

The final, formatted version of the article will be published soon.

    The audio-visual benefit in speech perception-where congruent visual input enhances auditory processing-is well-documented across age groups, particularly in challenging listening conditions and among individuals with varying hearing abilities. However, most studies rely on highly controlled laboratory environments with scripted stimuli. Here, we examine the audio-visual benefit using unscripted, natural speech from untrained speakers within a virtual acoustic environment. Using electroencephalography (EEG) and cortical speech tracking, we assessed neural responses across audio-visual, audio-only, visual-only, and masked-lip conditions to isolate the role of lip movements. Additionally, we analysed individual differences in acoustic and visual features of the speakers, including pitch, jitter, and lip-openness, to explore their influence on the audio-visual speech tracking benefit. Results showed a significant audio-visual enhancement in speech tracking with background noise, with the masked-lip condition performing similarly to the audio-only condition, emphasizing the importance of lip movements in adverse listening situations. Our findings reveal the feasibility of cortical speech tracking with naturalistic stimuli and underscore the impact of individual speaker characteristics on audio-visual integration in real-world listening contexts.Auditory attention decoding (AAD) has traditionally aimed to distinguish between target and non-target speakers in environments with competing voices, capturing selective attention mechanisms in complex auditory scenes. Significant strides have been made in this field by decoding the speaker to whom a listener is attending, based on the brain's response to multiple simultaneous speakers (Ding & Simon, 2014;Luo & Poeppel, 2007;Mirkovic et al., 2015). AAD studies typically rely on controlled, multi-speaker environments, often using professional speakers and scripted speech for consistency and precision (Holtze et al., 2023;Jaeger et al., 2020;Mirkovic et al., 2016). Although these studies have been foundational, their reliance on controlled settings presents challenges for generalizing findings to more naturalistic auditory environments.

    Keywords: Neural speech tracking, EEG, Virtual acoustic environment, continuous speech, Unscripted conversation

    Received: 14 Jan 2025; Accepted: 27 Mar 2025.

    Copyright: © 2025 Daeglau, Otten, Grimm, Mirkovic, Hohmann and Debener. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

    * Correspondence: Mareike Daeglau, Division of Neuropsychology, Department of Psychology, Faculty of Medicine and Health Sciences, University of Oldenburg, Oldenburg, Germany

    Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

    Research integrity at Frontiers

    Man ultramarathon runner in the mountains he trains at sunset

    94% of researchers rate our articles as excellent or good

    Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


    Find out more