Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 06 November 2024
Sec. Auditory Cognitive Neuroscience

Novel bias-reduced coherence measure for EEG-based speech tracking in listeners with hearing impairment

  • 1Centre for Mathematical Sciences, Lund University, Lund, Sweden
  • 2Eriksholm Research Centre, Oticon A/S, Snekkersten, Denmark
  • 3Department of Electrical Engineering, Linköping University, Linköping, Sweden

In the literature, auditory attention is explored through neural speech tracking, primarily entailing modeling and analyzing electroencephalography (EEG) responses to natural speech via linear filtering. Our study takes a novel approach, introducing an enhanced coherence estimation technique to assess the strength of neural speech tracking. This enables effective discrimination between attended and ignored speech. To mitigate the impact of colored noise in EEG, we address two biases–overall coherence-level bias and spectral peak-shifting bias. In a listening study involving 32 participants with hearing impairment, tasked with attending to competing talkers in background noise, our coherence-based method effectively discerns EEG representations of attended and ignored speech. We comprehensively analyze frequency bands, individual frequencies, and EEG channels. Frequency bands of importance are shown to be delta, theta and alpha, and the important EEG channels are the central. Lastly, we showcase coherence differences across different noise reduction settings implemented in hearing aids (HAs), underscoring our method's potential to objectively assess auditory attention and enhance HA efficacy.

1 Introduction

Listening is a biological process that engages the entire auditory system. The process, from sound pressures to neural activations, includes (non-linear) transforms of the peripheral auditory system as well as complex processing within the central auditory system. The listening process also affects the electrical activity of the brain, which can be measured on scalp-level using electroencephalography (EEG). Linear filter estimation, referred to as temporal response functions (TRFs) (Crosse et al., 2016; Alickovic et al., 2019; Geirnaert et al., 2021), has been shown to capture the relations between the auditory system and EEG signatures (O'sullivan et al., 2015). State-of-the art methods employ TRFs to analyze phonemes (Di Liberto et al., 2015; Brodbeck et al., 2018; Carta et al., 2023) and semantic content (Broderick et al., 2018; Gillis et al., 2021). The TRF methods are also particularly valuable for decoding auditory attention, specifically in identifying the attended talker from EEG signals in challenging listening situations like cocktail-party environments (Cherry, 1953). Effectively decoding auditory attention to distinguish and enhance attended speech in multi-talker situations holds significance for hearing aid (HA) applications (Lunner et al., 2020; Alickovic et al., 2020, 2021; Andersen et al., 2021).

The mentioned studies, along with others, provide evidence that neural speech processing exhibits a linear component. Specifically, there are robust linear relationships between the speech envelope and the concurrent EEG signatures. Hence, spectral coherence analysis between the speech envelope and EEG signals can confidently identify and analyze the underlying system to a significant degree, as demonstrated in Viswanathan et al. (2019) and Vander Ghinst et al. (2021). Spectral coherence is also a measure of the linear coupling between two signals, calculated as the cross-spectrum normalized by the respective auto-spectrum of both signals.

Various methods exist for estimating spectral coherence, where the Welch method is the most commonly applied. In this approach, uncorrelated Fourier-based spectrum estimates from multiple data segments (that may overlap) are averaged (Welch, 1967). Another state-of-the-art approach is Thomson's multitaper method, which multiplies data with various tapering windows before Fourier analysis and subsequent averaging (Thomson, 1982; Walden, 2000; Karnik et al., 2022). Multitapers are often applied in various applications for robust spectral estimation of EEG (Alickovic et al., 2023; Reinhold and Sandsten, 2022; Viswanathan et al., 2019; Babadi and Brown, 2014; Hansson-Sandsten, 2010).

The statistics in coherence estimation using Welch method (Carter and Nuttall, 1972; Carter et al., 1973) as well as the Thomson's multitaper method (Thomson, 1982; Bronez, 1992; McCoy et al., 1998; Brynolfsson and Hansson-Sandsten, 2014; Hansson-Sandsten, 2011) has been thoroughly studied. The coherence of 1 signifies a perfect linear connection between signals, while no coupling yields the ideal coherence is 0. In practice, the lack of infinite data always introduces a significant bias upwards toward positive values in the no coupling case (Carter and Nuttall, 1972). In comparison, the variance can be assumed to be small when a reasonable amount of data is available (Carter et al., 1973; Thomson, 1982).

This paper introduces a new method for EEG-speech envelope coherence estimation, denoted bias-reduced coherence (BRC). The method aims to decrease the bias of estimation, while assessing the effectiveness of coherence in analyzing the relationship between speech envelope and EEG responses. With a certain choice in the cross-spectra estimation, as part of the coherence measure, the bias at low linear coupling can be reduced compared to traditional methods. Considering that EEG responses are susceptible to random noise, any detected linear coupling is expected to be weak. Real data is employed to evaluate the coherence methods and showcase their potential in decoding auditory attention. Furthermore, since coherence is applied to signals influenced by 1/f spectrally shaped noise prevalent in EEG data (Bénar et al., 2019), this study also investigates biases arising from this phenomenon. Coherence “levels out” the slanted noise in the coherence spectrum, and peaks that are widened by the taper kernel will have a bias toward higher frequencies. This paper quantitatively analyzes this bias and its implications for EEG and hearing application. The study's hypotheses are as follows.

Two types of anticipated biases in speech-EEG coherence estimation

1. Due to the high level of noise present in EEG data, we anticipate an overall upward bias across all frequencies from all methods. We propose that the new BRC can mitigate this bias, due to including taper phase information.

2. We expect coherence peaks from all coherence estimates to shift toward higher frequencies, due to the slanted nature of EEG noise. The extent and nature of this shift is hypothesized to be dependent on factors such as the number of tapers, data length, and the shape and power of the noise present in signals.

Auditory attention and hearing aid effects manifested in speech-EEG coherence changes

1. Differences in speech-EEG coherence between attended and ignored speech, specifically within the delta, theta, and alpha frequency bands associated with auditory processing, are anticipated when applying these methods to a population using HAs. For these bands, we expect clearer distinctions in coherence estimates with the BRC, since this improves bias and variance of coherence estimates.

2. Speech-EEG coherence differences are expected to be larger when HA noise reduction feature is activated compared to when it is deactivated.

This work extends our previous study (Keding et al., 2023), which introduces the novel BRC method that utilizes more of the phase information from tapered data segments. In Keding et al. (2023) the case when coherent signals are affected heavily by noise (causing low expected true coherence levels) is analyzed. Also, the base-level biases are compared for the traditional and the new BRC method, showing lower bias for BRC. Here, we present further analytical analysis of aspects evident during application of coherence as a measure within EEG analysis, focusing on neural speech tracking. Further comparisons of coherence methods topographically over channels are provided. Additionally, we investigate coherence as a metric for objectively assessing the effect of HAs on auditory attention during listening tasks.

The paper is structured as follows. Firstly, the traditional coherence method and the BRC are presented in Section 2, where the reduction in upward magnitude bias of BRC is explained. Analysis of the positional bias of peaks in the coherence spectrum due to 1/f noise is evaluated analytically as well as in simulations. Section 3 outlines the experimental design, pre-processing of data and statistical methods used in real data analysis. Results on real data are presented in Section 4, with related discussion in Section 5, highlighting the capabilities and limitations of coherence methods in auditory attention decoding and in objective evaluation of HA benefits. Conclusions are found in Section 6.

2 Coherence estimation

Spectral coherence Cxy(f) is the measure of the magnitude of a linear coupling between two signals in a system, ranging 0 ≤ Cxy(f) ≤ 1, here defined by its squared form

Cxy(f)=Sxy(f)2Sxx(f)Syy(f)    (1)

Sxy(f) is the cross-amplitude spectrum between signals x(n) and y(n). Sxx(f) and Syy(f) are the respective auto-spectra. Coherence, as denoted in Equation 1, is referred to as Magnitude Squared Coherence (MSC).

2.1 Magnitude squared coherence

Estimating the MSC entails estimating the different spectra according to

C^xy(f)=S^xy(f)2S^xx(f)S^yy(f)    (2)

The auto-spectra are estimated through averaging sub-spectra for each pair of L number of data segments and K number of data tapers as

S^xx(f)=1KLk=1Kl=1LXk,l(f)Xk,l(f)*    (2)
S^yy(f)=1KLk=1Kl=1LYk,l(f)Yk,l(f)*    (4)

where Fourier transforms of the l:th data segments xl(n), yl(n) over time samples n and total lengths N, and k:th tapering window hk(t) of respective signal are

Xk,l(f)=n=0N-1xl(n)hk(n)e-i2πfn    (5)
 Yk,l(f)=n=0N-1yl(n)hk(n)e-i2πfn    (6)

where i is the imaginary unit.

There are multiple options when constructing the cross-spectra from sub-spectra. Here we will consider two such options. The first has been used in Viswanathan et al. (2019) in speech envelope to EEG coherence estimation

S^xyTRAD(f)=1KLk=1K|l=1LXk,l(f)Yk,l(f)*|    (7)

Coherence estimation using this method is denoted the traditional method. In Equation 7 and from now on in the paper, the L data segments are assumed non-overlapping. A second option was introduced in our previous work (Keding et al., 2023), where both sums are taken before the absolute value as

S^xyBRC(f)=1KL|k=1Kl=1LXk,l(f)Yk,l(f)*|    (8)

Note that these two options gives the same auto-spectra, since all sub-spectra are real and positive. The resulting estimator of coherence with S^xyBRC(f) is shown to decrease the bias upwards in magnitude for low coherence scenarios. This due to the traditional method averaging power over tapers without taking the phase information of tapers into account, which BRC does. This is shown in Keding et al. (2023). Coherence estimation using this method is therefore denoted bias-reduced coherence (BRC). The method in Equation 8 utilizes the phase information of the signals in estimation of the final cross-spectrum. A third option where the order of sums in Equation 7 is reversed, so absolute values of the sum of tapers are taken, is not considered here as L>K in most applications.

A common choice, in accordance with the Thomson's multitaper method, for the set of tapering windows {hk(t)}1, ..., K are the Discrete Prolate Spheroidal Sequences (DPSS), which can be seen in time and frequency domain in Figure 1 (Slepian, 1978). These are the windows that maximize the main lobe (or within-band) power relative to broadband power for a normalized frequency bandwidth W, but have no closed-form expression. They have been shown to have effective variance-reduction characteristics. Although it is argued here that these are suitable choices for tapering windows, the tapering windows can be chosen from another set, with the general methodology staying identical.

Figure 1
www.frontiersin.org

Figure 1. Spectral estimation windows. The temporal representation and power spectra of each of the 10 first DPSS windows used in the multitaper coherence estimation. The tapers are enumerated from bottom to top. The sum of the taper kernels is shown above the separate spectral kernels, effectively showing the width of the narrowband leakage effects.

The number of DPSS windows theoretically can be infinite, but using a large amount of tapering windows is not advisable due to its potential to significantly increase the narrowband leakage. This leads to the smearing of the spectrum, making spectral peaks of neighboring frequencies to be indistinguishable from one another. This phenomenon become particularly problematic when estimating coherence in 1/f-shaped noise, as discussed in Section 2.2. Instead, a practical approach involves selecting the desired bandwidth of the tapering windows according to the specific application and limit the number of windows accordingly. The DPSS windows are defined in terms of both their shape and the number of windows required to achieve this bandwidth. A common choice for the number of windows is K = 2NW, where N is the lengths of signals and W is the normalized frequency bandwidth.

A overall bias analysis of the traditional coherence estimation technique, compared to the BRC has been given in Keding et al. (2023). In this work, the BRC was shown to significantly reduce the bias of the coherence estimate level when there is no phase-locking between signals in channel x and y. The traditional method wrongly shows a peak in the coherence spectrum even though there is no phase-locking present. If 1/f-shaped noise is added to the phase-locked signal in either or both channels, the peaks of coherence for narrowband signals are also spectrally shifted toward higher frequencies. This simple observation made in Keding et al. (2023) is further expanded and analyzed in the following section.

2.2 Spectral positional bias in EEG-like noise

In various EEG-related applications, coherence measures are employed to analyze coherence spectra between EEG channels, or between EEG channels and sensory stimuli. This analysis allows the detection of key frequencies showing notable correlation. The identification of these specific correlation frequencies finds applications in tasks like filtering, denoising, and drawing functional insights about the brain.

However, a common challenge faced in visually inspecting the coherence spectra, or any coherence method based on the Fourier transform, is to identify key frequencies in the correlation between channels in the presence of unwanted spectral shifting of relevant coherence energy peaks. This shift is attributed to the presence of slanted 1/f noise, indicative of irrelevant brain activity. As the coherence normalizes the base levels of spectrum, the 1/f pattern becomes flattened, making it difficult to perceive its impact during peak analysis. This effect is particularly pronounced at lower frequencies, where the 1/f noise spectrum exhibits the most prominent degree of gradient.

2.2.1 Approximate coherence for sinusoidal signal

To illustrate the positional bias of peaks in the coherence spectrum when one or more signals are disturbed by 1/f noise, we introduce an expression for the coherence of a single frequency coupled model. Although the positional bias impacts any peaked spectra similarly, a simple case showing the bias involves single frequency oscillations is:

x(n)=s(n),  y(n)=s(n)+σpep(n)    (9)

where the signal s(n)=Ffn[δf0(f)] represents a complex sinusoidal and the noise has a spectral distribution E[|Fnf[ep(n)]|2]=1/fα, α>0.

In this subsection, it is assumed that Htot(f)=k=1K|Hk(f)|2=rect(f/B), the combined taper kernel, is a box function of energy one and frequency range of B = 2Wfs, as shown in Thomson (1982), where fs denotes the sampling frequency. The expected value of MSC between x and y can be approximated by the ratio of smeared cross-spectra and smeared auto-spectra of signals, since these are the expectations of the cross/auto-spectra. This gives a fairly simple expression for the final expectation of the coherence estimate. For f>B/2, an initial application of a zero:th order Taylor expansion of the expectation yields

E[C^xy(f)]|E[S^xy(f)]|2E[S^xx(f)]E[S^yy(f)]    (10)
=|E[Sxy(f)]*Htot(f)|2(E[Sxx(f)]*Htot(f))(E[Syy(f)]*Htot(f))    (11)
=|δf0(f)*Htot(f)|2(δf0(f)*Htot(f))((δf0(f)+σp2fα)*Htot(f))    (12)
=Htot(f-f0)Htot(f-f0)+σp2fα*Htot(f)    (13)
={11+σp2fα*Htot(f)f[f0B2,f0+B2]0 otherwise    (14)

where

1fα*Htot(f)=B/2B/2(νf)αdν=    (15)
={1α+1((f+B/2)α+1(fB/2)α+1)α<1log(f+B/2)log(fB/2)α=11α+1((f+B/2)α+1(fB/2)α+1)α>1    (16)

As can be easily observed in Equation 14, the actual bias beyond the main lobe of the taper kernel is disregarded. In reality, these spectra have some estimated coherence, although lower than the main lobe coherence shown, as discussed in Keding et al. (2023). One can identify this expression as an increasing coherence spectrum within the bandwidth of the tapers used. Perhaps the most frequent choice is α = 1, which gives 1/f noise. A bias error in Equation 14 is expected, since the approximation is crude in reality. Additionally, an overall bias upwards is missing, that is observed in the no-coherence case, as estimation of zero-coherence is impossible with finite data. However, this does not affect the frequency shifting of peaks.

Figure 2 presents a comparison between the empirical expectation of Nreal = 10, 000 simulations of signals according to Equation 9 to the Equation 14. Signal parameters were set to f0 = 10, N = 128, sampling frequency fs = 128, σ = [0.5, 1, 5] and α = 1, yielding a taper bandwidth B = 6 Hz. The simulations demonstrate the positional bias of the peak clearly in all noise levels of the slanted noise. One can see that the expression follows the expectation of the coherence estimates very well, up to a scale factor.

Figure 2
www.frontiersin.org

Figure 2. Evaluation of expectation expression. The expression for coherence expectation in dashed lines is shown along the corresponding average coherence of 10,000 estimates on a sinusoidal signal in two channels, with added 1/f noise in one of the channels. The comparison is made for three different noise levels, σ = [0.5, 1, 5], which corresponds the pairs of expectations in ascending order. One can see that the expression estimates the behavior of the coherence expectation very well, up to a constant. Ideal coherence lies in a single peak at f0 = 10 Hz.

2.2.2 Effects of signal and method parameters

Characterizing parameter effects on the frequency shifts of peaks in coherence is possible with some a priori knowledge about signals and the coherence estimation method. These shifts depend on the data segment length and the number of tapers used in a multitaper coherence estimation method.

Firstly, the data length N modulates the taper kernel's narrowband width. Decreasing N will linearly increase frequency length of the bias of coherence peaks. Secondly, introducing more tapers in a multitaper scenario will further amplify this bias, although there is a trade-off involving variance reduction with higher K values. Simulations are made to visualize this effect. In these simulations, signals are defined as per Equation 9, with f0 = 10 Hz. Colored noise ep(n) is made by Fourier transforming white noise and then multiplying it by the 1/f-spectrum before transforming back. Coherence estimates were computed using BRC on Nreal = 10, 000 generated signals with parameters N = 128, fs = 128, L = 50 and varying K from 1 to 10. Figures 3A, B show the empirical expectation and variance of the coherence estimates across all repetitions. These figures reveal that the coherence peak shifts more in the spectral domain as the number of included window increases. It is worth noting that method parameters, such as data lengths N and number of windows K, are typically chosen in an application and can be adjusted accordingly during analysis.

Figure 3
www.frontiersin.org

Figure 3. Expectation (A) and variance (B) of coherence estimation using BRC with a varying amount of tapers included. The number of tapers K ranges from 1 to 10.

On the other hand, parameters that influence the shifting bias are not as easily quantified prior to analysis. Examples of such parameters include the noise-level of the 1/f noise and the spectral position of the coherence peak in question. Referring back to Equation 14, the SNR of 1/f noise doesn't affect how much the peaks are shifted spectrally; it only impacts the magnitude of the resulting coherence estimate. An exception occurs with very high SNR, where the coherence approaches one, making the slanted peak indistinguishable from a plateau. In this case, the risk of misinterpretation is minimal. Nevertheless, in many applications, coherence values are not substantially large.

It is notable that the true frequency of the peak does not alter how wide peaks are shifted, but that it does influence the slope of the widened band, resulting in a sharper or more blunt peak. This occurs due to the flattening of slope of the 1/f noise spectrum at higher frequencies. Additionally, there is often noise in the x channel. Parameter effects on the frequency shifting of peaks are consistent when 1/f noise (or white) is added to y, albeit with an overall lower coherence.

3 Experimental setup and statistical methods

The methods discussed in this study were assessed using an EEG dataset collected from individuals with hearing impairment. This dataset has been previously analyzed in several studies using different analysis approaches (Andersen et al., 2021; Keding et al., 2023). Data collection started 1st of June 2020 and ended on 7th of September 2020. Experimental protocol has been approved by the Science Ethics Committee for the Capital Region of Denmark (journal no. H20028542). The study was conducted according to the Declaration of Helsinki, and all the participants signed a written consent prior to the experiment. More information and details about data can be found in Andersen et al. (2021).

3.1 Participants and experimental design

The study included 31 experienced HA users, aged 65.5 years on average, with mild to moderately severe sensorineural hearing loss. All participants were native Danish speakers without a history of neurological disorders, dyslexia or diabetes mellitus. The experiment consisted of 84 trials, with the first four trials designated for training and the remaining 80 for testing and analysis. Each trial commenced with a short silence, succeeded by 5 s of background noise. Subsequently, 33 s of concurrent attended and ignored speech, along with background noise, were presented. Post-trial, participants responded to a two-choice question about the content of the attended speech (Alickovic et al., 2020, 2021; Andersen et al., 2021).

Speech stimuli were presented through a sound card (RME Hammerfall DSP multiface II, Audio AG, Germany) to six loudspeakers (Genelec 8040A; Genelec Oy, Finland). The loudspeakers were positioned at 30° (T1-T2), 112.5° (B1-B2), and 157.5° (B3-B4) azimuth. The audio was played at a 44.1 kHz sampling rate. Attended and ignored speech were presented at 73 dB SPL in T1 or T2, randomized evenly over trials. Background noise was presented at 70 dB SPL in B1-B4, with a mix of four talkers in each loudspeaker. T1 and T2 had talkers of opposite sexes presented at each trial. EEG data were recorded at 1024 Hz using a BioSemi ActiveTwo 64-channel EEG system in a 10–20 layout.

Participants wore HAs during all trials, split into four sessions of 20 trials each. These sessions were randomly ordered and had different HA noise reduction (NR) settings. Two sessions had NR turned OFF (referred to as HA NR OFF), and two sessions had NR turned ON (referred to as HA NR ON) with two different NR algorithms, as described in Andersen et al. (2021). However, these NR algorithms are not discussed in this study.

3.2 Preprocessing steps

The attended and ignored speech signals in each trial were decimated to a sampling rate of 256 Hz, and envelopes were calculated as the absolute value of the analytic version of speech signals, following similar studies (O'sullivan et al., 2015; Alickovic et al., 2019). The EEG data were re-referenced to the average of mastoid channels, bandpass filtered between 0.5 and 70 Hz, notch filtered to remove line noise, and resampled to 256 Hz. On average, 0.87 channels per subject were identified as bad channels and were interpolated using neighboring channels. Independent component analysis (Bell and Sejnowski, 1995) was performed to remove components not related to brain activity. On average, 17 components per subject were removed. The analysis involved two speech envelope signals and 64 EEG channels of EEG for each trial.

3.3 Statistical analysis

Statistical p-value testing assessed the significance of method choices (traditional or BRC estimation method) on differences of speech-EEG coherence between attended and ignored speech. Speech-EEG coherence estimates were computed for each trial using both the traditional method and the BRC. This was done using L = 33 number of data segments that are T = 1 s long, and K = 10 data tapers. Coherence values were then averaged across each EEG band (denoted B), or retained as estimates for each frequency bin. Bands are defined as delta (1–4 Hz), theta (4–8 Hz), alpha (8–12 Hz), beta (12–30 Hz), and gamma (30–128 Hz). A null hypothesis was formulated, positing that the grand average (over all subjects and trials) of attended speech coherence CattB and ignored speech coherence CignB are equal. The alternative hypothesis suggests that CattB is greater than CignB.

H0:CattB-CignB0
H1:CattB-CignB>0

To estimate the null distribution, DB=CattB-CignB, by bootstrap sampling, the sign of DB was randomized for each subject trial. Next, the mean over all trials and subjects was taken to make one grand average sample from the null hypothesis (Viswanathan et al., 2019; Te and Ap, 2002). This procedure is repeated 500,000 times to approximate the underlying null distribution. Finally, the probability of actual observed coherence grand average difference was derived from the null distribution.

4 Results

This section applies the BRC method and the traditional method of coherence estimation to the experimental data detailed in Sections 3.1 and 3.2. Speech-EEG coherence is shown to be useful in decoding attention in a multi-talker scenario. This is done by comparing the resulting p-values of the difference of coherence when x is either envelopes of attended or ignored speech. As we are interested in not simply the significance of results, but the overall sensitivity of the novel method BRC compared to the traditional method, comparing p-values illuminates in this aspect. Smaller values of p-values indicate found larger differences between attended speech and ignored speech, which indicate a more sensitive model in this context.

To start, Figures 4A, B show the grand average of coherence estimates using the traditional method and BRC, respectively. These are shown beside the estimated power spectra of speech envelopes, averaged over all speech material, in Figure 4C. A general correspondence is found between the shapes of the different spectra, which is expected since the spectra are all based on the Fourier transform of speech envelopes. Both methods successfully differentiate attended and ignored speech, which is expected due to the success of linear filters in discriminating speakers in multi-talker situations for hearing impaired listeners. Notably, BRC provides less noisy coherence estimates across frequencies and exhibits much lower baseline coherence. The lower baseline coherence is a result of the reduced bias for low coherence estimates when using BRC, described in Keding et al. (2023). While both methods show a similar frequency pattern, the traditional method shows two clear coherence peaks, whereas BRC features one more pronounced peak at 3–5 Hz. The BRC captures a smaller peak in the 6–9 Hz range for attended speech-EEG coherence, highlighting a difference between speech streams. In contrast, the traditional method lacks this clear pattern as it results in peaks for both attended and ignored speech.

Figure 4
www.frontiersin.org

Figure 4. Grand averages of speech-EEG coherence between attended (green) and ignored (red) speech envelopes estimated using (A) the traditional method and (B) the BRC method. Colored areas show the empirical 95% confidence interval of estimates across the spectral range. Note the rather large difference in overall absolute value of estimates between the two methods on the y-axis. The average power spectra of all speech envelopes (both attended and ignored) is shown in (C). One can observe a general correspondence in the shape of the spectra, comparing (A, B) to (C).

The difference in grand averages seen in Figure 4 becomes more apparent through grand-level mean testing of the coherence. As detailed in Section 3.3, this approach provides three ways for assessing the significance of speech-EEG coherence differences between attended and ignored speech: at the band level, frequency bin level, and channel level. Figure 5A displays p-values, which evaluate speech-EEG difference between attended and ignored speech, for each frequency bin. The figure is cut at 40 Hz, but the pattern of non-significant p-values seen in the range of 10–40 Hz continues upwards in the frequency range above 40 Hz, omitted to enlarge the differences in p-values seen at lower frequencies. Figure 5B shows the significance of coherence differences when coherence estimates are averaged within each EEG band.

Figure 5
www.frontiersin.org

Figure 5. P-values, indicating the difference between observed speech-EEG coherence estimates for attended and ignored speech envelopes, are shown for (A) each frequency bin and (B) each EEG band. Traditional method-derived values are indicated in red, while the values in blue are computed using the BRC method. BRC consistently yields lower p-values, facilitating stronger statistical testing with reduced data requirements. With a few frequency bins as an exception, this is true for the frequencies and bands where p-values for any method are significant. The 95% significance level of p-values is marked by the horizontal line, with the < symbol indicating p-values below the cutoff.

Speech-EEG coherence estimates for the delta, theta and alpha bands across each EEG channel were extracted and the grand averages of these were tested, resulting in topographic plots of logarithms of corresponding p-values, shown in Figure 6. Importantly, enhanced speech-EEG coherence for attended speech envelope compared to ignored speech envelope is more significant when p-values and logarithm of p-values are lower, as depicted by blue shading in the figure.

Figure 6
www.frontiersin.org

Figure 6. Topographic plots of the logarithm of p-values showing significant difference between attended and ignored coherence, (A) the delta band, (B) the theta band and (C) the alpha band. Upper figures show log p-values calculated with the traditional method, and lower figures are calculated with the BRC method. Color intensity of log p-values is matched over each column for comparison.

As a final piece of analysis, the effects of HA NR on the coherence measures of attended and ignored speech were investigated. Half of the trials involved participants with the NR algorithm OFF in their HAs, while the other half had it ON. After splitting the trials into conditions with HA NR OFF and HA NR ON and conducting identical statistical analysis of attended/ignored speech-EEG coherence differences (now with half the total number of trials for each condition), the resulting p-values are shown in Figures 7A, B. P-values are generally lower when HA NR is ON in frequencies showing a linear relationship between speech envelopes and EEG stronger than the same for when HA NR is OFF.

Figure 7
www.frontiersin.org

Figure 7. P-values, indicating the difference between observed speech-EEG coherence estimates for attended and ignored speech envelopes, are shown for (A) each frequency bin and (B) each EEG band. Traditional method-derived values are indicated in red, while the values in blue are computed using the BRC method. BRC consistently yields lower p-values, facilitating stronger statistical testing with reduced data requirements. The 95% significance level of p-values is marked by the horizontal line, with the < symbol indicating p-values below the cutoff. (A) Frequencies, HA NR OFF. (B) Frequencies, HA NR ON.

5 Discussion

The potential of the coherence measure for neural tracking of speech, compared to commonly used methods, is discussed in Section 5.1. In this section, discussion around the results of the BRC and traditional coherence method is also provided. Furthermore, Section 5.2 provides promising results that highlight speech-EEG coherence as a valuable objective measure for assessing the effects of HA signal processing algorithms on auditory attention.

5.1 Separating attended and ignored speech

As an initial discussion to motivate the use of coherence methods, we compare characteristics of the coherence methodology to the state-of-the-art TRF analysis paradigm. Commonly within the field of neural tracking of sound, a TRF filter is trained on attended speech, and then applied on other trials to distinguish between attended and ignored speech on these new trials. Given a certain decision window on unseen data, the previous fitting of TRF on other data will make the method better at classifying speech as attended or ignored, although this requires the fitting of the filter on previous data. The methodology of analyzing the difference in attended and ignored coherence, is algorithmically simpler. This approach requires no training or data splitting (train-test or cross-validation), and can be applied directly to new data to make conclusions about auditory attention. Additionally, in the fitting of filters, a method and level of regularization are required to be chosen. This problem, of choosing regularization hyperparameters, is avoided with the coherence methodology. Instead, hyperparameters are chosen during coherence estimation on new data. These hyperparameters can be deduced from the signal characteristics of speech features and EEG responses, to a very large extent. Therefore, coherence can be suitable for early screening of data where no training data is available, and data splitting is not possible.

Furthermore, as the coherence measure is calculated over each channel and frequency, analysis in these domains is straightforward. Comparatively, using TRF requires training and test of individual filters for each channel. Certain regularization techniques can also smear which frequencies are important in the filters. If classification between attended and ignored speech is of main interest, then coherence measures presented in this work can be used in further analysis to make classification models. In this work, the difference of coherence in frequencies and channels is used to differentiate between attended and ignored speech, although one can just as well use coherence as inputs to more complicated models.

With comments about the role of coherence in neural tracking of speech, interesting observations can be made when applying these methods. Initially, one can observe in Figure 4 that the coherence measures differ between are at the lower end of spectral axis, where also speech envelope power is strong (seen in Figure 4C). This is reasonable because it is the lower frequencies of the speech envelopes that carry information. High oscillations of envelopes are typically not informative for the speech envelope, as these not as prevalent.

For the statistical analysis of p-values in the frequency domain, shown in Figures 5A, B, two main observations can be made. Firstly, significant differences in coherence are observed in the delta, theta and alpha bands. Secondly, within the delta band, multiple individual bins exhibit significance with the new coherence estimation method BRC. This is particularly relevant, since the delta band should indeed exhibit a significant difference between attended and ignored speech envelopes, when estimated using linear filter methods, making BRC superior in regard to sensitivity, compared to the traditional approach (Bröhl and Kayser, 2021; McHaney et al., 2021). It is important to note that BRC and traditional coherence methods yield similar (not significant) p-values for frequencies and bands where a linear relationship between speech envelope and EEG is not anticipated. Also, considering the bias described in Section 2.2 and visualized in Figure 3, frequency bins showing coherence contains significant information from lower frequencies as the coherence peaks are shifted toward higher frequencies in slanted noise of EEG.

Focusing on the spatial patterns of coherence, Figures 6B, C strengthen the evidence for coherence differences in central and frontal regions, consistent with previous research (Fiedler et al., 2019; Lesenfants and Francart, 2020; Schmitt et al., 2022). p-values remain low across the parietal channels, in line with prior studies (Shahsavari Baboukani et al., 2022). Figures 6B, C also illustrate patterns where BRC results in smaller jumps in significance between neighboring channels, compared to the traditional one. The traditional method has more sensitivity over frontal and central channels in the theta band, and BRC has more sensitivity in the alpha band. Although it may be hard to make strong claims in regard to which alternative is preferable, each band include an equal amount of frequency bins. The significance gain in the alpha band using BRC is larger than the gain in the theta band using the traditional coherence method.

5.2 Coherence as an objective evaluation of HAs

This work suggests using coherence methods to objectively evaluate benefits of signal processing algorithms in HAs in real-life, multi-talker environments. As seen in Figures 7A, B, the p-values for the ON condition are lower than the p-values for the OFF condition. This suggests a more pronounced distinction between the coherence associated with attended speech compared to coherence associated with ignored speech. It implies that activating HA NR algorithm enhances the separation of attended speech from irrelevant sounds, thus aiding in the auditory attention task. This is corroborated by previous studies (Alickovic et al., 2020, 2021).

While beyond the scope of this study, a future research could explore the link between behavioral outcomes and coherence estimates in context of HA signal processing algorithms. Such an investigation could strengthen the proposed objective measure introduced for evaluating the benefits of HA technology. By delving into channels and frequencies of interest, it may demonstrate sensitivity of HAs with less EEG data. This could further facilitate clinical adoption of coherence measures as an objective assessment tool.

6 Conclusions

Linear system analysis has previously proven useful for inferring about the connection between speech features, in specific speech envelopes, and measured EEG representing cortical responses. This paper aligns with previous work, and show the functionality of using coherence to decode attention in a cocktail-party environment. Although using higher-order models is useful in their own right, the use of linear models to understand the brain is of importance since their effects on results can be easily quantified.

When interpreting brain electrical activity through spectral aspects of coherence, it is crucial to address two types of biases. Firstly, an upward bias toward higher coherence is analyzed due to responses having a very weak linear connection, showing an improvement with BRC. Moreover, this bias is mitigated by including more trials. Secondly, and of particular significance in EEG applications, there is the shifting of coherence peaks in spectral domain toward higher frequencies. This peak shifting is amplified when widening the effective bandwidth of the spectral estimation technique.

Finally, a novel coherence method, namely BRC, was introduced for objectively evaluating the efficacy of HA noise reduction algorithms in realistic listening environments. Observable differences in speech-EEG coherence between attended and ignored speech were observed when the noise reduction feature in HAs was activated compared to when it was deactivated. While no rigorous test is presented, the speech-EEG coherence difference between noise reduction schemes remains consistent over frequencies related to the listening task. Further work is needed to explore the suitability of coherence measures for objective evaluation of HAs, where parameters such as data amount are taken into account.

Data availability statement

The data analyzed in this study is subject to the following licenses/restrictions: due to this regulation and the way data was collected with low number of participants, it is not possible to fully anonymize the dataset and hence cannot be shared. Requests to access these datasets should be directed to Claus Nielsen, Y2xuaSYjeDAwMDQwO2VyaWtzaG9sbS5jb20=.

Ethics statement

The studies involving humans were approved by Etikprövningsmyndigheten, Sweden. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

OK: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing. EA: Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing. MAS: Conceptualization, Formal analysis, Funding acquisition, Investigation, Project administration, Validation, Writing – review & editing. MS: Conceptualization, Formal analysis, Funding acquisition, Methodology, Project administration, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. The research in this paper was supported by the ELLIIT strategic research environment, Sweden.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Alickovic, E., Lunner, T., Gustafsson, F., and Ljung, L. (2019). A tutorial on auditory attention identification methods. Front. Neurosci. 13:153. doi: 10.3389/fnins.2019.00153

PubMed Abstract | Crossref Full Text | Google Scholar

Alickovic, E., Lunner, T., Wendt, D., Fiedler, L., Hietkamp, R., Ng, E. H. N., et al. (2020). Neural representation enhanced for speech and reduced for background noise with a hearing aid noise reduction scheme during a selective attention task. Front. Neurosci. 14:846. doi: 10.3389/fnins.2020.00846

PubMed Abstract | Crossref Full Text | Google Scholar

Alickovic, E., Mendoza, C. F., Segar, A., Sandsten, M., and Skoglund, M. A. (2023). “Decoding auditory attention from EEG data using cepstral analysis,” in 2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW) (Rhodes Island: IEEE), 1–5.

Google Scholar

Alickovic, E., Ng, E. H. N., Fiedler, L., Santurette, S., Innes-Brown, H., and Graversen, C. (2021). Effects of hearing aid noise reduction on early and late cortical representations of competing talkers in noise. Front. Neurosci. 15:636060. doi: 10.3389/fnins.2021.636060

PubMed Abstract | Crossref Full Text | Google Scholar

Andersen, A. H., Santurette, S., Pedersen, M. S., Alickovic, E., Fiedler, L., Jensen, J., et al. (2021). “Creating clarity in noisy environments by using deep learning in hearing aids,” in Seminars in Hearing, Vol. 42 (New York, NY: Thieme Medical Publishers, Inc.), 260–281.

PubMed Abstract | Google Scholar

Babadi, B., and Brown, E. N. (2014). A review of multitaper spectral analysis. IEEE Transact. Biomed. Eng. 61, 1555–1564. doi: 10.1109/TBME.2014.2311996

PubMed Abstract | Crossref Full Text | Google Scholar

Bell, A. J., and Sejnowski, T. J. (1995). An information-maximization approach to blind separation and blind deconvolution. Neural Comput. 7, 1129–1159. doi: 10.1162/neco.1995.7.6.1129

PubMed Abstract | Crossref Full Text | Google Scholar

Bénar, C.-G., Grova, C., Jirsa, V., and Lina, J. (2019). Differences in MEG and EEG power-law scaling explained by a coupling between spatial coherence and frequency: a simulation study. J. Comput. Neurosci. 47, 31–41. doi: 10.1007/s10827-019-00721-9

PubMed Abstract | Crossref Full Text | Google Scholar

Brodbeck, C., Hong, L. E., and Simon, J. Z. (2018). Rapid transformation from auditory to linguistic representations of continuous speech. Curr. Biol. 28, 3976–3983. doi: 10.1016/j.cub.2018.10.042

PubMed Abstract | Crossref Full Text | Google Scholar

Broderick, M. P., Anderson, A. J., Di Liberto, G. M., Crosse, M. J., and Lalor, E. C. (2018). Electrophysiological correlates of semantic dissimilarity reflect the comprehension of natural, narrative speech. Curr. Biol. 28, 803–809. doi: 10.1016/j.cub.2018.01.080

PubMed Abstract | Crossref Full Text | Google Scholar

Bröhl, F., and Kayser, C. (2021). Delta/theta band EEG differentially tracks low and high frequency speech-derived envelopes. Neuroimage 233:117958. doi: 10.1016/j.neuroimage.2021.117958

PubMed Abstract | Crossref Full Text | Google Scholar

Bronez, T. P. (1992). On the performance advantage of multitaper spectral analysis. IEEE Transact. Signal Process. 40, 2941–2946. doi: 10.1109/78.175738

PubMed Abstract | Crossref Full Text | Google Scholar

Brynolfsson, J., and Hansson-Sandsten, M. (2014). “Multitaper estimation of the coherence spectrum in low SNR,” in Proceedings of the EUSIPCO (Lisbon).

Google Scholar

Carta, S., Alickovic, E., Zaar, J., Lopez Valdes, A., and Di Liberto, G. (2023). Cortical over-representation of phonetic onsets of ignored speech in hearing impaired individuals. bioRxiv [preprint]. doi: 10.1101/2023.06.26.546549

Crossref Full Text | Google Scholar

Carter, G. C., Knapp, C. H., and Nuttall, A. H. (1973). Statistics of estimate of magnitude-coherence function. IEEE Transact. Audio Electroacoust. AU21, 388–389. doi: 10.1109/TAU.1973.1162487

Crossref Full Text | Google Scholar

Carter, G. C., and Nuttall, A. H. (1972). Statistics of the estimate of coherence. Proc. IEEE 60, 465–466. doi: 10.1109/PROC.1972.8671

Crossref Full Text | Google Scholar

Cherry, E. C. (1953). Some experiments on the recognition of speech, with one and with two ears. J. Acoust. Soc. Am. 25, 975–979. doi: 10.1121/1.1907229

Crossref Full Text | Google Scholar

Crosse, M. J., Di Liberto, G. M., Bednar, A., and Lalor, E. C. (2016). The multivariate temporal response function (mTRF) toolbox: a MATLAB toolbox for relating neural signals to continuous stimuli. Front. Hum. Neurosci. 10:604. doi: 10.3389/fnhum.2016.00604

PubMed Abstract | Crossref Full Text | Google Scholar

Di Liberto, G. M., O'sullivan, J. A., and Lalor, E. C. (2015). Low-frequency cortical entrainment to speech reflects phoneme-level processing. Curr. Biol. 25, 2457–2465. doi: 10.1016/j.cub.2015.08.030

PubMed Abstract | Crossref Full Text | Google Scholar

Fiedler, L., Wöstmann, M., Herbst, S. K., and Obleser, J. (2019). Late cortical tracking of ignored speech facilitates neural selectivity in acoustically challenging conditions. Neuroimage 186, 33–42. doi: 10.1016/j.neuroimage.2018.10.057

PubMed Abstract | Crossref Full Text | Google Scholar

Geirnaert, S., Vandecappelle, S., Alickovic, E., de Cheveigne, A., Lalor, E., Meyer, B. T., et al. (2021). Electroencephalography-based auditory attention decoding: toward neurosteered hearing devices. IEEE Signal Process. Mag. 38, 89–102. doi: 10.1109/MSP.2021.3075932

Crossref Full Text | Google Scholar

Gillis, M., Vanthornhout, J., Simon, J. Z., Francart, T., and Brodbeck, C. (2021). Neural markers of speech comprehension: measuring EEG tracking of linguistic speech representations, controlling the speech acoustics. J. Neurosci. 41, 10316–10329. doi: 10.1523/JNEUROSCI.0812-21.2021

PubMed Abstract | Crossref Full Text | Google Scholar

Hansson-Sandsten, M. (2010). Evaluation of the optimal lengths and number of multiple windows for spectrogram estimation of SSVEP. Medi. Eng. Phys. 32, 372–383. doi: 10.1016/j.medengphy.2010.01.009

PubMed Abstract | Crossref Full Text | Google Scholar

Hansson-Sandsten, M. (2011). “Cross-spectrum and coherence function estimation using time-delayed Thomson multitapers,” in Proceedings of the ICASSP (Prague: IEEE).

Google Scholar

Karnik, S., Romberg, J., and Davenport, M. (2022). Thomson's multitaper method revisited. IEEE Transact. Inf. Theory 68, 4864–4891. doi: 10.1109/TIT.2022.3151415

Crossref Full Text | Google Scholar

Keding, O., Alickovic, E., Skoglund, M., and Sandsten, M. (2023). “Coherence estimation tracks auditory attention in listeners with hearing impairment,” in 24th Interspeech Conference (Dublin).

Google Scholar

Lesenfants, D., and Francart, T. (2020). The interplay of top-down focal attention and the cortical tracking of speech. Sci. Rep. 10:6922. doi: 10.1038/s41598-020-63587-3

PubMed Abstract | Crossref Full Text | Google Scholar

Lunner, T., Alickovic, E., Graversen, C., Ng, E. H. N., Wendt, D., and Keidser, G. (2020). Three new outcome measures that tap into cognitive processes required for real-life communication. Ear Hear. 41(Suppl. 1):39S. doi: 10.1097/AUD.0000000000000941

PubMed Abstract | Crossref Full Text | Google Scholar

McCoy, E., Walden, A., and Percival, D. (1998). Multitaper spectral estimation of power law processes. IEEE Transact. Signal Process. 46, 655–668. doi: 10.1109/78.661333

Crossref Full Text | Google Scholar

McHaney, J. R., Gnanateja, G. N., Smayda, K. E., Zinszer, B. D., and Chandrasekaran, B. (2021). Cortical tracking of speech in delta band relates to individual differences in speech in noise comprehension in older adults. Ear Hear. 42, 343–354. doi: 10.1097/AUD.0000000000000923

PubMed Abstract | Crossref Full Text | Google Scholar

O'sullivan, J. A., Power, A. J., Mesgarani, N., Rajaram, S., Foxe, J. J., Shinn-Cunningham, B. G., et al. (2015). Attentional selection in a cocktail party environment can be decoded from single-trial EEG. Cereb. Cortex 25, 1697–1706. doi: 10.1093/cercor/bht355

PubMed Abstract | Crossref Full Text | Google Scholar

Reinhold, I., and Sandsten, M. (2022). The multitaper reassigned spectrogram for oscillating transients with Gaussian envelopes. Signal Process. 198:108570. doi: 10.1016/j.sigpro.2022.108570

Crossref Full Text | Google Scholar

Schmitt, R., Meyer, M., and Giroud, N. (2022). Better speech-in-noise comprehension is associated with enhanced neural speech tracking in older adults with hearing impairment. Cortex 151, 133–146. doi: 10.1016/j.cortex.2022.02.017

PubMed Abstract | Crossref Full Text | Google Scholar

Shahsavari Baboukani, P., Graversen, C., Alickovic, E., and Østergaard, J. (2022). Speech to noise ratio improvement induces nonlinear parietal phase synchrony in hearing aid users. Front. Neurosci. 16:932959. doi: 10.3389/fnins.2022.932959

PubMed Abstract | Crossref Full Text | Google Scholar

Slepian, D. (1978). Prolate spheroidal wave functions, Fourier analysis, and uncertainty—v: the discrete case. Bell Syst. Tech. J. 57, 1371–1430. doi: 10.1002/j.1538-7305.1978.tb02104.x

Crossref Full Text | Google Scholar

Te, N., and Ap, H. (2002). Nonparametric permutation tests for functional neuroimaging: a primer with examples. Hum. Brain Mapp. 15:1058. doi: 10.1002/hbm.1058

PubMed Abstract | Crossref Full Text | Google Scholar

Thomson, D. J. (1982). Spectrum estimation and harmonic analysis. Proc. IEEE 70, 1055–1096. doi: 10.1109/PROC.1982.12433

Crossref Full Text | Google Scholar

Vander Ghinst, M., Bourguignon, M., Wens, V., Naeije, G., Ducène, C., Niesen, M., et al. (2021). Inaccurate cortical tracking of speech in adults with impaired speech perception in noise. Brain Commun. 3:fcab186. doi: 10.1093/braincomms/fcab186

PubMed Abstract | Crossref Full Text | Google Scholar

Viswanathan, V., Bharadwaj, H., and Shinn-Cunningham, B. (2019). Electroencephalographic signatures of the neural representation of speech during selective attention. eNeuro 6:ENEURO.0057-19.2019. doi: 10.1523/ENEURO.0057-19.2019

PubMed Abstract | Crossref Full Text | Google Scholar

Walden, A. T. (2000). A unified view of multitaper multivariate spectral estimation. Biometrika 87, 767–788. doi: 10.1093/biomet/87.4.767

Crossref Full Text | Google Scholar

Welch, P. D. (1967). The use of fast Fourier transform for the estimation of power spectra: a method based on time averaging over short, modified periodograms. IEEE Transact. Audio Electroacoust. 15, 70–73. doi: 10.1109/TAU.1967.1161901

Crossref Full Text | Google Scholar

Keywords: coherence, EEG, neural speech tracking, auditory attention, multitapers, hearing impairment

Citation: Keding O, Alickovic E, Skoglund MA and Sandsten M (2024) Novel bias-reduced coherence measure for EEG-based speech tracking in listeners with hearing impairment. Front. Neurosci. 18:1415397. doi: 10.3389/fnins.2024.1415397

Received: 10 April 2024; Accepted: 16 October 2024;
Published: 06 November 2024.

Edited by:

Torsten Dau, Technical University of Denmark, Denmark

Reviewed by:

Edmund C. Lalor, University of Rochester, United States
Waldo Nogueira, Hannover Medical School, Germany

Copyright © 2024 Keding, Alickovic, Skoglund and Sandsten. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Oskar Keding, b3NrYXIua2VkaW5nJiN4MDAwNDA7bWF0c3RhdC5sdS5zZQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.