Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci. , 28 March 2025

Sec. Brain Imaging Methods

Volume 19 - 2025 | https://doi.org/10.3389/fnins.2025.1550581

Functional connectivity in EEG: a multiclass classification approach for disorders of consciousness

\r\nSreelakshmi RaveendranSreelakshmi Raveendran1Kala SKala S1Ramakrishnan A G,Ramakrishnan A G2,3Raghavendra KenchaiahRaghavendra Kenchaiah4Jayakrushna SahooJayakrushna Sahoo5Santhos KumarSanthos Kumar1Farsana M KFarsana M K4Ravindranadh Chowdary MundlamuriRavindranadh Chowdary Mundlamuri4Sonia BansalSonia Bansal6Binu V SBinu V S7Subasree R
Subasree R4*
  • 1Department of Electronics and Communication Engineering, Indian Institute of Information Technology Kottayam, Kerala, India
  • 2Department of Electrical Engineering, Indian Institute of Science, Bangalore, India
  • 3Centre for Neuroscience, Indian Institute of Science, Bangalore, India
  • 4Department of Neurology, NIMHANS, Bangalore, India
  • 5Department of Computer Science and Engineering, Indian Institute of Information Technology Kottayam, Kerala, India
  • 6Department of Neuroanaesthesia and Neurocritical Care, NIMHANS, Bangalore, India
  • 7Department of Biostatistics, NIMHANS, Bangalore, India

Characterizing functional connectivity (FC) in the human brain is crucial for understanding and supporting clinical decision making in disorders of consciousness. This study investigates FC using sliding window correlation (SWC) analysis of electroencephalogram (EEG) applied to three connectivity measures: phase-lag index (PLI) and weighted phase-lag index (wPLI), which quantify phase synchronization, and amplitude envelope correlation (AEC), which captures amplitude-based coactivation patterns between pairs of channels. SWC analysis is performed across the five canonical frequency bands (delta, theta, alpha, beta, gamma) of EEG data from four distinct groups: coma, unresponsive wakefulness syndrome, minimally conscious state, and healthy controls. The extracted SWC metrics, mean, reflecting the stability of connectivity, and standard deviation, indicating variability, are analyzed to discern FC differences at the group level. Multiclass classification is attempted using various models of artificial neural networks that include different multilayer perceptrons (MLP), recurrent neural networks, long-short-term memory networks, gated recurrent units, and a hybrid CNN-LSTM model that combines convolutional neural networks (CNN) and long-short-term memory network to validate the discriminative power of these FC features. The results show that MLP model 2 achieves a classification accuracy of 96.3% using AEC features obtained with a window length of 16s, highlighting the effectiveness of AEC. An evaluation of the model performance for different window sizes (16 to 20 s) shows that MLP model 2 consistently achieves high accuracy, ranging from 95.5% to 96.3%, using AEC features. When AEC and wPLI features are combined, the maximum accuracy increases to 96.9% for MLP model 2 and 96.7% for MLP model 3, with a window size of 17 seconds in both cases.

1 Introduction

Functional connectivity (FC) has emerged as a transformative concept in neuroscience, offering a deeper understanding of brain function. FC allows researchers to explore brain's adaptability to external stimuli, cognitive challenges, and internal states (Sastry and Banerjee, 2024; Anusha et al., 2025). FC provides a more detailed picture of neural interactions by examining the synchronization patterns between different brain regions. This approach improves our understanding of the functional organization and flexibility of the brain. The ability to capture changes in FC is important for understanding complex cognitive processes and the brain's responses to ever-changing demands (Sastry and Banerjee, 2024).

Technologies such as functional magnetic resonance imaging (fMRI) (Mäki-Marttunen, 2014), electroencephalography (EEG) (Duclos et al., 2021; Dey et al., 2023), magnetoencephalography (MEG) (Jin et al., 2023), and functional near-infrared spectroscopy (fNIRS) (Tang et al., 2021) have been instrumental in advancing the study of functional connectivity. These neuroimaging modalities, supported by advancements in sensor resolution and computational techniques, enable the recording and analysis of FC patterns with high precision. Each modality offers unique advantages: fMRI detects blood oxygenation changes at a high spatial resolution, enabling whole-brain connectivity mapping (Mäki-Marttunen, 2014). EEG delivers exceptional temporal resolution by capturing electrical activity from the scalp, tracking rapid neural changes (Duclos et al., 2021). MEG measures magnetic fields produced by neural activity (Jin et al., 2023) at good spatial and temporal precision, while fNIRS uses near-infrared light to monitor cortical blood flow, offering portability and robustness to movement artifacts (Tang et al., 2021). These modalities have been particularly effective in studying resting-state networks, such as the default mode network and the salience network (Matsui and Yamashita, 2023). The spontaneous fluctuations within these networks illuminate baseline cognitive functions and introspective processes. A commonly employed method to analyze FC is the sliding window approach, which calculates connectivity metrics within overlapping time windows (Mokhtari et al., 2019). This technique allows researchers to track fluctuations in connectivity, providing insights into brain networks over time. The sliding window approach segments time series data into consecutive, overlapping windows of a specified length. Within each window, connectivity measures, such as the Pearson correlation coefficient, are calculated between pairs of brain regions, or electrode pairs (Saideepthi et al., 2023). Shifting the window across the data set can generate a time-resolved representation of connectivity. This method helps identify transient connectivity states and understand the brain's adaptability to various cognitive tasks and resting conditions.

Functional connectivity has facilitated significant advancements in medical research, particularly in the diagnosis and monitoring of neurological disorders such as Alzheimer's disease (Matsui and Yamashita, 2023), epilepsy (Qin et al., 2024), and stroke (Wu et al., 2024). Alterations in FC patterns have been associated with cognitive impairments in Alzheimer's patients, providing biomarkers of disease progression (Arbabyazd et al., 2023). In epilepsy, FC has offered critical insight into seizure mechanisms, which has helped to develop targeted therapies (Li et al., 2024b). It has also deepened our understanding of altered states of consciousness, such as those induced by anesthesia (Miao et al., 2023), psychedelics (Soares et al., 2024), and sleep (Xu et al., 2024). Under anesthesia, studies using techniques such as sliding window correlation (SWC) have revealed a reduction in the repertoire of FC states, reflecting a loss of the characteristic of dynamic flexibility of the brain of consciousness (Miao et al., 2023). However, psychedelics such as psilocybin expand the repertoire of FC and increase entropy, indicating heightened consciousness (Soares et al., 2024). Similarly, sleep studies reveal that deep sleep narrows the FC repertoire, reducing variability and diminishing the capacity for consciousness (Xu et al., 2024).

Beyond altered states, FC has proven valuable in neuropsychiatry (Matsui and Yamashita, 2023) by identifying disrupted connectivity patterns in conditions such as schizophrenia, depression, and anxiety. In schizophrenia, disruptions in FC correlate with cognitive deficits (Cattarinussi et al., 2023), while in depression, altered patterns are linked to symptom severity and treatment response, enabling personalized therapy. FC also elucidates developmental trajectories, showing increased stability and adaptability during maturation and reduced flexibility in aging and neurodegeneration, which reflects declining neural plasticity. Advances in computational techniques, including sliding window correlation, hidden Markov models, and multivariate autoregressive models, have enhanced the precision of FC analyses (Hutchison et al., 2013).

Beyond medical applications, FC contributes to cognitive science by revealing network interactions in processes like memory, attention, and learning. This knowledge informs strategies to enhance cognitive performance and supports advancements in brain-computer interfaces (BCIs). BCIs leverage FC to decode brain activity, enabling communication for individuals with disabilities, controlling prosthetic devices, and enhancing immersive gaming experiences (Fallani and Bassett, 2019). FC has also been used successfully in EEG-based biometrics (Kumar et al., 2022). FC also informs neuroergonomics by optimizing task environments to align with human cognitive capabilities, reducing mental fatigue, and improving performance (Ismail and Karwowski, 2020). It sheds light on social interactions by analyzing synchronized FC patterns during communication, offering insights into the neural basis of empathy and social connectivity. FC has found transformative applications in understanding and managing disorders of consciousness (DOC). In these cases, FC differentiates between minimally conscious states (MCS) and unresponsive wakefulness syndrome (UWS), since reduced variability and connectivity integration are associated with lower levels of consciousness. Stronger connectivity patterns in MCS patients align with a relatively higher degree of awareness (Naro et al., 2018). Clinically, FC assists in prognosis by identifying recovery markers, such as restored corticothalamic connectivity, and reveals covert awareness in unresponsive patients, guiding personalized care and rehabilitation. While existing literature highlights the potential of FC in understanding and managing consciousness in DOC cohorts, significant ambiguity persists in effectively distinguishing between various consciousness states and healthy controls. This lack of clarity hampers the development of reliable diagnostic and prognostic tools, creating a need for advanced analytical approaches.

This work investigates the classification of different states of consciousness: coma, UWS, MCS, and healthy controls (HC) using functional connectivity features derived from resting state EEG data. By leveraging metrics such as wPLI and AEC, the study aims to analyze brain connectivity patterns to differentiate between these states. The performance analysis has been conducted for different window sizes (16 to 20 s with a step size of 1 s). The analysis incorporates sliding window correlation to explore the variability in connectivity, providing insights into how brain network characteristics vary across different levels of consciousness.

2 Materials and methods

2.1 Resting state EEG measurement

The resting state brain activity of the subjects recruited for the study is acquired using the EEG machine with Galileo software at NIMHANS, Banglore. The recording duration is chosen as 30 minutes. Electrodes are set according to the international 10/10 system, using a sampling rate of 256 Hz and ensuring that impedance remains below 5 kΩ. The EEG data is referenced to the left (A1) and right earlobes (A2). The dataset includes 16 patients from coma, 20 subjects from UWS, MCS, and healthy controls. EEG recordings are performed while participants are positioned supine with their eyes closed, and methods are implemented to reduce artifacts (Raveendran et al., 2023). Consultant neurologists have recruited and classified the patients into various consciousness categories based on established behavioral assessment methods of CRS-R (coma recovery scaled-revised) and GCS (Glasgow coma scale). Further details regarding patient recruitment, inclusion, and exclusion criteria, as well as the signal acquisition protocol, are available in Raveendran et al. (2024).

2.2 Data preprocessing

The resting-state EEG signals are visually inspected by expert EEG technicians to detect bad channels, artifact-affected epochs, and recorded annotations. Subsequently, appropriate channels are selected and average referencing is applied to the signal. The preprocessing steps applied include trimming the first and last two minutes to remove initial setup artifacts and ending noise followed by adding a standard 10-20 montage to ensure correct spatial mapping of electrodes. A band-pass filter (1–48 Hz) is applied to remove low-frequency drifts and high-frequency muscle noise, followed by a notch filter at 50 Hz to eliminate power line interference. Bad segments were automatically rejected based on annotations. The five distinct EEG frequency bands were extracted from the preprocessed signals using a zero-phase, non-causal FIR bandpass filter. This filter was designed with a windowed time-domain approach (firwin), setting the lower and upper cut-off frequencies according to the respective EEG bands. A Hamming window was applied, featuring a passband ripple of 0.0194 and a stopband attenuation of 53 dB (Raveendran et al., 2024). To maintain the same number of subjects in each group, 15 patients are selected for analysis. The selection of the 15 subjects was based on two key criteria. First, only subjects with available follow-up data were included to ensure longitudinal consistency in the analysis. Second, subjects with the longest EEG signal duration after preprocessing were chosen to maximize the number of extracted windows, thereby enhancing the reliability and robustness of the analysis. A 20-s sliding window with 50% overlap is applied to each record to generate analysis segments. Windows are extracted from each channel and frequency band. 95 windows are extracted from each signal of each patient, taking it to a total of 475 windows from each subject. The total number of windows in each group is 7,125; making a data sample set of 28,500 for all four classes. For analysis, the FC features wPLI and AEC are extracted from these 28,500 windows. The study also explores the impact of different window lengths (ranging from 16 to 20 s) on data segmentation and the overall dataset size. A 16 s window size results in 36,000 data samples, with each group contributing 9,000 samples.

3 Methodology

3.1 Feature extraction

3.1.1 Phase lag index (PLI)

The phase lag index (PLI) is a metric widely used in neuroscience for quantifying functional connectivity between two signals based on their phase relationships. PLI focuses on the consistency of non-zero phase lags, making it robust against the effects of common-source artifacts. This robustness is advantageous in studying brain connectivity, where eliminating spurious correlations is critical for uncovering genuine interactions. PLI is equal to 0 when there is no consistent phase lag between signals, indicating that the phase differences are symmetrically distributed around zero, which can occur in cases of perfect synchrony or random phase differences. When the phase difference is consistently ±π (exactly 180 degrees out of phase), then the rate of change or slope of phase difference with respect to frequency is zero. Hence, PLI is zero.

The phase, which encodes information about the temporal coordination of oscillatory activity, is extracted from the analytic representation of a real-valued signal x(t) using its Hilbert transform H(x(t)). The analytic signal z(t) of x(t) is expressed as:

z(t)=x(t)+iH(x(t))

The amplitude envelope Ax(t) is the magnitude of the analytic signal:

Ax(t)=|z(t)|=x(t)2+H(x(t))2

and its instantaneous phase is given by:

ϕx(t)=tan-1(Im(z(t))Re(z(t)))

For two signals x(t) and y(t), the instantaneous phase difference is:

Δϕ(t)=ϕx(t)-ϕy(t)

The PLI is derived from these phase differences and emphasizes the asymmetry of consistent phase lags (Stam et al., 2007). It is defined as:

PLI=|𝔼[sgn(sin(Δϕ(t))]|

Here, 𝔼[·] denotes the expectation (or average) over time, sgn(sin(Δϕ(t))) is the sign of the sine of the phase difference. The PLI quantifies the consistency of directional phase relationships by focusing on whether one signal consistently leads or lags the other. The normalized range between 0 and 1 makes it well-suited for comparing connectivity patterns across different datasets and experimental conditions, establishing it as a robust measure for studying functional brain networks.

3.1.2 Weighted phase lag index (wPLI)

The weighted phase lag index (wPLI) extends the PLI by incorporating the magnitude of phase lag asymmetry, providing a more sensitive measure of functional connectivity. Like PLI, the wPLI focuses on non-zero phase lags. However, by weighting the contribution of phase differences based on the magnitude of their imaginary components, the wPLI enhances sensitivity to true underlying interactions, particularly in noisy data (Vinck et al., 2011). With the instantaneous phases of the signals derived from their analytic representations, the wPLI is calculated as:

wPLI=|𝔼[Im(Δϕ(t))·sgn(Im(Δϕ(t)))]|𝔼[|Im(Δϕ(t))|]

where, Im(Δϕ(t)) is the imaginary part of the phase difference, and sgn(Im(Δϕ(t))) is the sign function (Vinck et al., 2011). The numerator emphasizes the magnitude of phase lag asymmetry, while the denominator normalizes the result to bound wPLI values between 0 and 1. A wPLI value close to 0 indicates no consistent phase lag, while values close to 1 indicate strong phase synchronization with consistent lag. The bounded range of wPLI facilitates comparisons across datasets and conditions, making it a reliable measure for studying functional brain networks.

Both PLI and wPLI quantify the consistency of phase differences between signals; however they differ in their weighting mechanisms. PLI is based on the sign of the imaginary component of the cross-spectrum, measuring whether phase differences are systematically positive or negative. PLI equally weights all samples, including those where the phase difference is close to 0 or ±π which can introduce noise and reduce sensitivity. wPLI addresses this limitation by weighting the contribution of each sample by the magnitude of the imaginary component of the cross-spectrum. Consequently, wPLI provides a more reliable measure of true brain interactions by emphasizing phase differences with stronger imaginary components, thereby improving robustness against common-source artifacts.

3.1.3 Amplitude envelope correlation (AEC)

While phase-based measures like wPLI assess temporal coordination in oscillatory activity, amplitude-based methods such as amplitude envelope correlation (AEC) provide complementary insights by focusing on the co-variation of signal intensities (Bruns et al., 2000). AEC specifically quantifies the linear relationship between the amplitude envelopes of two signals over time, capturing how their intensities fluctuate in tandem.

The amplitude envelopes required for AEC are obtained from the same analytic signal derived using the Hilbert transform, as described earlier. Given the amplitude envelopes Ax(t) and Ay(t) of two signals x(t) and y(t), AEC is calculated as the Pearson correlation coefficient between these envelopes over a specified time interval T:

AEC=t=1T(Ax(t)-Āx)(Ay(t)-Āy)t=1T(Ax(t)-Āx)2t=1T(Ay(t)-Āy)2

where Āx and Āy are the respective mean values of Ax(t) and Ay(t) over T.

AEC captures slow fluctuations in signal power, often associated with large-scale brain network activity, such as in studies of the resting state. AEC focuses on the comodulation of signal envelopes, making it a functionally relevant metric. Although AEC does not inherently eliminate volume conduction due to the presence of zero-lag correlations, it provides valuable insights into frequency-dependent connectivity beyond pure phase synchrony.

4 Sliding window correlation (SWC)

Sliding window correlation is a widely used method for examining the time-varying relationships between two signals or datasets. Traditional static correlation assumes that the relationship between signals remains constant over time. However, these relationships are often nonstationary and evolve in real-world scenarios, such as neural activity, physiological signals, and financial systems. Sliding window correlation has emerged as a prominent method for capturing these time-varying dependencies.

SWC involves partitioning the signals into smaller temporal windows, calculating the correlation within each window, and observing how the correlation evolves as the window “slides” across the signals. This method has been instrumental in understanding functional connectivity, particularly in neuroscience, where it provides insights into the temporal fluctuations of brain network interactions.

SWC builds on the standard Pearson correlation coefficient, a widely used metric for measuring the linear relationship between two-time series. The Pearson correlation coefficient is defined as:

r=t=1T(Xt-X̄)(Yt-Ȳ)t=1T(Xt-X̄)2t=1T(Yt-Ȳ)2

where; Xt, Yt are the values of the two signals at time t and X̄, Ȳ are the mean values of Xt and Yt over the time interval T. A 20-second sliding window with 50% overlap is applied to obtain the windows of the input EEG signal. Within each window k, the Pearson correlation is computed as:

rk=t=tktk+W-1(Xt-X̄k)(Yt-Ȳk)t=tktk+W-1(Xt-X̄k)2t=tktk+W-1(Yt-Ȳk)2

Here tk is the starting time of the k-th window, Xk̄ and Yk̄ are the mean values of Xt and Yt within the k-th window. The window is then shifted by a step size S and the process is repeated until the entire signal is analyzed. The result is a time series of correlation coefficients rk representing the temporal evolution of the relationship between X and Y. The proposed methodology is graphically depicted in Figure 1.

Figure 1
www.frontiersin.org

Figure 1. Graphical depiction of the proposed methodology: FC-based multiclass classification of DOC patients.

5 Choice of ANN models for multiclass classification

The proposed work employs several advanced artificial neural network (ANN) architectures to classify states of consciousness–coma, UWS, MCS, and healthy subjects–using functional connectivity features, namely PLI, wPLI and AEC. The architectures include three variations of multilayer perceptron (MLP), simple recurrent neural network (RNN), long short-term memory network (LSTM), gated recurrent units (GRUs), and a hybrid CNN-LSTM model, each tailored to optimize the learning process for different data representations.

The preprocessed signal has 19 channels: F3, F4, C3, C4, F7, F8, P3, P4, T7, T8, P7, P8, O1, O2, Fp1, Fp2, Fz, Pz, and Cz. Each channel is divided into five frequency bands (delta, theta, alpha, beta, and gamma). Thus each connectivity feature gives rise to a symmetric feature matrix of size 19 × 19 for each frequency band. Due to its symmetry, only the upper triangular elements, excluding the diagonal, are used for classification. Hence, the number of unique elements is 171 per frequency band. With 5 frequency bands, the total feature vector for each window contains 5 × 171 = 855 values. For 20 s window size, each subject generates 95 windows, resulting in a feature matrix of shape (95,855) per subject. With 15 subjects per group, the total number of samples per group is 15 × 95 = 1,425. Considering all four groups, the dataset comprises 4 × 1,425 = 5,700 samples, each represented by a feature vector of dimension 855. After reshaping the data into a 2D array, the final input has a shape of (5,700,855). For a window size of 19 s, each subject produces 103 windows, leading to a feature matrix of shape (103, 855) per subject. Given 15 subjects per group, the total number of samples per group is 1,545 (15 × 103). Considering four groups, the input comprises 6,180 samples (4 × 1,545). When the window size is 18 s, each subject generates 109 windows, resulting in a final feature matrix of shape (6,540,855) for all the groups together (109 X 15 4 = 6,540). For 17 s window size, each subject produces 115 windows, resulting in a dataset of 6,900 feature vectors. With a window size of 16 s, each subject generates 120 windows, finally resulting in 7,200 feature vectors.

5.1 MLP models

The MLP model 1 starts with an input layer that accepts flattened input data of shape (X_train.shape[1]), where the number of neurons corresponds to the total features in the dataset. This ensures compatibility with fully connected layers and simplifies computations. It employs a dense hidden layer with 128 neurons and ReLU activation, followed by dropout regularization with a rate of 0.5 to prevent overfitting. A second hidden layer contains 64 neurons, also with ReLU activation, and applies a dropout of 0.3 for additional regularization. The output layer has four neurons (one for each class), with a softmax activation function to produce class probabilities. The optimizer used is Adam with default hyperparameters, and the loss function is categorical crossentropy, suitable for multiclass classification. Training involves 50 epochs with a batch size of 32, using 10% of the training data as a validation split.

The MLP model 2 builds upon the first model and starts with an input layer identical to model 1, where the number of neurons matches the feature size (X_train.shape[1]). It incorporates a dense hidden layer with 256 neurons and ReLU activation, followed by batch normalization to stabilize and accelerate training. Dropout is applied with a rate of 0.4. A second hidden layer with 128 neurons employs ReLU activation, batch normalization, and dropout at 0.4. A third hidden layer contains 64 neurons with ReLU activation and dropout at 0.3. The second and third layers introduce kernel regularization with L2 penalty (0.001) to further combat overfitting. Like model 1, the output layer uses softmax activation with four units, optimized by Adam with a learning rate of 0.001. Training parameters are identical to those of model 1.

MLP model 3 resembles model 2 but maintains an identical input layer, where the number of neurons is X_train.shape[1]. It includes two hidden layers with 256 and 128 neurons, respectively, both employing ReLU activation, batch normalization, dropout (0.4), and L2 regularization (0.001). The output layer is configured identically to the first two models, and the optimizer is Adam. Early stopping and learning rate reduction callbacks are used to halt training when validation performance stops improving, ensuring optimal resource usage.

5.2 Sequential models: simple RNN, LSTM, and GRU

The simple RNN model processes data as sequences, requiring input to be reshaped into (time_steps, features). The first recurrent layer employs 128 simple RNN units with tanh activation and return_sequences = True to retain outputs for the next layer. Dropout regularization is applied with a rate of 0.4. A second recurrent layer with 64 simple RNN units has return_sequences = False, followed by dropout at 0.3. Batch normalization is applied after each layer to stabilize gradients. Dense layers are used post-RNN, starting with a layer of 32 units, ReLU activation, and dropout of 0.2, culminating in an output layer with four softmax-activated neurons. Adam optimizer is used with a learning rate of 0.0005, and categorical cross-entropy is the loss function. The performance evaluation metrics include accuracy, precision, recall, and F1-score.

The LSTM model leverages the long-term memory capabilities of LSTM layers. The first LSTM layer has 128 units, return_sequences = True, and applies dropout at 0.4 to reduce overfitting. The second LSTM layer, with 64 units and return_sequences = False, uses dropout at 0.3. Batch normalization is applied after each layer. Post-LSTM dense layers include a 32-unit layer with ReLU activation and dropout at 0.2, followed by the output layer with four softmax-activated neurons. Adam optimizer with a learning rate of 0.0005 is used, along with early stopping and learning rate reduction callbacks.

The GRU model is structured similarly to the LSTM model but replaces LSTM layers with GRU layers, which are computationally more efficient while retaining the ability to model sequential dependencies. The first GRU layer has 128 units with return_sequences = True and dropout at 0.4. The second GRU layer, with 64 units and return_sequences = False, uses dropout at 0.3. Batch normalization follows each GRU layer. The dense layers include a 32-unit layer with ReLU activation and dropout at 0.2, followed by the output layer with four softmax-activated neurons. Like the LSTM model, this configuration uses Adam optimizer (learning rate 0.0005), categorical cross-entropy, and early stopping to ensure optimal training performance.

A hybrid CNN-LSTM model integrates convolutional and recurrent architectures for effective sequential learning. The input data is structured as (time_steps, num_features), capturing temporal dependencies. The model begins with two Conv1D layers (64 and 128 filters) using ReLU activation, followed by MaxPooling1D layers to downsample the feature maps. Batch normalization is applied after each convolutional layer to stabilize training. The extracted spatial features are then passed through two LSTM layers (64 and 32 units) with tanh activation, modeling sequential dependencies. Dropout (0.3 and 0.2) is used in LSTM layers to prevent overfitting. A fully connected dense layer with 32 neurons (ReLU activation) and dropout (0.2) is followed by a softmax-activated output layer for multi-class classification. The model is trained using categorical cross-entropy loss and optimized with the Adam optimizer (learning rate 0.0005). It is trained with a batch size of 32 for 50 epochs, incorporating early stopping and a learning rate reduction scheduler to enhance convergence and performance. Table 1 describes the network configurations and the model details.

Table 1
www.frontiersin.org

Table 1. Description of the implemented ANN models for classification.

Training and Evaluation: All models are trained using standardized inputs to ensure consistency across features. Labels are one-hot encoded to suit the categorical classification task. During training, 10% of the data is reserved for validation to monitor overfitting. Metrics such as accuracy, weighted precision, recall, and F1-score provide a holistic evaluation of model performance, especially for imbalanced class distributions.

6 Results

Table 2 compares the six models using the performance metrics of accuracy, precision, recall, and F1-score. The metrics used compare the effectiveness of various artificial neural network models–3 MLP models, simple RNN, LSTM, and GRU–for classifying states of consciousness based on the functional connectivity features of phase lag index, weighted phase lag index, and amplitude envelope correlation. Table 2 demonstrates the distinct trends influenced by the choice of features and model architectures in leveraging the FC data to classify states of consciousness into coma, UWS, MCS, and healthy subjects. The analysis has been conducted using different window sizes to examine the impact of window size on model performance. The applied sizes of the window are 16, 17, 18, 19, and 20 s, with a 50% overlap maintained to capture variability.

Table 2
www.frontiersin.org

Table 2. Performance comparison of different ANN models for multiclass classification with 20s window length and 80/20 train-test split (coma vs. UWS vs. MCS vs. HC) using three different features and a combination of two of them.

6.1 Performance analysis on PLI features

For the PLI feature, MLP model 2 exhibits the highest performance across all metrics, with an accuracy of 0.622, precision of 0.626, recall of 0.622, and an F1-score of 0.624. These results demonstrate that MLP model 2 achieves a slightly better balance between precision and recall, leading to the highest F1-score among all the evaluated models. MLP models 1 and 3 follow closely in terms of performance, with MLP model 3 showing slightly better results than MLP model 1. The RNN-based models perform poorer. The simple RNN model achieves an accuracy of 0.52, precision of 0.529, recall of 0.52, and an F1-score of 0.524. Among the RNN-based architectures, the GRU model marginally outperforms the simple RNN and LSTM models, with an F1-score of 0.544, reflecting a slightly better precision-to-recall balance. The LSTM model achieves slightly better than the simple RNN with accuracy, precision, recall, and F1-score values of 0.533, 0.543, 0.533, and 0.537, respectively.

Table 3 presents the accuracy of different models for PLI across varying window sizes (16–20 s) using an 80-20 fixed train-test split. The results showed that larger window sizes (19 and 20 s) lead to better model performance, particularly for MLP architectures. Among the models, MLP Model 2 consistently achieves higher accuracy, peaking at 0.622 for the 20 s window, followed closely by MLP Model 3 at 0.6183. Recurrent models (simple RNN, LSTM, and GRU) exhibit lower overall performance. Both LSTM and GRU models improve in performance as the window size increases, reaching higher accuracy at 20 s (0.5325 and 0.5292, respectively).

Table 3
www.frontiersin.org

Table 3. Accuracy with PLI features for different window sizes with 80-20 fixed train-test split.

Table 4 presents model performance results using group k-fold cross-validation with PLI as the feature extraction method. The results show that MLP-based models consistently outperform RNN-based models, with MLP model 2 achieving the highest accuracy and F1-score across all window sizes. An increase in window size leads to improved performance, particularly for MLP models, highlighting their ability to leverage longer temporal dependencies more effectively. Among the cross-validation folds, folds 10 and 8 produce better results, especially for MLP models 2 and 3, suggesting that these data splits are more favorable for efficiently training the models. In contrast, recurrent models simple RNN, LSTM, and GRU exhibit greater variability in performance and underperform compared to MLP-based models.

Table 4
www.frontiersin.org

Table 4. Best fold obtained using group k-fold cross validation and the performance matrix for each model using PLI.

Better results are obtained with MLP-based models, particularly models 2 and 3, using folds 8 and 10, and larger window sizes (19 and 20 s). The performance highlights that PLI-based features are better leveraged by MLP architectures rather than sequential models, potentially due to their ability to learn complex, high-dimensional representations more effectively.

6.2 Performance analysis on wPLI features

For the wPLI feature, which captures phase-based connectivity between brain regions, the performance shows significant variability across models. MLP model 1 emerges as the highest-performing model in terms of accuracy, achieving a value of 0.772. The high accuracy is supported by its precision of 0.773, which shows that it has a good ability to correctly identify positive predictions with relatively fewer false positives. Its recall value matches its accuracy at 0.772, suggesting that the model captures a large proportion of true positives. The F1-score of MLP model 1 is 0.773, reflecting a strong balance between precision and recall, which is critical in cases where the dataset may have an uneven class distribution or when both metrics are equally important for evaluation. This performance highlights that MLP model 1 can extract meaningful patterns from the wPLI feature set, leading to high classification success.

However, MLP models 2 and 3 significantly underperform. MLP model 2 achieves an accuracy of only 0.532. The corresponding precision (0.613), recall (0.532), and F1-score (0.570) further emphasize its struggle with generalization. MLP model 3 performs even worse, with an accuracy of 0.515. Both models exhibit poor adaptability to wPLI feature set, which may stem from their inability to effectively process the complex relationships embedded in the data.

The simple RNN model provides moderate performance, with an accuracy of 0.668. While it does not match the highest-performing MLP model 1 or GRU in accuracy, it surpasses MLP models 2 and 3 by a significant margin. The precision of 0.671 and recall of 0.668 are relatively balanced, and the F1-score of 0.669 confirms its consistent classification ability.

The LSTM network, a more advanced recurrent model, achieves an accuracy of 0.638. While this value is lower than that of simple RNN, its precision (0.646), recall (0.638), and F1-score (0.642) suggest that its performance is stable across metrics. However, LSTM does not perform as well as GRU, which surpasses it in all the metrics. This relative underperformance could be attributed to the additional complexity of LSTM, which may require more data and fine-tuning to optimize its performance on the wPLI feature set. LSTM's slightly lower accuracy compared to simple RNN highlights the trade-off between model complexity and effective learning, especially when working with limited datasets.

The GRU model stands out as a better performer in this comparison. With accuracy, precision, recall, and F1-score of 0.713, 0.718, 0.713, and 0.716, respectively, GRU demonstrates a well-rounded ability to classify instances in the wPLI feature set. Its F1-score is notable, reflecting a fine balance between precision and recall. This performance advantage is likely due to GRU's simpler architecture than LSTM, which allows it to efficiently capture the patterns without overfitting or requiring excessive computational resources. GRU's high precision also suggests that it minimizes false positives, making it a reliable choice for tasks where avoiding incorrect classifications is critical.

When comparing the models holistically, MLP model 1 achieves the highest accuracy, since it excels in extracting patterns. GRU, on the other hand, provides a more balanced performance across all metrics. This makes GRU the most effective model for the wPLI feature, since it leverages the dataset's unique temporal properties while maintaining strong generalization.

Accuracy across multiple deep learning models based on wPLI and different window sizes (16–20 s) using an 80-20 fixed train-test split is shown in Table 5. The results imply that increasing the window size improved accuracy for the GRU and LSTM models. However, MLP Model 1 has remained the most robust choice, showing stable and high accuracy across all window sizes.

Table 5
www.frontiersin.org

Table 5. Accuracy with wPLI features for different window sizes with 80-20 fixed train-test split.

The group k-fold cross-validation results for models trained using wPLI-based features demonstrate varying performances across different architectures and window sizes is shown in Table 6. MLP Model 1 consistently outperforms all other models, achieving the highest accuracy of 0.785 at 20 s using fold 10, indicating its robustness in capturing relevant patterns from the data. The simple RNN model also performed well, reaching an accuracy of 0.7171 at 20 s, with Fold 8 yielding the best results in multiple cases.

Table 6
www.frontiersin.org

Table 6. Best fold obtained using group k-fold cross validation and the performance matrix for each model using wPLI.

Among the MLP models, models 2 and 3 exhibit moderate performance, with model 2 reaching its peak accuracy of 0.5798 at 20 s (fold 8) and model 3 achieving 0.5703 at 20 s (fold 5). However, these models show lower recall and F1-score than MLP Model 1, suggesting that they do not generalize as effectively.

Recurrent models, such as LSTM and GRU, demonstrate varying performance across different folds. The LSTM model performs the best at 19 s (fold 8) with an accuracy of 0.6378, while the GRU model reaches its peak at 19 s (fold 8) with an accuracy of 0.6515. This suggests that the performance of recurrent models remains highly dependent on the data split.

6.3 Performance analysis on AEC features

Using the AEC features, MLP model 2 achieves the highest overall performance, with an accuracy of 95.7%, precision of 95.8%, recall of 95.7%, and an F1-score of 95.8%. These values demonstrate the model's exceptional ability to classify instances correctly, minimize false positives, and identify true positives. The F1-score, which harmonizes precision and recall, confirms that MLP model 2 maintains a balanced performance without overemphasizing one aspect at the expense of the other. Its ability to leverage the AEC feature set likely stems from its optimized architecture, enabling it to capture intricate interconnections between features.

MLP model 3 also performs well, with an accuracy of 94.1%, precision of 94.2%, recall of 94.1%, and an F1-score of 94.2%. While its performance is lower than MLP model 2, the model demonstrates a robust ability to process the AEC feature set. It effectively generalizes across the dataset. The slight difference in performance compared to MLP model 2 could be attributed to architectural variations that make it less optimal in extracting fine-grained patterns.

MLP model 1 achieves an accuracy of 87.9%, precision of 87.9%, recall of 87.9%, and an F1-score of 87.9%. While these values are not as high as those of MLP models 2 and 3, model 1 still demonstrates an acceptable classification performance. It reliably identifies patterns in the AEC feature set, although its simpler architecture may limit its performance.

The simple RNN model performs well, achieving an accuracy of 90.5%, precision of 90.7%, recall of 90.5%, and an F1-score of 90.6%. These results highlight simple RNN's ability to generalize effectively on the AEC feature set, despite its recurrent architecture designed primarily for sequential data. Though its performance surpasses that of MLP model 1, it does not match the performance of MLP models 2 and 3, indicating that it does not exploit the full potential of the AEC feature set as efficiently as the MLP architectures.

The LSTM model exhibits a performance similar to simple RNN, with an accuracy of 90.6%, precision of 90.8%, recall of 90.6%, and an F1-score of 90.7%. This shows that while LSTM is a powerful model for sequential data, its complexity may not provide substantial benefits when applied to static connectivity features like AEC.

The GRU model achieves an accuracy of 90.3%, precision of 90.4%, recall of 90.3%, and an F1-score of 90.3%. While these values are close to those of simple RNN and LSTM, the slightly lower values suggest that it may be less suited for AEC than the other recurrent models.

Table 7 presents the accuracy of different deep learning models for AEC across various window sizes with an 80-20 fixed train-test split. MLP Model 2 consistently achieves higher accuracy across different window sizes, with values above 95%, making it the better-performing model for AEC. Among recurrent models, simple RNN performs better than LSTM and GRU, with its best accuracy at 17 s.

Table 7
www.frontiersin.org

Table 7. Accuracy with AEC features for different window sizes with 80-20 fixed train-test split.

The group k-fold cross-validation results using AEC-based features shown in Table 8 indicate that MLP models achieve higher accuracy across all tested models, with MLP Models 2 and 3 performing better. MLP Model 2 records a high accuracy of 0.9648 at 20 s (fold 8), followed closely by MLP Model 3 at 0.9648 at 19 s (fold 10). These results demonstrate the robust performance of MLP models when leveraging AEC-based features, suggesting that these models effectively capture the distinct patterns in the data. MLP Model 1 also shows strong performance, achieving its highest accuracy of 0.8895 at 16 s (fold 8), but is outperformed by the other two MLP models.

Table 8
www.frontiersin.org

Table 8. Best fold obtained using group k-fold cross validation and the performance matrix for each model using AEC.

Among RNN-based models, simple RNN exhibits the best performance, achieving an accuracy of 0.9225 at 18 s (fold 5), indicating that it effectively leverages AEC features. GRU and LSTM follow, with GRU reaching its highest accuracy of 0.8705 at 20 s (fold 8), while LSTM peaks at 0.8595 at 17 s (fold 8). These results highlight that RNN-based models, particularly simple RNN, still perform well with AEC-based features, though they did not surpass the MLP models in accuracy.

6.4 Performance analysis on combined features

MLP model 2 performs the best with accuracy, precision, recall, and F1-score values of 95.1%, 95.2%, 95.1%, and 95.1%, respectively. The high accuracy value indicates that the model reliably distinguishes between the classes, while the closely aligned precision and recall values reflect its ability to minimize false positives and false negatives effectively. However, these values are marginally lower than the values achieved by using AEC alone as the feature.

MLP model 3 performs slightly poorer than model 2. Model 3 achieves identical values of 94.8% for accuracy, precision, recall, and F1-score. The model's architecture appears well-suited for capturing the intricate relationships between wPLI and AEC features.

MLP model 1, with an accuracy of 87.4%, precision of 87.6%, recall of 87.4%, and an F1-score of 87.5%, exhibits a noticeable drop in performance compared to the other two MLP models. These values suggest that while MLP model 1 can handle the combined wPLI and AEC feature set, its simpler architecture might limit its capacity to fully capture the relationships between the two types of connectivity information.

Among the recurrent models, GRU achieves the highest performance, with equal values of accuracy, precision, recall, and F1-score of 92.1%. GRU's architecture captures the patterns of the combined wPLI-AEC features. LSTM performs poorer, with equal values of accuracy, precision, recall, and F1-score of 91.8%. The simple RNN model, while achieving an accuracy of 90.2%, precision of 90.4%, recall of 90.2%, and an F1-score of 90.3%, performs moderately well but falls short of the other recurrent and MLP architectures.

Table 9 displays the accuracy of various models utilizing AEC+wPLI for different window sizes (16-20 s) under an 80-20 train-test split. MLP-based models outperform recurrent models, with MLP model 2 consistently achieving better accuracy across all window sizes. Simple RNN performs better than LSTM and GRU in most cases, though GRU shows the highest accuracy of 0.9209 at 20 s, surpassing even LSTM. The best performance occurs at 18 s for MLP and simple RNN, while LSTM and GRU peak at 20 s.

Table 9
www.frontiersin.org

Table 9. Accuracy with AEC+wPLI features for different window sizes with 80-20 train test split.

Table 10 shows that the best fold obtained using group k-fold cross-validation across different models utilizing a combination of AEC and wPLI features varies based on the evaluation metrics. Among the models, MLP model 1 shows overall higher performance, achieving an accuracy of 0.9453 in its best fold (CV: 10). It also achieves a better precision (0.9463), recall (0.9453), and F1-score (0.9458). MLP Model 2 followed closely, with a best accuracy of 0.9383 (CV: 8), while MLP Model 3 has a slightly lower peak accuracy of 0.928 (CV: 8). Among recurrent models, GRU outperforms simple RNN and LSTM, achieving a better accuracy of 0.8581 (CV: 10), along with an F1-score of 0.8597, making it the strongest RNN-based model. The simple RNN and LSTM models display relatively lower accuracy, with the simple RNN peaking at 0.7518 (CV: 5) and LSTM at 0.756 (CV: 8), indicating that MLP models perform significantly better in this setup. Smaller window sizes (16-18 s) tend to yield better performance, especially for MLP models, which consistently achieve higher accuracy in this range. GRU models perform reliably across different window sizes but achieve their high results with 16 s, indicating that this window provides better informative features. Recurrent models (RNN and LSTM) exhibit some variability, with LSTM performing best at 16 s, while simple RNN shows slightly better results at 17 and 20 s.

Table 10
www.frontiersin.org

Table 10. Best performing fold in group k-fold cross validation and the performance metrics for each model using combined AEC and wPLI features.

The 95% confidence intervals are calculated for accuracy, precision, recall, and F1-score to evaluate the performance of different models using the four feature sets, namely PLI, wPLI, AEC, and combined AEC and wPLI. Among the models, MLP models 2 and 3 demonstrate higher accuracy, particularly with AEC (0.964, 95% CI = 0.964–0.965), while MLP model 1 performs slightly lower but improves significantly when using the combined AEC and wPLI (0.931, 95% CI = 0.909–0.953). The recurrent models (Simple RNN, LSTM, and GRU) show varying results, with GRU outperforming LSTM and RNN, particularly when using AEC and wPLI combinedly (0.853, 95% CI = 0.846–0.86). The CNN-LSTM hybrid model achieves better performance, attaining 0.931 accuracy (95% CI = 0.928–0.933) with AEC and 0.918 (95% CI = 0.908–0.929) with AEC and wPLI combined, making it comparable to the better-performing MLP models. Across all models, AEC consistently yields the highest performance, while the combination of AEC with wPLI further improves the results, particularly for MLP and GRU models. Table 11 presents the detailed confidence intervals for the implemented models.

Table 11
www.frontiersin.org

Table 11. Confidence interval for the implemented models using different features.

6.5 Sliding window correlation analysis

The sliding window correlation analysis of weighted phase-lag index and amplitude envelope correlation connectivity matrices across four groups: coma, UWS, MCS, and healthy controls provides critical insights into the brain's functional connectivity and its relationship to levels of consciousness. By examining the variability of connectivity patterns across five frequency bands (delta, theta, alpha, beta, gamma), the study captures the nature of neural interactions that underpin different conscious states. Two key metrics, mean SWC and standard deviation (Std. SWC), are used to quantify these patterns. The mean SWC reflects the overall similarity of connectivity across consecutive windows, while Std. SWC captures the variability of these patterns over time, offering a detailed view of the network dynamics.

6.5.1 SWC analysis for PLI

6.5.1.1 Group-level interpretation of mean SWC

In the delta band, mean SWC values of PLI rise steadily from coma (0.0196) to healthy controls (0.0593), reflecting enhanced functional connectivity as the brain transitions to more conscious states. This pattern is consistent across other frequency bands, such as theta (0.0208 in coma to 0.0540 in healthy controls), alpha (0.0222–0.0602), beta (0.0363–0.0554), and gamma (0.0183–0.0586). These increases suggest that higher levels of consciousness are associated with stronger interactions within brain networks. Among the groups, healthy controls consistently exhibit the highest mean SWC values across all frequency bands, indicating robust connectivity in fully conscious individuals. Conversely, coma patients display the lowest values, signifying diminished network interactions. UWS and MCS groups generally fall between these extremes, with subtle differences observed particularly in theta and beta bands.

6.5.2 Variability in connectivity

The standard deviation provides a measure of variability in connectivity across groups and reveals distinct trends. Healthy controls consistently show the highest variability across all frequency bands, with Std SWC values peaking in the delta (0.1446), theta (0.1428), alpha (0.1431), beta (0.1381), and gamma (0.1447) bands. This variability underscores the complexity and adaptability of brain network interactions in fully conscious individuals. In contrast, coma patients exhibit lower variability, indicating more uniform connectivity values. UWS and MCS groups show intermediate variabilities, with overlapping patterns in theta and gamma bands. This suggests that while both states share similarities in their connectivity, the greater heterogeneity in MCS may reflect the emergence of more complex neural interactions.

6.5.3 SWC analysis for wPLI

6.5.3.1 Group-level interpretation of mean SWC

Across all frequency bands, healthy controls consistently exhibit the highest mean SWC values, highlighting robust connectivity. In the delta band, HC have a mean SWC of 0.034, while coma patients have only 0.00003, showing diminished connectivity. Similarly, in the theta band, mean SWC is progressively higher across the groups: coma (0.0038), UWS (0.0061), MCS (0.029), and HC (0.054), reflecting a gradual recovery of stable network with improving consciousness levels. This trend is consistent across the alpha band, where mean SWC ranges from -0.0025 in coma to 0.035 in HC, and the beta band, where values range from -0.0067 in coma to 0.034 in HC. The gamma band, associated with high-frequency neural processes, shows a slightly different pattern. While HC has a high mean SWC of 0.030, the MCS group exhibits the highest mean SWC of 0.036, potentially indicating a unique pattern of gamma connectivity during partial recovery of consciousness. On the other hand, coma patients show reduced mean SWC values in the gamma band (0.0075), revealing a lack of high-frequency connectivity in them.

6.5.4 Variability in connectivity

High variability is often interpreted as an indicator of flexible and adaptive neural networks, which are hallmarks of healthy brain function. Here, healthy controls demonstrate significantly greater Std. SWC values across most frequency bands. In the theta band, HC show a Std. SWC of 0.236, compared to 0.178 in coma and 0.200 in UWS. Also in the alpha band, where HC have the highest Std. SWC (0.243), indicating network flexibility, coma patients exhibit the lowest Std. SWC (0.178). A similar pattern is seen in the beta band, with Std. SWC values increasing progressively from coma (0.178) to HC (0.241). In the gamma band, HC again displays the highest Std. SWC (0.244); MCS follows closely with 0.188, potentially reflecting a recovery in high-frequency network flexibility as consciousness is partially restored. Lower Std. SWC in coma patients (0.175) suggests that their FC networks are less capable of adapting, which is consistent with the rigidity observed in severely impaired states.

6.5.5 SWC analysis for AEC

6.5.5.1 Group-level interpretation of mean SWC

Healthy controls exhibit the highest mean SWC values across all frequency bands, underscoring their robust connectivity. HC show a mean SWC of 0.6487, significantly higher than coma (0.3093) and MCS (0.3098) in delta band. This indicates that low-frequency connectivity in healthy brains is consistently more stable than those in pathological states, reflecting a well-integrated and synchronized network. UWS patients, with a mean SWC of 0.4406 in the delta band, occupy an intermediate position, suggesting partial recovery of network compared to coma. In the theta band, HC maintains the highest mean SWC (0.6609), compared to coma (0.3458), UWS (0.4119), and MCS (0.3271). Theta oscillations are critical for memory and attention, and the marked difference in mean SWC between HC and pathological groups highlights their diminished capacity for cognitive processes. The alpha band, often associated with sensory processing and resting-state connectivity, shows a similar progression. HC exhibit a mean SWC of 0.6409, more than double that of coma (0.3072). UWS and MCS show slightly closer values (0.3681 and 0.3267, respectively), indicating some preservation of alpha connectivity in these groups, albeit significantly disrupted compared to healthy individuals. In the beta band, which supports higher-order cognitive functions, HC has the highest connectivity, with a mean SWC of 0.6591, compared to 0.3214 for coma, 0.3962 for UWS, and 0.3264 for MCS. In the gamma band, critical for neural synchronization and cognitive integration, HC (mean SWC = 0.6490) exhibit far more connectivity than coma (0.3445) and MCS (0.3395).

6.5.6 Variability in connectivity

Healthy controls exhibit moderate variability in most bands, with Std. SWC values consistently lower than mean SWC, reflecting an optimal balance between stability and flexibility. For instance, in the delta band, HC have a Std. SWC of 0.1984, significantly lower than their mean SWC of 0.6487, suggesting that low-frequency networks in healthy individuals are highly stable with low fluctuation. Coma and MCS groups exhibit slightly higher variability in the delta band (0.2526 and 0.2506, respectively), likely reflecting unstable network. UWS patients, with a Std. SWC of 0.2396, exhibit intermediate variability, consistent with their partial preservation of flexibility. In the theta band, HC show a Std. SWC of 0.2064, again reflecting a stable yet adaptable network. Comparatively, coma (0.2444) and MCS (0.2617) show higher variability, possibly indicating disorganized connectivity patterns. The UWS group exhibits variability comparable to HC (0.2467), suggesting that theta-band networks retain some flexibility in this group. The alpha band provides a clearer differentiation between groups. HC, with a Std. SWC of 0.2120, demonstrate the lowest variability, consistent with well-regulated alpha connectivity. In contrast, coma and MCS groups have significantly higher variability (0.2616 and 0.2560, respectively), indicating erratic alpha connectivity. UWS patients again occupy an intermediate position, with a Std. SWC of 0.2519, suggesting partial preservation of alpha flexibility. In the beta band, HC exhibit the lowest Std. SWC (0.1952), indicating that their high-frequency FC is both stable and efficient. Pathological groups show higher variability, with Std. SWC values of 0.2378 (coma), 0.2664 (UWS), and 0.2616 (MCS). The gamma band follows a similar trend, with HC showing a Std. SWC of 0.2266, while coma, UWS, and MCS have values of 0.2394, 0.2579, and 0.2541, respectively.

7 Discussion

The study analyses the classification ability of functional connectivity features, namely PLI, wPLI, and AEC, using various artificial neural network models. Across all feature sets, MLP model 2 consistently outperforms others in terms of accuracy, precision, recall, and F1-score. With the wPLI feature, MLP model 2 achieves an accuracy of 0.772. With the AEC feature, it delivers higher performance, achieving an accuracy of 0.963. When combining wPLI and AEC features, MLP model 2 gives a marginally higher accuracy of 0.969.

Whereas, PLI feature performs poorly across all models and metrics, consistently showing results inferior to other features. On MLP model 1, it achieves an accuracy of only 0.559, which is markedly lower than AEC (0.886) and wPLI (0.772). Its performance remains poor in recurrent models like LSTM and GRU, where it records accuracies of just 0.533 and 0.529, respectively. Similarly, its F1-scores are weak, hovering around 0.544. These results indicate that PLI is less effective in capturing connectivity patterns than wPLI and AEC in all evaluations.

These results highlight that MLP model 2 is the best-performing model for this classification task, with the AEC feature providing the best result across all performance metrics. This emphasizes the strength of amplitude-based connectivity metrics for distinguishing between consciousness states.

Table 12 presents a detailed comparison of our results with the techniques in the literature for distinguishing states of consciousness, including coma, UWS, MCS, and HC. Among existing methods for classifying disorders of consciousness, Naro et al. (2018) focused on phase-based connectivity and analyzed debiased weighted phase lag index (dwPLI) feature in 15 MCI and 17 UWS subjects. However, the absence of key performance metrics like accuracy and precision limits the comparability of their findings. Similarly, Di Gregorio et al. (2022) employed dominant frequency and phase coherence features with a linear discriminant analysis classifier to classify DOC subjects based on traumatic brain injury status. The study achieved accuracies of 80% and 83.3% for TBI and Non-TBI groups, with precision values of 0.777 and 0.857. While these results demonstrate the utility of simple classifiers, reliance on linear methods constrains their ability to capture complex relationships characteristic of brain waves.

Table 12
www.frontiersin.org

Table 12. Comparison of our performance with the techniques in the literature.

The study by Duclos et al. (2021) explored the application of FC measures like wPLI and AEC to assess anesthetic-induced unconsciousness and classify various altered consciousness states. SVM classifier was employed to analyze the anesthetic-induced unconsciousness in nine control subjects, reporting accuracies of 69.4% for wPLI, 83.7% for AEC, and 84.1% for their combination. However, the study included only nine subjects. Ensemble approaches, such as the CatBoost classifier used by Li et al. (2024a) and the bagged tree model in Raveendran et al. (2024), demonstrate improved adaptability but may lack the flexibility of ANN models when processing diverse and complex feature sets.

The inclusion of both accuracy and precision metrics strengthens the proposed method's evaluation. Accuracy measures the overall correctness of classification, while precision quantifies the model's ability to avoid false positives. This dual assessment addresses a notable gap in existing studies, where a sole focus on accuracy may obscure the clinical relevance of results. By providing a balanced evaluation, the proposed method ensures reliability and effectiveness, critical for clinical applications where misclassification can have significant consequences.

Our findings support integrating FC features, since combining wPLI and AEC features marginally improves the classification performance.

7.1 Limitations and future directions

While this study provides valuable insights, several limitations must be acknowledged. Firstly, the study relies on training and testing data from a fixed set of participants, which limits the generalizability of the findings. Despite demonstrating better performance metrics, the models' applicability to new datasets or broader population remains uncertain. Further, the analysis is restricted to three functional connectivity features: PLI, wPLI, and AEC. The study assumes that input features are standardized and preprocessed consistently across all states of consciousness. While this approach ensures comparability, it may overlook subtle variations in feature distributions across classes. To mitigate these limitations, future research will consider integrating multiple connectivity features to enhance classification performance.

7.1.1 Implications for clinical applications

The ability to accurately classify states of consciousness has significant clinical implications for the diagnosis and management of DOC. The study's findings suggest that MLP architectures using AEC as feature offer a reliable and practical solution for clinical applications. The high accuracy and balanced performance values achieved by these models indicate their potential for automated diagnostic systems. For applications focusing on phase-based features, GRU models leveraging wPLI features provide a complementary approach. By combining both approaches, a multimodal framework can be developed that integrates amplitude and phase-based connectivity measures to improve diagnostic accuracy and patient outcome.

8 Conclusion

This study demonstrates the importance of aligning feature selection and model design in classifying states of consciousness. AEC's superior predictive power highlights its utility as a standalone feature, while wPLI quantifies the phase relationship between channels. The findings emphasize the complementary nature of these features and the necessity of tailoring ANN architectures to their unique characteristics. By leveraging these insights, researchers can develop more effective tools for diagnosing and understanding disorders of consciousness, paving the way for improved patient care and deeper insights into the neural mechanisms underlying consciousness. Future work can explore other features, advanced architectures, and independent datasets to further refine these approaches and enhance their clinical utility.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by Institute Ethics Committee (Basic Sciences and Neurosciences), NIMHANS, Bangalore, India. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

SR: Conceptualization, Methodology, Software, Visualization, Writing – original draft, Writing – review & editing. KS: Conceptualization, Methodology, Supervision, Validation, Writing – review & editing. RG: Conceptualization, Methodology, Software, Supervision, Validation, Writing – review & editing. RK: Data curation, Supervision, Validation, Writing – review & editing. JS: Methodology, Software, Supervision, Validation, Writing – review & editing. SK: Supervision, Validation, Writing – review & editing. FK: Data curation, Validation, Writing – review & editing. RM: Data curation, Supervision, Validation, Writing – review & editing. SB: Data curation, Supervision, Validation, Writing – review & editing. BS: Data curation, Validation, Writing – review & editing. SR: Conceptualization, Data curation, Methodology, Supervision, Validation, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Acknowledgments

The authors sincerely thank the technicians at the EEG lab in NIMHANS for their invaluable assistance. They also express their deep gratitude to the consciousness-impaired subjects, their caretakers, and the healthy participants for their partaking in the study. This research has been partially supported by NIMHANS and also by IGSTC under Grant No. IGSTC/WISER 2023/KS/35/2023-24/765.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Anusha, A., Kumar, P., and Ramakrishnan, A. (2025). Brain-scale theta band functional connectome as signature of slow breathing and breath-hold phases. Comput. Biol. Med. 184:109435. doi: 10.1016/j.compbiomed.2024.109435

PubMed Abstract | Crossref Full Text | Google Scholar

Arbabyazd, L., Petkoski, S., Breakspear, M., Solodkin, A., Battaglia, D., and Jirsa, V. (2023). State-switching and high-order spatiotemporal organization of dynamic functional connectivity are disrupted by Alzheimer's disease. Netw. Neurosci. 7, 1420–1451. doi: 10.1101/2023.02.19.23285768

PubMed Abstract | Crossref Full Text | Google Scholar

Bruns, A., Eckhorn, R., Jokeit, H., and Ebner, A. (2000). Amplitude envelope correlation detects coupling among incoherent brain signals. Neuroreport 11, 1509–1514. doi: 10.1097/00001756-200005150-00029

PubMed Abstract | Crossref Full Text | Google Scholar

Cattarinussi, G., Di Giorgio, A., Moretti, F., Bondi, E., and Sambataro, F. (2023). Dynamic functional connectivity in schizophrenia and bipolar disorder: A review of the evidence and associations with psychopathological features. Prog. Neuro-Psychopharmacol. Biol. Psychiat. 127:110827. doi: 10.1016/j.pnpbp.2023.110827

PubMed Abstract | Crossref Full Text | Google Scholar

Dey, S., Anusha, A., and Ramakrishnan, A. (2023). “Differential effects of slow deep inhalation and exhalation on brain functional connectivity,” in 2023 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT) (Bangalore: IEEE), 1–5.

Google Scholar

Di Gregorio, F., La Porta, F., Petrone, V., Battaglia, S., Orlandi, S., Ippolito, G., et al. (2022). Accuracy of EEG biomarkers in the detection of clinical outcome in disorders of consciousness after severe acquired brain injury: preliminary results of a pilot study using a machine learning approach. Biomedicines 10:1897. doi: 10.3390/biomedicines10081897

PubMed Abstract | Crossref Full Text | Google Scholar

Duclos, C., Maschke, C., Mahdid, Y., Berkun, K., da Silva Castanheira, J., Tarnal, V., et al. (2021). Differential classification of states of consciousness using envelope-and phase-based functional connectivity. Neuroimage 237:118171. doi: 10.1016/j.neuroimage.2021.118171

PubMed Abstract | Crossref Full Text | Google Scholar

Fallani, F. D. V., and Bassett, D. S. (2019). Network neuroscience for optimizing brain-computer interfaces. Phys. Life Rev. 31, 304–309. doi: 10.1016/j.plrev.2018.10.001

PubMed Abstract | Crossref Full Text | Google Scholar

Hutchison, R. M., Womelsdorf, T., Allen, E. A., Bandettini, P. A., Calhoun, V. D., Corbetta, M., et al. (2013). Dynamic functional connectivity: promise, issues, and interpretations. Neuroimage 80, 360–378. doi: 10.1016/j.neuroimage.2013.05.079

PubMed Abstract | Crossref Full Text | Google Scholar

Ismail, L. E., and Karwowski, W. (2020). A graph theory-based modeling of functional brain connectivity based on EEG: a systematic review in the context of neuroergonomics. IEEE Access 8, 155103–155135. doi: 10.1109/ACCESS.2020.3018995

Crossref Full Text | Google Scholar

Jin, H., Ranasinghe, K. G., Prabhu, P., Dale, C., Gao, Y., Kudo, K., et al. (2023). Dynamic functional connectivity MEG features of Alzheimer's disease. Neuroimage 281:120358. doi: 10.1016/j.neuroimage.2023.120358

PubMed Abstract | Crossref Full Text | Google Scholar

Kumar, G. P., Dutta, U., Sharma, K., and Ganesan, R. A. (2022). “EEG-based biometrics: phase-locking value from gamma band performs well across heterogeneous datasets,” in 2022 International Conference of the Biometrics Special Interest Group (BIOSIG) (Darmstadt: IEEE), 1–6.

Google Scholar

Li, Y., Gao, J., Yang, Y., Zhuang, Y., Kang, Q., Li, X., et al. (2024a). Temporal and spatial variability of dynamic microstate brain network in disorders of consciousness. CNS Neurosci. Therapeut. 30:e14641. doi: 10.1111/cns.14641

PubMed Abstract | Crossref Full Text | Google Scholar

Li, Y., Ran, Y., Yao, M., and Chen, Q. (2024b). Altered static and dynamic functional connectivity of the default mode network across epilepsy subtypes in children: a resting-state fMRI study. Neurobiol. Dis. 192:106425. doi: 10.1016/j.nbd.2024.106425

PubMed Abstract | Crossref Full Text | Google Scholar

Mäki-Marttunen, V. (2014). Brain dynamic functional connectivity in patients with disorders of consciousness. BMC Neurosci. 15:P105. doi: 10.1186/1471-2202-15-S1-P105

Crossref Full Text | Google Scholar

Matsui, T., and Yamashita, K. (2023). Static and dynamic functional connectivity alterations in Alzheimer's disease and neuropsychiatric diseases. Brain Connect. 13, 307–314. doi: 10.1089/brain.2022.0044

PubMed Abstract | Crossref Full Text | Google Scholar

Miao, J., Tantawi, M., Alizadeh, M., Thalheimer, S., Vedaei, F., Romo, V., et al. (2023). Characteristic dynamic functional connectivity during sevoflurane-induced general anesthesia. Sci. Rep. 13:21014. doi: 10.1038/s41598-023-43832-1

PubMed Abstract | Crossref Full Text | Google Scholar

Mokhtari, F., Akhlaghi, M. I., Simpson, S. L., Wu, G., and Laurienti, P. J. (2019). Sliding window correlation analysis: modulating window shape for dynamic brain connectivity in resting state. Neuroimage 189, 655–666. doi: 10.1016/j.neuroimage.2019.02.001

PubMed Abstract | Crossref Full Text | Google Scholar

Naro, A., Bramanti, A., Leo, A., Cacciola, A., Manuli, A., Bramanti, P., et al. (2018). Shedding new light on disorders of consciousness diagnosis: the dynamic functional connectivity. Cortex 103:316–328. doi: 10.1016/j.cortex.2018.03.029

PubMed Abstract | Crossref Full Text | Google Scholar

Qin, L., Zhou, Q., Sun, Y., Pang, X., Chen, Z., and Zheng, J. (2024). Dynamic functional connectivity and gene expression correlates in temporal lobe epilepsy: insights from hidden Markov models. J. Transl. Med. 22:763. doi: 10.1186/s12967-024-05580-2

PubMed Abstract | Crossref Full Text | Google Scholar

Raveendran, S., Kenchaiah, R., Kumar, S., Sahoo, J., Farsana, M. K., Chowdary Mundlamuri, R., et al. (2024). Variational mode decomposition-based EEG analysis for the classification of disorders of consciousness. Front. Neurosci. 18:1340528. doi: 10.3389/fnins.2024.1340528

PubMed Abstract | Crossref Full Text | Google Scholar

Raveendran, S., Kumar, S., Kenchiah, R., Farsana, M., Choudary, R., Bansal, S., et al. (2023). “Scalp EEG-based classification of disorder of consciousness states using machine learning techniques,” in 2023 11th International Symposium on Electronic Systems Devices and Computing (ESDC) (Sri City: IEEE), 1–6.

Google Scholar

Saideepthi, P., Chowdhury, A., Gaur, P., and Pachori, R. B. (2023). Sliding window along with EEGNet-based prediction of EEG motor imagery. IEEE Sens. J. 23, 17703–17713. doi: 10.1109/JSEN.2023.3270281

Crossref Full Text | Google Scholar

Sastry, N. C., and Banerjee, A. (2024). Dynamicity of brain network organization & their community architecture as characterizing features for classification of common mental disorders from whole-brain connectome. Transl. Psychiatry 14:268. doi: 10.1038/s41398-024-02929-5

PubMed Abstract | Crossref Full Text | Google Scholar

Soares, C., Lima, G., Pais, M. L., Teixeira, M., Cabral, C., and Castelo-Branco, M. (2024). Increased functional connectivity between brain regions involved in social cognition, emotion and affective-value in psychedelic states induced by n, n-dimethyltryptamine (DMT). Front. Pharmacol. 15:1454628. doi: 10.3389/fphar.2024.1454628

PubMed Abstract | Crossref Full Text | Google Scholar

Stam, C. J., Nolte, G., and Daffertshofer, A. (2007). Phase lag index: Assessment of functional connectivity from multi channel EEG and MEG with diminished bias from common sources. Hum. Brain Mapp. 28, 1178–1193. doi: 10.1002/hbm.20346

PubMed Abstract | Crossref Full Text | Google Scholar

Tang, T. B., Chong, J. S., Kiguchi, M., Funane, T., and Lu, C.-K. (2021). Detection of emotional sensitivity using fNIRS based dynamic functional connectivity. IEEE Trans. Neural Syst. Rehabilitat. Eng. 29, 894–904. doi: 10.1109/TNSRE.2021.3078460

PubMed Abstract | Crossref Full Text | Google Scholar

Vinck, M., Oostenveld, R., Van Wingerden, M., Battaglia, F., and Pennartz, C. M. (2011). An improved index of phase-synchronization for electrophysiological data in the presence of volume-conduction, noise and sample-size bias. Neuroimage 55, 1548–1565. doi: 10.1016/j.neuroimage.2011.01.055

PubMed Abstract | Crossref Full Text | Google Scholar

Wu, K., Jelfs, B., Neville, K., Cai, A., and Fang, Q. (2024). fMRI-based static and dynamic functional connectivity analysis for post-stroke motor dysfunction patient: a review. IEEE Access 12, 133067–133085. doi: 10.1109/ACCESS.2024.3445580

Crossref Full Text | Google Scholar

Xu, Z., Chang, Y., Guo, F., Wang, C., Chai, N., Zheng, M., et al. (2024). The restoration ability of a short nap after sleep deprivation on the brain cognitive function: a dynamic functional connectivity analysis. CNS Neurosci. Therapeut. 30:e14413. doi: 10.1111/cns.14413

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: functional connectivity, sliding window correlation, amplitude envelope correlation, weighted phase lag index, disorders of consciousness, artificial neural networks, LSTM, hybrid CNN-LSTM

Citation: Raveendran S, S K, A G R, Kenchaiah R, Sahoo J, Kumar S, M K F, Mundlamuri RC, Bansal S, V S B and R S (2025) Functional connectivity in EEG: a multiclass classification approach for disorders of consciousness. Front. Neurosci. 19:1550581. doi: 10.3389/fnins.2025.1550581

Received: 23 December 2024; Accepted: 12 March 2025;
Published: 28 March 2025.

Edited by:

Jürgen Dammers, Helmholtz Association of German Research Centres (HZ), Germany

Reviewed by:

David Purger, Stanford University, United States
Chunyun Zhang, Qilu Hospital of Shandong University (Qingdao), China
Zoran Šverko, University of Rijeka, Croatia

Copyright © 2025 Raveendran, S, A G, Kenchaiah, Sahoo, Kumar, M K, Mundlamuri, Bansal, V S and R. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Ramakrishnan Subasree, c3ViYXNyZWUucmFtYWtyaXNobmFuQGdtYWlsLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Research integrity at Frontiers

Man ultramarathon runner in the mountains he trains at sunset

95% of researchers rate our articles as excellent or good

Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


Find out more