
94% of researchers rate our articles as excellent or good
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.
Find out more
ORIGINAL RESEARCH article
Front. Neurosci. , 05 February 2025
Sec. Brain Imaging Methods
Volume 19 - 2025 | https://doi.org/10.3389/fnins.2025.1543508
Introduction: A brain-computer interface (BCI) is an emerging technology that aims to establish a direct communication pathway between the human brain and external devices. Motor imagery electroencephalography (MI-EEG) signals are analyzed to infer users’ intentions during motor imagery. These signals hold potential for applications in rehabilitation training and device control. However, the classification accuracy of MI-EEG signals remains a key challenge for the development of BCI technology.
Methods: This paper proposes a composite improved attention convolutional network (CIACNet) for MI-EEG signals classification. CIACNet utilizes a dual-branch convolutional neural network (CNN) to extract rich temporal features, an improved convolutional block attention module (CBAM) to enhance feature extraction, temporal convolutional network (TCN) to capture advanced temporal features, and multi-level feature concatenation for more comprehensive feature representation.
Results: The CIACNet model performs well on both the BCI IV-2a and BCI IV-2b datasets, achieving accuracies of 85.15 and 90.05%, respectively, with a kappa score of 0.80 on both datasets. These results indicate that the CIACNet model’s classification performance exceeds that of four other comparative models.
Conclusion: Experimental results demonstrate that the proposed CIACNet model has strong classification capabilities and low time cost. Removing one or more blocks results in a decline in the overall performance of the model, indicating that each block within the model makes a significant contribution to its overall effectiveness. These results demonstrate the ability of the CIACNet model to reduce time costs and improve performance in motor imagery brain-computer interface (MI-BCI) systems, while also highlighting its practical applicability.
Brain-computer interface (BCI) has ushered in a new era of human-technology interaction. It establish a direct communication bridge between brain neural signals and external devices (Wolpaw et al., 2002). BCI systems can accurately measure and interpret brain activity, translating an individual’s intentions, thoughts, or perceptions into executable control signals. These signals can operate various external devices, including wheelchairs, prostheses, computer cursors, and advanced robotic systems. This allows users to perform specific tasks (Do et al., 2013; Ramadan and Vasilakos, 2017; Vilela and Hochberg, 2020).
Electroencephalography (EEG) is a non-invasive technique for capturing electrical signals generated by neuronal activity in the brain. These signals are detected by electrodes placed on the scalp. The electrodes can detect small current fluctuations generated by neuronal populations in the cerebral cortex. Due to its high temporal resolution and portability, EEG is not only an indispensable tool in clinical medicine but also widely used in engineering and psychology (Biasiucci et al., 2019).
Motor imagery (MI) refers to the mental rehearsal of movements without actual physical execution (Pfurtscheller and Neuper, 2001). This process includes recalling past actions and imagining future ones. Monitoring sensory motor rhythms (SMR) shows that MI induces event-related synchronization (ERS) and event-related desynchronization (ERD). This marks it as an actively evoked EEG signal (Grandchamp and Delorme, 2011). Motor imagery electroencephalography (MI-EEG) signals are widely used in rehabilitation medicine to support the recovery of compromised motor functions. A key advantage of these signals is that they can autonomously activate motor-related brain regions without relying on external stimuli. Motor imagery brain-computer interface (MI-BCI) systems have been applied across a range of medical and non-medical fields. In medicine, MI-BCI applications include stroke rehabilitation, prosthetic control, wheelchair navigation, psychological therapy, and cognitive training (Khan et al., 2020). Beyond medical applications, MI-BCI systems are used in vehicle and drone control, gaming, skill development, and virtual reality.
Despite the significant potential of MI-BCI across various domains, challenges persist in accurately interpreting users’ intentions. EEG patterns generated during MI can vary substantially between subjects, increasing the time and complexity required for system customization. This limits the universal applicability of MI-BCI technology. The non-stationarity of MI-EEG signals is a primary obstacle in MI-BCI development. Variations in user fatigue and attention levels further exacerbate signal non-stationarity (Abdulkader et al., 2015). Motion artifacts, such as electromyographic activity, eye movements, and blinking-related electrooculogram (EOG) signals, can significantly degrade EEG signal quality (Rashmi and Shantala, 2022). This ultimately impacts the performance of MI-BCI systems. Consequently, inter-individual variability and signal non-stationarity present substantial challenges for the practical implementation of MI-BCI. These challenges make it important to classify MI-EEG signals accurately.
Researchers have proposed various methods for capturing MI-EEG signals features and classifying MI tasks. Conventional feature extraction methods typically involve fewer parameters and have lower computational complexity. These methods include common spatial patterns (CSP) (Ramoser et al., 2000), fourier transform (FT) (Wang P. et al., 2018), and wavelet transform (WT) (Xu et al., 2018). CSP is a typical feature extraction algorithm. It is capable of extracting the spatial distribution components of each class from multi-channel EEG signals. It is suitable for processing small datasets and EEG signals with minimal motion artifacts. In MI-EEG signal classification, support vector machine (SVM) (Chatterjee and Bandyopadhyay, 2016) is the most representative method. SVM perform well in classification tasks with small sample sizes and well-defined features. However, its performance drops when handling large datasets. Ang et al. (2012) proposed filter bank common spatial pattern (FBCSP). This method combines a set of band-pass filters with CSP. The accuracy obtained by combining this method with SVM shows a slight improvement over CSP, but due to the high noise level in EEG signals, the results remain unsatisfactory. Traditional approaches to MI-EEG feature extraction have limited efficacy in non-linear models, especially in complex or multi-class classification scenarios. Their performance is significantly reduced in these cases.
In recent years, the use of deep learning (DL) techniques for classifying MI-EEG signals has increased rapidly. DL approaches can automatically extract features from data and possess strong nonlinear fitting capabilities. They introduce nonlinear factors through activation functions, making them effective in handling the complex nonlinear characteristics of EEG signals. This has led to significant improvements in the performance of MI-EEG signal classification. In recent years, many DL architectures have been introduced for MI task classification, including convolutional neural network (CNN), recurrent neural network (RNN), long short-term memory (LSTM), and autoencoder. Additionally, various CNN modifications have been proposed, such as multi-branch CNNs, multi-scale CNNs, and Residual-based CNNs. Lawhern et al. (2018) proposed EEGNet, a compact CNN architecture designed for EEG data, which has been effectively applied in various BCI paradigms. Its lightweight design makes EEGNet particularly suitable for EEG analysis. Chen W. et al. (2024) proposed the EEGNeX model, which replaces the two-dimensional (2D) convolutions and separable convolutions in EEGNet with a pair of 2D convolutions. Luo et al. (2018) developed an RNN model that incorporates a sliding window cropping strategy (SWCS) for classifying MI-EEG signals. Wang Z. et al. (2018) proposed a classification framework that utilizes LSTM and uses a one-dimensional aggregation approximation (1d-AX) technique to extract effective signal representations. In the field of algorithmic integration, Lu et al. (2019) combined one-dimensional CNN with LSTM for classifying MI tasks. Later, Gao et al. (2022) integrated gated recurrent units (GRU) with CNN to create a parallel feature fusion network. Overall, CNN models have found broader application in MI task classification than other DL models. They have also proven to be compatible with integration into other DL models.
Temporal convolutional network (TCN) has been increasingly applied in various fields (Bai et al., 2018). TCN is particularly useful in time series forecasting and sequential annotation tasks. It effectively extracts both high-frequency and low-frequency information from sequences. TCN is a specialized one-dimensional CNN composed of three main components: causal convolution, dilated convolution, and residual blocks. Causal convolution ensures strict temporal constraints, while dilated convolution enhances the network’s ability to process long sequences (Bi et al., 2021). Unlike conventional RNN, TCN uses convolutional operations to capture temporal interdependencies. This effectively expands the receptive field and supports long-range dependencies (Hewage et al., 2020). Recent research has applied the TCN framework to classify MI-EEG signals. Ingolfsson et al. (2020) proposed the EEG-TCNet model. It combines the EEGNet and TCN architectures. Subsequently, Musallam et al. (2021) proposed an enhanced version of EEG-TCNet through feature concatenation, named TCNet-Fusion. Salami et al. (2022) proposed the EEG-ITNet model. It is a tri-branch structure combining CNN and TCN to extract rich temporal and spatial information.
Attention is the cognitive ability to selectively focus on specific objects. Drawing inspiration from this capacity, researchers have introduced the attention mechanism. In deep learning, the attention mechanism dynamically adjusts the model’s focus across different segments of input data. This enhances the model’s performance and efficiency by emphasizing more significant information. In 2017, Google researchers introduced the Transformer architecture, a novel neural network framework based entirely on attention mechanisms. Later, Hu et al. (2018) proposed the squeeze-and-excitation (SE) model. It focuses on inter-channel relationships and enables the model to autonomously determine the significance of different channel features. In the same year, Woo et al. (2018) proposed the convolutional block attention module (CBAM), an attention mechanism that retains conventional channel attention and adds a spatial attention mechanism. It enhances network performance across both channel and spatial domains. Recently, the scientific community has increasingly adopted DL models with attention mechanisms for classifying MI-EEG signals. Song et al. (2022) combined CNN with self-attention to classify tasks involving MI and emotion recognition. Altaheri et al. (2022) proposed ATCNet, a model that integrates multi-head self-attention (MSA), TCN, and CNN to decode MI-EEG signals. Hu et al. (2023) proposed MSATNet, a model that combines a dual-branch CNN and Transformer to classify MI-EEG signals. Xie et al. (2024) proposed the BFATCNet model, designing a tri-branch structure with attention mechanisms and TCN. Similarly, Deng et al. (2024) developed a tri-branch architecture that combines parallel multi-head attention with SE and CBAM alongside EEGNet and TCN for MI-EEG signal classification.
This study proposes a novel composite improved attention convolutional network, CIACNet, for MI-EEG signals classification. The proposed CIACNet model utilizes a dual-branch convolutional architecture, combining an improved CBAM with TCN for feature extraction from MI-EEG signals. The structure of this model differs from the repetitive multi-branch structures used by others, as it employs multi-level feature concatenation, allowing the model to consider the features obtained from each block component. This lightweight model removes the requirement for human intervention in the decoding process. Its applicability within MI-BCI systems is highlighted by its automated decoding, high accuracy, and reduced training costs for subjects. The steps for decoding within the proposed CIACNet model are as follows: First, conduct simple preprocessing. Second, extract rich temporal features using a dual-branch CNN with distinct parameter sets. Third, apply the improved CBAM to enhance the model’s ability to extract key information. Fourth, capture high-level temporal features using the TCN. Finally, in the fully connected (FC) layer, we concatenate the features from multiple branches and perform classification using the softmax function. This research highlights the following contributions:
1. We have proposed the high-performance CIACNet model, which combines CNN, improved CBAM, and TCN to form a composite deep learning architecture. The model has demonstrated exceptional performance on the BCI Competition IV-2a and BCI Competition IV-2b datasets.
2. The dual-branch CNN architecture effectively extracts features from MI-EEG signals at multiple scales, thereby enhancing the comprehensiveness of the feature learning process.
3. The improved CBAM further enhances the model’s ability to extract key information by modifying the pooling approach of the standard CBAM. This attention mechanism effectively identifies the importance of different regions within the input feature map across both channel and spatial dimensions.
4. In the FC layer, concatenating features from different levels enhances the model’s ability to represent complex features. This improves the overall classification accuracy.
The proposed CIACNet model consists of three core components: the convolutional (CV) block, the improved CBAM attention (IAT) block, and the temporal convolution (TC) block, as shown in Figure 1. The CV block includes two subblocks, CV1 and CV2. Each subblock comprises three distinct convolutional layers: temporal, channel depth-wise, and spatial convolutions. This block is responsible for encoding the spatiotemporal features of MI-EEG signals. The CV block captures temporal features across different time steps and outputs a comprehensive representation of the signals. The IAT block is equipped with an improved CBAM. It effectively captures features from both channel and spatial perspectives. The TC block utilizes TCN to extract high-level temporal features from sequential data. In the final stage of the model, the outputs from the IAT, TC, and CV2 blocks are concatenated. The resulting features are then passed to the FC layer with a softmax classifier. Further details about the CIACNet model will be provided in the subsequent sections.
In the proposed CIACNet model, the raw MI-EEG signals undergo a simple preprocessing phase. During this phase, to preserve the original characteristics of EEG signals and reduce the latency of the decoding system, we utilize the full spectrum of EEG frequencies and all channels without removing any artifacts. A single trial dataset of MI-EEG can be denoted as , where the corresponding class label is . , , and represent the channel count, the number of sampling points, and the class count, respectively. To address the non-stationarity of the signals, we apply Z-score normalization, as shown in Equation 1.
Where and denote the input and normalized dataset, respectively. and represent the mean and variance of the training dataset.
The normalization process ensures that each channel in the raw EEG signals has a mean of zero and a standard deviation of one. MI-EEG datasets usually contain multiple trials conducted by different subjects. For a given subject, the MI-EEG data is represented as a collection , where and indicate the n-th subject and the i-th trial, respectively. The goal of the CIACNet model is to decode each input trial , determine its predicted category, and compare it to the actual class label .
Drawing inspiration from the EEGNet model, we implemented a series of improvements and introduced the CV1 and CV2 blocks. EEGNet is a compact CNN that demonstrates strong classification capabilities. Its simple architecture limits its effectiveness in processing noisy data. To improve the comprehensiveness of MI-EEG signal decoding, this study incorporates the enhanced CV1 and CV2 blocks as branches within the CV block. The CV block adopts a multi-branch, multi-scale convolutional strategy to enrich the extracted features of the EEG signals.
The CV1 and CV2 blocks serve as dual branches for feature extraction within the model, each receiving the same input. They follow EEGNet’s structure but with different parameter settings. While the original EEGNet employs 2D, depthwise, and separable convolutional layers, we have replaced its separable convolutions with 2D convolutions. This adjustment enhances the model’s ability to learn features and enables it to capture more complex spatial and temporal feature relationships.
The CV block consists of three convolutional layers, as shown in Figure 2. The data in the figure is based on the BCI IV-2a dataset. The first layer uses 2D convolution, where the CV1 and CV2 blocks apply and filters, respectively, for temporal convolution. These filters have dimensions of and based on and . The CV2 block uses larger filters to cover longer time windows, and a greater number of filters to extract more diverse features. After the first layer, EEG signals produce temporal feature maps. The second layer uses depthwise convolution with and filters for CV1 and CV2, respectively, each with size , where represents the number of MI-EEG channels. Based on empirical findings, is set to 2. Following depthwise convolution, an average pooling layer with dimensions (1, 8) reduces the sampling frequency and temporal sequence length to 32 Hz and 140, respectively. The third layer applies 2D convolution similar to the first layer. In this layer, the CV1 and CV2 blocks use filters of size , with and filters, respectively. This is followed by an average pooling layer with size , which reduces the sampling frequency and temporal sequence length to 4 Hz and 17, respectively. After each convolutional layer in the CV block, batch normalization (BN) is applied to standardize the input and accelerate the training process. Additionally, ridge regularization is incorporated into each convolutional layer to control the weight magnitude. Following the second and third convolutional layers, exponential linear unit (ELU) activation functions are utilized to provide non-linear mapping, enhancing network performance. Dropout layers are also included to mitigate overfitting by randomly deselecting neurons. The CV block uses varying filter sizes and quantities across its dual branches. This approach effectively extracts rich features from different scales. As a result, it significantly improves classification accuracy.
Figure 2. The basic structure of the CV block. The preprocessed MI-EEG data is encoded by the CV block. This encoding outputs a sequence with TC elements. Each element is a vector of size F2 or F4.
Inspired by the human ability to selectively focus on specific information, attention mechanisms in deep learning have proven effective at decoding MI-EEG signals. CBAM integrates both channel and spatial attention. It is a highly effective attention mechanism (Altuwaijri and Muhammad, 2022). CBAM allows the model to learn the significance of different regions within the input feature map across both channel and spatial dimensions. Notably, CBAM is efficient and lightweight. It imposes minimal additional computational burden on the CNN architecture. The channel attention component of CBAM applies global average and max pooling to the input feature map, generating distinct channel-wise representations. The spatial attention component performs average and max pooling for each channel in the feature map, producing two separate channel feature maps. These maps are then concatenated to form a comprehensive spatial attention map.
Max pooling focuses solely on the largest element and disregards the others within the pooling region. This helps preserve the most significant features but may overlook some valuable information. Average pooling computes the mean value of all elements within the pooling region, which makes it well-suited for output smoothing. In this work, we improve CBAM by incorporating stochastic pooling (Zeiler and Fergus, 2013). This method assigns probabilities to feature map elements based on their numerical values. The likelihood of an element’s selection is proportional to its magnitude. As a result, the pooled features neither overemphasize local maxima nor are dominated by mean smoothing effects. By combining these three pooling strategies, the model is able to extract features more effectively. It can also assess the significance of various regions within the input feature map. This ultimately improves the performance of the attention block.
Instead of using deterministic selection methods, stochastic pooling adopts a probabilistic approach to select elements. It randomly selects elements based on a probability distribution derived from the activations within the pooling region. The likelihood of selecting a given position is proportional to the normalized activation values. This reflects their numerical magnitude, as shown in Equation 2.
Where is the probability of each region , denotes an element within the pooling region, and refers to the pooling region.
The location of the region is sampled from a multinomial distribution according to the probabilities. Stochastic pooling is defined as Equation 3.
Where is the output of the pooling operation associated with the j-th feature map, and denotes the activation value of the pooling.
The specific process of stochastic pooling is as follows: The elements in the pooling region are normalized to obtain a probability matrix. Regions are then randomly selected based on these probabilities. The pooled value is the value of the selected region location.
The improved CBAM attention block proposed in this paper is shown in Figure 3. The block receives the output feature map from the CV1 block. , , and denote the channel count, height, and width of the feature map, respectively. This map serves as the input for the IAT block. First, the feature map passes through the channel attention block and produces a one-dimensional channel attention map . Second, is element-wise multiplied with to emphasize important channels. This produces a modified feature map . Third, enters the spatial attention block. It generates a two-dimensional spatial attention map. . Finally, is element-wise multiplied with . This applies spatial attention weights to enhance spatial regions of the feature map. The resulting output is denoted as .
The channel attention block within the IAT block integrates global information from the feature map. This provides a detailed representation of each channel, as shown in the upper half of Figure 4. First, three pooling strategies, average, max, and stochastic pooling, are applied to the input feature map . This broadens the spectrum of captured information. The resulting vectors are then reshaped to dimensions. Second, these three channel description vectors are input into a multilayer perceptron (MLP) consisting of two FC layers. The MLP generates the attention weights for each channel. Rectified linear unit (ReLU) activation is incorporated within the MLP to enhance non-linearity. To reduce computational cost, the hidden layer’s size is set to , with as the reduction ratio. Finally, the aligned positions of the three feature vectors are summed, followed by the application of a sigmoid function to produce the channel attention map . The channel attention calculation is shown in Equation 4.
Where is sigmoid activation function, while , , and denote the features derived from the channel attention block after applying average, max, and stochastic pooling, respectively. The MLP weights are represented by and , with a ReLU activation following . , and sequentially assign weights to the pooled features from each of the three pooling operations.
The spatial attention block in the IAT block focuses on the spatial aspects of the input feature map. It highlights important spatial locations and reduces less critical ones, as shown in the lower half of Figure 4. First, average, max, and stochastic pooling are applied across each channel of the feature map . Second, the resulting feature maps have identical dimensions. They are then concatenated and passed through a 2D convolution layer with a 7 × 7 kernel. Finally, a sigmoid function generates the spatial attention feature map . The computation for spatial attention is shown in Equation 5.
Where is the convolutional operation performed with a 7 × 7 kernel, while , , and denote the features derived from the spatial attention block after applying average, max, and stochastic pooling, respectively.
The TCN is particularly suited for processing sequential data. The TC block in this study follows the TCN architecture and consists of two residual blocks. Each residual block contains two dilated causal convolutional layers. After each convolutional layer, BN and ELU activation functions are applied. The architecture of the TC block is shown in Figure 5.
The TC block utilizes causal convolution. This ensures that model predictions at each step depend solely on prior inputs and are unaffected by future data. Additionally, it incorporates dilated convolution by introducing gaps in the convolution process. This allows the kernel to capture a broader temporal range. In residual blocks, residual connections alleviate vanishing and exploding gradient issues by summing input and output feature maps element-wise. With dilated convolution, the receptive field size (RFS) in the TC block increases exponentially with the dilation factor, as shown in Equation 6.
Where is the kernel size, and denotes the number of residual blocks.
Within the TC block, parameters and are applied, along with 32 filters. Given these settings, the RFS is 19. This allows the TCN to process sequences up to a maximum length of 19. Additionally, a parallel branch adjacent to the TC block outputs features from the IAT block. This setup enhances feature diversity by combining original features with those obtained after temporal convolution. Finally, the outputs from these two branches are concatenated with the CV2 block’s output. These are then fed into the FC layer with a softmax classifier. The hyperparameters used in the experiments are consistent across all subjects, as shown in Table 1.
The BCI IV-2a dataset (Brunner et al., 2008) serves as the benchmark for evaluating our proposed model. This publicly available MI-EEG dataset developed by Graz University of Technology is known. It contains a large number of artifacts, which increases the complexity of decoding MI tasks. It includes EEG recordings from nine healthy subjects, recorded with 22 electrodes following the standard 10–20 system at a sampling rate of 250 Hz. Signals are band-pass filtered between 0.5 and 100 Hz, with a 50 Hz notch filter applied to remove line noise. Each participant performs 576 trials of MI tasks, each lasting 4 s. Of these trials, 288 are used for training and the remaining 288 are reserved for evaluation. The MI tasks consist of four types: left-hand movement, right-hand movement, both feet movement, and tongue movement.
The BCI IV-2b dataset (Leeb et al., 2008) is also used to evaluate our proposed model. This dataset consists of EEG recordings from nine healthy subjects, captured using three electrodes at a 250 Hz sampling rate. Signals are band-pass filtered between 0.5 and 100 Hz, with a 50 Hz notch filter applied to remove line noise. Each participant completes five sessions: the first two without feedback, and the remaining three with smiley face feedback. Each session includes 120 trials of MI tasks, with each tasks lasting 4 s. The MI tasks consist of two types: left-hand movement and right-hand movement. Participants perform these tasks only through mental imagery, without any physical movement.
In this study, we use accuracy and kappa score to evaluate the performance of our proposed model. Accuracy provides a straightforward and intuitive measure, especially suitable for datasets with balanced class distributions. Kappa score accounts for biases due to imbalanced class distributions. It offers a more robust and impartial assessment of classifier performance compared to chance-level predictions. The equation for calculating accuracy (Equation 7) and kappa scores (Equation 8) are presented below:
Where is the true positives. It represents the number of samples correctly predicted within category . denotes the total number of samples in category . refers to the total number of categories.
Where is the actual percentage of agreement, denotes the expected percentage of agreement, and refers to the total number of classes.
Our proposed CIACNet model is compared with several CNN-based models. A description of each model is provided below:
1. EEGNet (Lawhern et al., 2018): A compact CNN model comprising a 2D convolutional layer, a depthwise convolutional layer, and a separable convolutional layer.
2. TCNet-Fusion (Musallam et al., 2021): This multi-branch model combines CNN and TCN architectures. The first block is based on EEGNet, while the second block splits the output into three branches, two remain unchanged, and the third connects to a TCN module. The third block concatenates the outputs from all branches, followed by classification with the softmax function.
3. ATCNet (Altaheri et al., 2022): Integrates attention mechanisms with TCN and CNN architectures. The attention block applies multi-head self-attention to identify key features. The TC block utilizes TCN to extract advanced temporal features.
4. TBTSCTnet (Chen X. et al., 2024): A multi-branch model incorporating Transformer with three branches that apply temporal and spatial convolutional filters of varying sizes. A Transformer encoder captures global dependencies within the spatiotemporal features extracted by the convolutional layers.
Experiments involving the proposed CIACNet model and the comparison models are conducted on an NVIDIA GTX 1650 8G, using TensorFlow 2.7 framework. Uniform training settings are applied across all models in this study. All models were trained using the Adam optimizer with a learning rate of 0.001 and a batch size of 64. Categorical cross-entropy loss was employed over 1,000 epochs. Model weights are saved only when there is an improvement in accuracy, which ensures the retention of the best-performing models. All models undergo validation based on supervised validation loss, with a reduced learning rate applied if validation loss stabilizes over consecutive epochs. To prevent overfitting, early stopping is used with the patience parameter set to 300 epochs. Each subject was trained 10 times. The best-performing trial among them was selected as the final result.
Tables 2, 3 offer a detailed overview of classification accuracy and kappa scores for CIACNet model and four other CNN-based models on the BCI IV-2a and IV-2b datasets. They also present the average values and standard deviations of the model parameters. The CIACNet model consistently outperforms reproduced EEGNet, TCNet-Fusion, ATCNet, and TBTSCTNet models. It demonstrates the best performance on both datasets. On the BCI IV-2a dataset, CIACNet model achieves an average accuracy of 85.15% and a kappa score of 0.80. It exceeds the advanced ATCNet model by 1.97% and outperforms the other models by at least 3.52%. With an accuracy standard deviation of 8.41%, CIACNet model also demonstrates the highest stability in accuracy across all subjects. Additionally, on the BCI IV-2b dataset, CIACNet model achieves an average accuracy exceeding 90%. This demonstrates its significant advantage in classifying MI tasks for left and right hand movements. To show the role of our model in reducing training time costs, we add the average training time (in minutes) for each model when classifying all subjects in the last row of Tables 2, 3. On the BCI IV-2a dataset, EEGNet and TCNet-Fusion model require the shortest training times, but they also achieve the lowest classification accuracy. Compared with the remaining two models, our CIACNet achieves the shortest training time, averaging only 12.01 min. A similar conclusion is drawn for the BCI IV-2b dataset, where our model shows significantly shorter training times than the ATCNet model.
Table 2. Comparison of classification accuracy (%) and kappa between the proposed model and other reproduced models on the BCI IV-2a dataset.
Table 3. Comparison of classification accuracy (%) and kappa between the proposed model and other reproduced models on the BCI IV-2b dataset.
Boxplots are used to visually assess the performance of the CIACNet model and other models. They show the distribution of accuracy across each model. Figure 6 displays the boxplot for the classification results of the five models on the BCI IV-2a dataset. In the boxplot, the x-axis represents the model names and the y-axis shows accuracy. The horizontal lines within each boxplot indicate the median accuracy, while the whiskers denote the minimum and maximum values across all subjects. The edges of each box correspond to the upper and lower quartiles. The figure clearly shows that the CIACNet model exceeds the other models in terms of median accuracy, as well as in maximum, minimum, and quartile values, across all subjects.
Figure 7 displays the confusion matrices for the CIACNet model’s classification results on the BCI IV-2a dataset. These figures show the model’s strong performance across all subjects. In Figure 7, subjects S1, S3, and S7 achieve classification accuracies exceeding 90%. Figure 8 shows the mean confusion matrices across all subjects for BCI IV-2a and BCI IV-2b datasets as classified by the CIACNet model. The left panel of Figure 8 shows average accuracies of 86, 85, 87, and 84% for the left hand, right hand, both feet, and tongue tasks, respectively. The right panel shows average accuracies of 89% for the left hand and 91% for the right hand. These results suggest that the CIACNet model consistently achieves balanced and high classification accuracy on both datasets. This demonstrates its stability and precision in decoding MI-EEG signals. It also highlights the model’s potential for real-world applications.
Figure 8. The subject average confusion matrix of the CIACNet model on the BCI IV-2a and BCI IV-2b datasets.
Table 4 presents the classification results of the CIACNet model on the BCI IV-2a and BCI IV-2b datasets. It also provides a comparative analysis with other state-of-the-art (SOTA) models. The table details the main methodologies used by these models, which include multi-branch multi-scale and attention mechanisms. The CIACNet model outperforms these SOTA models in terms of both accuracy and kappa scores for the four-class and two-class classification tasks. This indicates the high performance of our model in classification tasks.
Table 4. Classification performance of different models based on the BCI IV-2a and BCI IV-2b datasets.
We conduct ablation studies to evaluate the contribution of each block within the CIACNet model. Table 5 shows the effects of removing one or more blocks on the CIACNet model’s performance. It presents the average accuracy and kappa scores for classification tasks on both the BCI IV-2a and BCI IV-2b datasets. For the BCI IV-2a dataset, the CV2 and IAT blocks improve average accuracy by 2.52 and 2.86%, respectively. Removing the TC block leads to a 3.21% decrease in average accuracy. This suggests that CBAM may not be suitable when positioned only in the final CNN layer. For the BCI IV-2b dataset, removing the TC block causes a more significant reduction in accuracy than removing the CV2, IAT, and TC blocks together. This highlights the importance of the TC block for this relatively simpler dataset. The CIACNet model utilizes a hybrid approach, integrating CBAM elements into both intermediate and final layers. This enhances its ability to focus attention across different feature levels. The ablation study in Table 5 shows that removing one or more blocks results in a decline in the overall performance of the CIACNet model. Therefore, each block within the model makes a significant contribution to its overall effectiveness.
Table 5. The contribution of each block in CIACNet model was evaluated using the BCI IV-2a and BCI IV-2b datasets.
Moving forward, we assess the impact of modifying the attention mechanisms within the IAT block on the CIACNet model’s performance. Table 6 presents the accuracy and kappa scores after substituting the improved CBAM in the IAT block with standard CBAM, SE, and MSA, respectively. The first row in the table represents the original setup, which uses the improved CBAM. Results in Table 6 indicate that the improved CBAM achieves the highest accuracy and kappa scores. Across both datasets, the accuracy with the improved CBAM exceeds other attention mechanisms by at least 0.77 and 0.35%, respectively. Additionally, we compare the performance of the improved CBAM and the standard CBAM on EEG data with high noise levels. On the BCI IV-2a dataset, we focus on high-noise subjects 2 and 6. Using the model with the improved CBAM, the accuracies for subjects 2 and 6 are 75.69 and 70.49%, respectively. In contrast, the model with the standard CBAM achieves accuracies of 73.96 and 68.40% for these subjects. The comparison results indicate that the improved CBAM outperforms the standard CBAM in handling high-noise EEG data. Finally, as shown in Table 6, the improved CBAM shows its advantages in classification tasks for both large datasets (BCI IV-2a) and small datasets (BCI IV-2b). Thus, our proposed IAT block significantly improves the model’s ability to extract key features, thereby enhancing its decoding performance.
Figure 9 presents a visualization analysis of the CIACNet model’s classification results on the BCI IV-2a and BCI IV-2b datasets. The analysis uses t-distributed stochastic neighbor embedding (t-SNE). The t-SNE technique projects high-dimensional samples into a low-dimensional space, preserving the relative proximity of similar samples. All feature map dimensions have been reduced to two dimensions using t-SNE for this analysis. In Figure 9, points in four distinct colors represent classification results for the left hand, right hand, both feet, and tongue tasks. The x and y axes correspond to these two projected dimensions. The analysis includes subject 3 from the BCI IV-2a dataset and subject 4 from the BCI IV-2b dataset exhibiting best performance. For subject 3, classification accuracy is 98.6% (71/72) for the left hand and 95.8% (69/72) for the right hand, both feet, and tongue tasks. Subject 4 achieves 98.6% (158/160) accuracy for both left and right hand tasks. These results demonstrate the model’s high classification proficiency and its capability to delineate clear cluster boundaries.
Figure 9. t-SNE visualization of subject 3 from the BCI IV-2a dataset and subject 4 from the BCI IV-2b dataset.
In this study, we propose a CIACNet model that integrates dual-branch CNN, improved CBAM, TCN, and multi-level feature concatenation to address the challenges in motor imagery EEG classification.
CIACNet is a lightweight EEG decoding model that requires only minimal preprocessing, limited to standardizing the input data. The CV block uses two branches to process MI-EEG signals in parallel. This allows extracting rich temporal features. The IAT block utilizes an improved CBAM to effectively focus on the key features of the EEG signals. The TC block employs TCN to capture high-level temporal features from sequential data. The multi-level feature concatenation enhances the model’s ability to represent complex features.
In the experiments, the CIACNet model achieves the best results on both the BCI IV-2a and BCI IV-2b datasets. Specifically, the model achieves an accuracy of 85.15% and a kappa score of 0.80 on the BCI IV-2a dataset. On the BCI IV-2b dataset, it achieves an accuracy of 90.05% and a kappa score of 0.80. Compared to the four other models included in the comparison, CIACNet demonstrates superior performance. The classification of MI-EEG signals remains a challenging task in current research. Existing results from SOTA models show limited improvements in MI-EEG signal classification algorithms. For instance, on the BCI IV-2a dataset, the EEG Conformer model proposed by Song et al. (2022) achieves an accuracy of 78.66%, representing a 2.46% improvement over previous models. The simplicity of the model limits its ability to achieve better classification performance. The MBMANet model proposed by Deng et al. (2024) achieves an accuracy of 83.18%, marking a 3.43% increase. On the BCI IV-2b dataset, the EEGNet Fusion V2 model proposed by Chowdhury et al. (2023) achieves an accuracy of 84.10%. This model focuses solely on the number of CNN branches and lacks additional blocks to further enhance performance. The proposed CIACNet model outperforms these studies on both datasets.
Ablation experiment confirms the necessity of each block within the CIACNet model for improving overall performance. When the CV2 block, IAT block, or TC block is removed from the model, its performance decreases. The performance decline becomes more pronounced when multiple blocks are removed. Additionally, we observe that the IAT block performs more effectively when placed in the middle of the model. By modifying the attention mechanism, we observe that our improved CBAM results in an increase of at least 0.77 and 0.35% in accuracy across both datasets compared to other attention mechanisms. This result verifies that the improved CBAM is highly effective at enhancing the model’s ability to extract key information. Our improved CBAM modifies the pooling approach of the standard CBAM. We add stochastic pooling to standard CBAM. Stochastic pooling assigns probabilities to feature map elements based on their numerical values. This increases the diversity of pooling, thereby enhancing the model’s ability to focus on key features.
We also present the confusion matrices and t-SNE images. In Figure 7, most subjects show good classification performance. However, subject 2 and subject 6 exhibit lower classification accuracy for left and right hand tasks. This indicates the presence of more artifacts in their EEG signals. In Table 2, the proposed model still significantly outperforms the other four models in terms of classification performance for subject 2. This demonstrates that our model maintains better performance when handling EEG data with more artifacts. The model captures key features and performs multi-level feature concatenation, reducing the impact of artifacts. The t-SNE images show that, for subjects with high classification accuracy, the proposed model effectively delineates cluster boundaries.
The proposed model in this study can be applied to the development of MI-BCI systems. By integrating with other devices, MI-BCI systems can achieve high accuracy. When utilized for upper limb rehabilitation in stroke patients, a decoding algorithm with high accuracy can enhance the precision of the feedback mechanism, thereby improving rehabilitation outcomes. Despite the advancements in MI-EEG signal decoding achieved in this study, several challenges remain. The proposed model is trained and validated individually for each subject, limiting its ability to effectively utilize information from other subjects. Future work will focus on incorporating algorithms such as transfer learning to enhance the generalizability of the models. We also plan to develop adaptive attention mechanisms. These mechanisms will automatically adjust attention allocation based on the varying needs of individuals and tasks. This enhances the versatility of decoding algorithms. We are confident that these efforts will drive further progress in BCI technology, bringing us closer to broader adoption of BCI applications.
This study proposes a composite improved attention convolutional network (CIACNet) for MI-EEG signals classification. The CIACNet model consists of three main components: the CV, IAT, and TC blocks. The experimental results from the BCI IV-2a and BCI IV-2b datasets shows that our model achieves accuracies of 85.15 and 90.05%, respectively. It outperforms the four other models under comparison. This demonstrates that our CIACNet model achieves high classification accuracy. It does not require artifact removal and has a short training time, thereby reducing the time cost in MI-BCI systems.
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/supplementary material.
WL: Writing – original draft, Writing – review & editing, Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision. ZM: Writing – original draft, Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization. SL: Writing – original draft, Data curation, Formal analysis, Methodology, Software, Validation, Visualization. LZ: Writing – original draft, Conceptualization, Data curation, Formal analysis, Methodology, Software. CL: Writing – review & editing, Conceptualization, Formal analysis, Funding acquisition, Investigation, Resources, Supervision, Validation.
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported in part by the Natural Science Foundation of Hebei Province (Grant No. F2024202019).
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declare that no Gen AI was used in the creation of this manuscript.
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Abdulkader, S. N., Atia, A., and Mostafa, M. S. M. (2015). Brain computer interfacing: applications and challenges. Egypt. Inf. J. 16, 213–230. doi: 10.1016/j.eij.2015.06.002
Altaheri, H., Muhammad, G., and Alsulaiman, M. (2022). Physics-informed attention temporal convolutional network for EEG-based motor imagery classification. IEEE Trans. Industr. Inform. 19, 2249–2258. doi: 10.1109/TII.2022.3197419
Altuwaijri, G. A., and Muhammad, G. (2022). Electroencephalogram-based motor imagery signals classification using a multi-branch convolutional neural network model with attention blocks. Bioengineering 9:323. doi: 10.3390/bioengineering9070323
Ang, K. K., Chin, Z. Y., Wang, C., Guan, C., and Zhang, H. (2012). Filter bank common spatial pattern algorithm on BCI competition IV datasets 2a and 2b. Front. Neurosci. 6:39. doi: 10.3389/fnins.2012.00039
Bai, S., Kolter, J. Z., and Koltun, V. (2018). An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv :1803.01271. doi: 10.48550/arXiv.1803.01271
Bi, J., Zhang, X., Yuan, H., Zhang, J., and Zhou, M. (2021). A hybrid prediction method for realistic network traffic with temporal convolutional network and LSTM. IEEE Trans. Autom. Sci. Eng. 19, 1869–1879. doi: 10.1109/TASE.2021.3077537
Biasiucci, A., Franceschiello, B., and Murray, M. M. (2019). Electroencephalography. Curr. Biol. 29, R80–R85. doi: 10.1016/j.cub.2018.11.052
Brunner, C., Leeb, R., Müller-Putz, G., Schlögl, A., and Pfurtscheller, G. (2008). BCI competition 2008–Graz data set a. institute for knowledge discovery (laboratory of brain-computer interfaces). Graz Univ. Technol. 16, 1–6. Available at: https://www.bbci.de/competition/iv/desc_2a.pdf
Cai, Z., Luo, T. J., and Cao, X. (2024). Multi-branch spatial-temporal-spectral convolutional neural networks for multi-task motor imagery EEG classification. Biomed. Signal Process. Control 93:106156. doi: 10.1016/j.bspc.2024.106156
Chatterjee, R., and Bandyopadhyay, T. (2016). “EEG based motor imagery classification using SVM and MLP.” in 2016 2nd international conference on computational intelligence and networks (CINE). pp. 84–89. IEEE.
Chen, W., Luo, Y., and Wang, J. (2024). Three-branch temporal-spatial convolutional transformer for motor imagery EEG classification. IEEE Access 12, 79754–79764. doi: 10.1109/ACCESS.2024.3405652
Chen, X., Teng, X., Chen, H., Pan, Y., and Geyer, P. (2024). Toward reliable signals decoding for electroencephalogram: a benchmark study to EEGNeX. Biomed. Signal Process. Control 87:105475. doi: 10.1016/j.bspc.2023.105475
Chowdhury, R. R., Muhammad, Y., and Adeel, U. (2023). Enhancing cross-subject motor imagery classification in EEG-based brain–computer interfaces by using multi-branch CNN. Sensors 23:7908. doi: 10.3390/s23187908
Deng, H., Li, M., Li, J., Guo, M., and Xu, G. (2024). A robust multi-branch multi-attention-mechanism EEGNet for motor imagery BCI decoding. J. Neurosci. Methods 405:110108. doi: 10.1016/j.jneumeth.2024.110108
Do, A. H., Wang, P. T., King, C. E., Chun, S. N., and Nenadic, Z. (2013). Brain-computer interface controlled robotic gait orthosis. J. Neuro Eng. Rehabil. 10, 111–119. doi: 10.1186/1743-0003-10-111
Gao, S., Yang, J., Shen, T., and Jiang, W. (2022). A parallel feature fusion network combining GRU and CNN for motor imagery EEG decoding. Brain Sci. 12:1233. doi: 10.3390/brainsci12091233
Grandchamp, R., and Delorme, A. (2011). Single-trial normalization for event-related spectral decomposition reduces sensitivity to noisy trials. Front. Psychol. 2:236. doi: 10.3389/fpsyg.2011.00236
Hewage, P., Behera, A., Trovati, M., Pereira, E., Ghahremani, M., Palmieri, F., et al. (2020). Temporal convolutional neural (TCN) network for an effective weather forecasting using time-series data from the local weather station. Soft. Comput. 24, 16453–16482. doi: 10.1007/s00500-020-04954-0
Hu, J., Shen, L., and Sun, G. (2018). “Squeeze-and-excitation networks.” in Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 7132–7141.
Hu, L., Hong, W., and Liu, L. (2023). MSATNet: multi-scale adaptive transformer network for motor imagery classification. Front. Neurosci. 17:1173778. doi: 10.3389/fnins.2023.1173778
Ingolfsson, T. M., Hersche, M., Wang, X., Kobayashi, N., Cavigelli, L., and Benini, L. (2020). “EEG-TCNet: an accurate temporal convolutional network for embedded motor-imagery brain–machine interfaces.” in 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC). pp. 2958–2965. IEEE.
Khabti, J., AlAhmadi, S., and Soudani, A. (2024). Optimal Channel selection of multiclass motor imagery classification based on fusion convolutional neural network with attention blocks. Sensors 24:3168. doi: 10.3390/s24103168
Khan, M. A., Das, R., Iversen, H. K., and Puthusserypady, S. (2020). Review on motor imagery based BCI systems for upper limb post-stroke neurorehabilitation: from designing to application. Comput. Biol. Med. 123:103843. doi: 10.1016/j.compbiomed.2020.103843
Lawhern, V. J., Solon, A. J., Waytowich, N. R., Gordon, S. M., Hung, C. P., and Lance, B. J. (2018). EEGNet: a compact convolutional neural network for EEG-based brain–computer interfaces. J. Neural Eng. 15:056013. doi: 10.1088/1741-2552/aace8c
Leeb, R., Brunner, C., Müller-Putz, G., Schlögl, A., and Pfurtscheller, G. J. G. U. O. T. (2008). BCI Competition 2008–Graz data set B. Graz Univ. Technol. Aust. 16, 1–6. Available at: https://www.bbci.de/competition/iv/desc_2b.pdf
Lu, P., Gao, N., Lu, Z., Yang, J., Bai, O., and Li, Q. (2019). “Combined CNN and LSTM for motor imagery classification.” in 2019 12th International Congress on Image and Signal Processing, Bio Medical Engineering and Informatics (CISP-BMEI). pp. 1–6. IEEE.
Luo, T. J., Zhou, C. L., and Chao, F. (2018). Exploring spatial-frequency-sequential relationships for motor imagery classification with recurrent neural network. BMC Bioinf. 19, 344–318. doi: 10.1186/s12859-018-2365-1
Ma, W., Gong, Y., Xue, H., Liu, Y., Lin, X., Zhou, G., et al. (2022). A lightweight and accurate double-branch neural network for four-class motor imagery classification. Biomed. Signal Process. Control 75:103582. doi: 10.1016/j.bspc.2022.103582
Musallam, Y. K., AlFassam, N. I., Muhammad, G., Amin, S. U., Alsulaiman, M., Abdul, W., et al. (2021). Electroencephalography-based motor imagery classification using temporal convolutional network fusion. Biomed. Signal Process. Control 69:102826. doi: 10.1016/j.bspc.2021.102826
Pfurtscheller, G., and Neuper, C. (2001). Motor imagery and direct brain-computer communication. Proc. IEEE 89, 1123–1134. doi: 10.1109/5.939829
Phunruangsakao, C., Achanccaray, D., and Hayashibe, M. (2022). Deep adversarial domain adaptation with few-shot learning for motor-imagery brain-computer interface. IEEE Access 10, 57255–57265. doi: 10.1109/ACCESS.2022.3178100
Ramadan, R. A., and Vasilakos, A. V. (2017). Brain computer interface: control signals review. Neurocomputing 223, 26–44. doi: 10.1016/j.neucom.2016.10.024
Ramoser, H., Muller-Gerking, J., and Pfurtscheller, G. (2000). Optimal spatial filtering of single trial EEG during imagined hand movement. IEEE Trans. Rehabil. Eng. 8, 441–446. doi: 10.1109/86.895946
Rashmi, C. R., and Shantala, C. P. (2022). EEG artifacts detection and removal techniques for brain computer interface applications: a systematic review. IJATEE 9, 354–383. doi: 10.19101/IJATEE.2021.874883
Salami, A., Andreu-Perez, J., and Gillmeister, H. (2022). EEG-ITNet: an explainable inception temporal convolutional network for motor imagery classification. IEEE Access 10, 36672–36685. doi: 10.1109/ACCESS.2022.3161489
Song, Y., Zheng, Q., Liu, B., and Gao, X. (2022). EEG conformer: convolutional transformer for EEG decoding and visualization. IEEE Trans. Neural Syst. Rehabil. Eng. doi: 10.1109/TNSRE.2022.3230250 [Epub ahead of print].
Song, Y., Zheng, Q., Wang, Q., Gao, X., and Heng, P. A. (2023). Global adaptive transformer for cross-subject enhanced EEG classification. IEEE Trans. Neural Syst. Rehabil. Eng. 31, 2767–2777. doi: 10.1109/TNSRE.2023.3285309
Vilela, M., and Hochberg, L. R. (2020). Applications of brain-computer interfaces to the control of robotic and prosthetic arms. Handb. Clin. Neurol. 168, 87–99. doi: 10.1016/B978-0-444-63934-9.00008-1
Wang, J., Yao, L., and Wang, Y. (2023). IFNet: an interactive frequency convolutional neural network for enhancing motor imagery decoding from EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 31, 1900–1911. doi: 10.1109/TNSRE.2023.3257319
Wang, P., Jiang, A., Liu, X., Shang, J., and Zhang, L. (2018). LSTM-based EEG classification in motor imagery tasks. IEEE Trans. Neural Syst. Rehabil. Eng. 26, 2086–2095. doi: 10.1109/TNSRE.2018.2876129
Wang, Z., Cao, L., Zhang, Z., Gong, X., Sun, Y., and Wang, H. (2018). Short time Fourier transformation and deep neural networks for motor imagery brain computer interface recognition. Concurr. Comput. Pract. Exper. 30:e4413. doi: 10.1002/cpe.4413
Wolpaw, J. R., Birbaumer, N., McFarland, D. J., Pfurtscheller, G., and Vaughan, T. M. (2002). Brain–computer interfaces for communication and control. Clin. Neurophysiol. 113, 767–791. doi: 10.1016/S1388-2457(02)00057-3
Woo, S., Park, J., Lee, J. Y., and Kweon, I. S. (2018). “Cbam: Convolutional block attention module.” in Proceedings of the European Conference on Computer Vision (ECCV). pp. 3–19.
Xie, X., Chen, L., Qin, S., Zha, F., and Fan, X. (2024). Bidirectional feature pyramid attention-based temporal convolutional network model for motor imagery electroencephalogram classification. Front. Neurorobot. 18:1343249. doi: 10.3389/fnbot.2024.1343249
Xu, B., Zhang, L., Song, A., Wu, C., Li, W., Zhang, D., et al. (2018). Wavelet transform time-frequency image and convolutional network-based motor imagery EEG classification. IEEE Access 7, 6084–6093. doi: 10.1109/ACCESS.2018.2889093
Zeiler, M. D., and Fergus, R. (2013). Stochastic pooling for regularization of deep convolutional neural networks. arXiv :1301.3557. doi: 10.48550/arXiv.1301.3557
Zhang, R., Liu, G., Wen, Y., and Zhou, W. (2023). Self-attention-based convolutional neural network and time-frequency common spatial pattern for enhanced motor imagery classification. J. Neurosci. Methods 398:109953. doi: 10.1016/j.jneumeth.2023.109953
Keywords: electroencephalography, convolution neural network, attention mechanism, temporal convolution network, motor imagery, classification
Citation: Liao W, Miao Z, Liang S, Zhang L and Li C (2025) A composite improved attention convolutional network for motor imagery EEG classification. Front. Neurosci. 19:1543508. doi: 10.3389/fnins.2025.1543508
Received: 11 December 2024; Accepted: 23 January 2025;
Published: 05 February 2025.
Edited by:
Qiu-Hua Lin, Dalian University of Technology, ChinaReviewed by:
Xinzhou Xie, Northwestern Polytechnical University, ChinaCopyright © 2025 Liao, Miao, Liang, Zhang and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Chen Li, bGljaGVuQHRtdS5lZHUuY24=
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
Research integrity at Frontiers
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.