- 1China University of Mining and Technology-Beijing, Beijing, China
- 2State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China
- 3Chinese Academy of Sciences (CAS) Center for Excellence in Brain Science and Intelligence Technology, Beijing, China
Introduction: During electroencephalography (EEG)-based motor imagery-brain-computer interfaces (MI-BCIs) task, a large number of electrodes are commonly used, and consume much computational resources. Therefore, channel selection is crucial while ensuring classification accuracy.
Methods: This paper proposes a channel selection method by integrating the efficient channel attention (ECA) module with a convolutional neural network (CNN). During model training process, the ECA module automatically assigns the channel weights by evaluating the relative importance for BCI classification accuracy of every channel. Then a ranking of EEG channel importance can be established so as to select an appropriate number of channels to form a channel subset from the ranking. In this paper, the ECA module is embedded into a commonly used network for MI, and comparative experiments are conducted on the BCI Competition IV dataset 2a.
Results and discussion: The proposed method achieved an average accuracy of 75.76% with all 22 channels and 69.52% with eight channels in a four-class classification task, outperforming other state-of-the-art EEG channel selection methods. The result demonstrates that the proposed method provides an effective channel selection approach for EEG-based MI-BCI.
1. Introduction
A brain-computer Interface (BCI) provides an interface between users and external devices by converting brain signals into commands (Fadel et al., 2020). Electroencephalography (EEG), magnetoencephalography (MEG), functional magnetic resonance imaging (fMRI), and functional near-infrared spectroscopy (fNIRS), et al., are commonly used to acquire signals from the brain (Herrera-Vega et al., 2017; Berger et al., 2019; Saha et al., 2021). Among these techniques, EEG is currently one of the most popular brain-imaging techniques for its non-invasive nature, portability, and low cost (Abiri et al., 2019; Aggarwal and Chugh, 2022). EEG-based BCI can be applied to various tasks, such as seizure detection, workload, motor imagery (MI), and emotion recognition (Das Chakladar et al., 2020; Lashgari et al., 2020; Gu et al., 2021). In the MI task, subjects are asked to imagine human movements without performing them. EEG signals generated during this process are collected, and the intention can thus be recognized (Chaisaen et al., 2020; Khan et al., 2020). MI-BCI has numerous applications for aiding the elderly and disabled (Lazarou et al., 2018; Palumbo et al., 2021; Saibene et al., 2023).
MI-BCI is built on the fact that the brain evokes event-related synchronization (ERS) and event-related desynchronization (ERD) at different locations over the scalp when imagining body movements (Lu et al., 2017; Zhang et al., 2021). Researchers have progressively placed more and more electrodes on the subject's scalp to record signals for more detailed information, making channel selection a critical stage for EEG-based BCI (Abdullah et al., 2022). Channel selection aims to identify a small subset of channels that involves the most classified information. It can reduce computational costs and the interference of irrelevant EEG channels. Thus, the most important issues in channel selection are: how many channels are suitable, and which channels should be selected. This work focuses on the number of channels, that is ensuring the high recognition accuracy while minimizing the channel subset.
Researchers have been making many effort to find a suitable channel selection method. These methods can be generally classified as filtering, wrapper, embedded, and hybrid techniques. Utilizing wrapper techniques for channel selection demands a substantial computational resource investment. Tang et al. (2022) introduced the sequential backward floating search (SBFS) method into EEG channel selection. While significant improvements have been made to reduce training time, it still necessitates ~2,000 seconds or more to complete channel selection. Therefore, this paper focuses on filtering and embedded techniques. The filtering technique is independent of subjects and classifiers, high-speed, stable, and does not consume significant computational resources (Baig et al., 2020), and the embedded technique can be used with deep learning techniques. Most filtering techniques are performed based on the statistical information of the EEG signal. Tam et al. (2011) proposed a channel selection method based on common spatial pattern (CSP) filter coefficient ranking for MI classification, called CSP-rank. They conducted a 20-session experiment with 64-channel EEG from five chronic stroke patients and found that the average classification accuracy of CSP-rank for 8–38 electrodes remained above 90%. Twenty-two electrodes achieved the highest average accuracy of 91.70%. Arvaneh et al. (2011) adopted a new filtering approach with a pre-specified subset channel selection scheme. They proposed a sparse common spatial pattern (SCSP) algorithm instead of CSP for optimal EEG channel selection. The results show that the SCSP algorithm outperformed the existing algorithms, including Fisher discriminant, mutual information, support vector machine and CSP. If the goal is to minimize the number of channels while maintaining comparable average accuracy to using all channels, SCSP achieved 79.07% accuracy with an average of 8.55 channels on the first dataset and 79.28% accuracy with an average of 7.6 channels on the second dataset. Shi et al. (2023) proposed an EEG channel selection method based on sparse logistic regression (SLR). This method was compared to conventional channel selection based on correlation coefficients (CCS) using a 64-channel two-class MI dataset. In the scenarios of selecting 10 channels and 16 channels, the accuracy achieved by the proposed method were 86.63 and 87.00%, respectively, demonstrating a performance advantage of 4.33 and 2.94% over CCS.
In recent years, deep learning techniques have shown advantages in big data processing. Several attempts have been made to explore an optimal embeded channel selection method. Zhang et al. (2021) proposed a deep learning-based approach to automatically select the relevant EEG channels while recognizing two MI states. A sparse squeeze-and-excitation (SE) module is used to learn EEG channels' contribution to MI classification, which developed into an automatic channel selection strategy. The results show that this method indicates a 3.30% improvement compared to CSP in their private dataset. Strypsteen and Bertrand (2021) employed a concrete selector layer to optimize both the channel selection and the network weights in an end-to-end manner. This layer uses a Gumbel-softmax method to deal with the discrete parameters inherent to a subset selection problem. It can freely specify the number of channel subsets by modifying the loss function so that the network does not pick the same channels as much as possible. Their method was evaluated on two EEG tasks: motor execution and auditory attention decoding. The Gumbel-softmax performs at least as well as (often better than) state-of-the-art methods: mutual information for motor execution and greedy channel selection with the utility metric for auditory attention decoding. These studies show that using deep learning techniques instead of traditional signal processing methods is credible and can substantially improve performance.
This paper proposes a new method for EEG channel selection by introducing efficient channel attention (ECA) modules into a convolutional neural network (CNN) (Wang et al., 2020). An ECA module can recalibrate the channels based on feature interdependencies, allowing the network to learn each EEG channel's importance to classification for each subject and improve performance. Using the proposed approach, a personalized optimal channel subset can be obtained based on the order of channels in the learned channel importance ranking. This ensures that the channel subset includes the most discriminative channels for each subject. The main contributions of this work are as follows:
1. An innovative method is proposed to find an optimal channel subset for each subject with the ECA module. The subset is formed based on the importance of each channel for that subject in the MI classification process. Researchers can easily adjust the number of channels according to actual needs and hardware conditions.
2. ECA modules are added between the convolutional layers of the CNN to recalibrate feature interdependencies between channels adaptively. A CNN structure called ECA-DeepNet based on DeepNet (Schirrmeister et al., 2017) is proposed, and classification accuracy of the network improves with minimal computational cost.
The rest of the paper is organized as follows. Section 2 describes the detail of the ECA module and the proposed method for channel selection. Section 3 presents the experimental results to validate and compare the method with state-of-the-art methods. Some discussions are given in Section 4.
2. Materials and methods
This section presents a detailed description of the proposed channel selection method. As depicted in Figure 1, the entire process consists of four main steps: (1) training the proposed model with the data from the subject, (2) extracting the channel weights from the channel attention (CA) layer of the trained network, (3) ranking the importance of each channel based on the extracted weights, and (4) selecting a certain number of channels from the ranking to form an optimal channel subset for that subject.
2.1. Dataset and data preprocessing
To analyze the EEG signals and evaluate the proposed method, the publicly available MI-EEG dataset “BCI Competition IV dataset 2a (BCIC IV 2a dataset)” is introduced (Brunner et al., 2008). This dataset comprises four different MI tasks: the movement of the left hand, right hand, feet, and tongue. Each task lasts for 4 s from cue onset to the end of the task, and no feedback is provided. The dataset contains 22-channel EEG data from nine subjects, and each subject has two sets of data, namely the training set and testing set. Each set includes 288 trials of MI data, with 72 trials for each of the four tasks. The EEG data were initially sampled at a frequency of 250 Hz and filtered with a bandpass filter between 0.5 and 100 Hz, as well as a notch filter at 50 Hz.
The EEG signals were preprocessed as follows. A bandpass filter between 1 and 40 Hz was applied to reduce the effect of eye blinking and extract the information related to MI. An exponential moving average with a decay factor of 0.999 was applied to each channel to normalize the continuous data. The data was then segmented into four-second time windows with a sliding window over the time period from −0.5 to 4 s for each trial. After cropped by the sliding window, there are 864 samples for each subject in the training set and testing set separately, with 216 samples for each of the four tasks. Each sample consists of MI data from 22 channels, and each channel contains 1,000 sampling points.
2.2. Channel attention layer
To perform channel selection, evaluating the correlation between channels is necessary. Irrelevant channels are deemed redundant and not included in the channel subset. The ECA module is a channel attention module based on the attention mechanism (Wang et al., 2020). It can learn the interdependencies among different feature maps from the channel dimension and adaptively adjust the channel features by assigning weights to each channel. By placing the ECA module as the first layer of the network, it can learn the interdependence among EEG channels. Specifically, it scores the importance of each EEG channel. It assigns weights to strengthen the crucial channels while weakening the influence of less important channels on the rest of the model. This score also serves as the basis for subsequent channel selection. As a result, this layer is referred to as the channel attention (CA) layer here. The ECA module's efficiency lies on avoiding channel dimensionality reduction and enabling appropriate local cross-channel interactions. Next, the working mechanism of the ECA module will be explained, illustrating on how it achieves these two objectives.
The input of the ECA module is denoted as X∈ℝW×H×C, where W, H and C are width, height and channel dimension. When the ECA module serves as the CA layer, the input is the EEG data, and at this point, W = 1. The ECA module first calculates the aggregated features y∈ℝ1 × 1 × C along the channel dimension without dimensionality reduction, which is represented as
where g(X) is channel-wise global average pooling (GAP). This operation aggregates feature maps of the EEG sample across time-space dimensions W×H, producing channel-wise statics embedded with the global distribution of feature responses. In general, channel attention can be learned by
where W involves C×C parameters, and σ is a Sigmoid function. However, the ECA module employs a band matrix Wk to learn channel attention:
where Wk is a k×C parameter matrix with much fewer parameters than W. As in Equation (3), the weight of yi only considers the interaction between yi and its adjacent k neighbors, and it will be more efficient to make all the channels share the same learning parameters, i.e.,
where indicates the set of k adjacent channels of yi. Such a strategy can be easily implemented by fast 1D convolution with a kernel size of k, which can be described as
where Conv1d is 1D convolution.
Since the ECA module aims to capture local cross-channel interaction appropriately, the kernel size k of 1D convolution must vary with different channel dimensions C. High channel dimension needs longer range interaction, and low channel dimension has shorter range interaction. Therefore, the kernel size k should be adaptively determined by channel dimension C, i.e.,
where |k|odd indicates the nearest odd number of k.
However, when the ECA module is used as the CA layer, it is required to no longer focus on local, but capture global cross-channel interactions in order to learn feature interdependencies between EEG channels thoroughly. This is because EEG channels interact between both adjacent and non-adjacent channels, especially within corresponding regions of the contralateral brain hemisphere. Therefore, the 1D fast convolution in the CA layer is replaced by a fully connected layer. The number of neurons in this fully connected layer matches the channel dimension C, thus avoiding the influence of dimension reduction on channel attention learning, i.e.,
where WCA represents the weight parameter matrix of the fully connected layer, and b is the bias coefficients. This fully connected layer will incur additional computational costs, but it ensures that the CA layer considers all channels when learning the importance of each channel. Then channel weights with the same length as channel dimension C are obtained, and the ECA module will output the recalibrated feature maps along the channel dimension without altering the dimensions of the input samples, i.e.,
Figure 2 illustrates the overview of an ECA module. After aggregating feature maps using GAP, the ECA module uses a 1D fast convolution with adaptive kernel size k, followed by a Sigmoid function to learn the channel attention. Finally, input features are recalibrated from the channel dimension by assigning weights.
Figure 2. Diagram of an ECA module. When the ECA module is utilized as the CA layer, 1D fast convolution is substituted with a fully connected layer.
2.3. ECA modules in DeepNet
In addition to placing the ECA module as the CA layer to learn the importance of EEG channels, positioning the ECA module between the convolutional layers of the network also serves to highlight significant feature maps and suppress irrelevant ones, thereby improving classification performance. To demonstrate the effects, this paper integrates the ECA modules into DeepNet (Schirrmeister et al., 2017), which is one of the most highly cited open-source models (Dai et al., 2020). It splits the first convolutional layer into a first convolution across time and a second convolution across space (electrodes), exploiting the ERS and ERD phenomena more effectively. Four ECA modules are added between convolution layers, and an additional ECA module is set before the first layer to act as the CA layer. Termed as ECA-DeepNet, the complete architecture is depicted in Figure 3.
Exponential linear units (ELUs) (Clevert et al., 2016), denoted as Equation (9), are selected as activation functions since they can speed up the learning process and improve classification accuracy.
The detailed parameters of the proposed ECA-DeepNet architecture are given in Table 1. Note that the number of filters in four convolution-max-pooling blocks are changed from (25, 50, 100, 200) to (32, 64, 128, 256) to fit the mapping as shown in Equation (6) better.
2.4. Proposed channel selection method
The channel selection method proposed in this paper is based on the weights within the CA layer of the network. As described in Section 2.2, during the network training on a subject's EEG data, the CA layer automatically learns the attention between channels, assigning corresponding weights based on the importance of each EEG channel in the classification task. More important channels are assigned higher weights, while less important channels are assigned lower weights. Therefore, ECA-DeepNet can be used to train a model for each subject on BCIC IV 2a dataset. The channel weights w in the CA layer of each model can be collected after the training process.
After reordering channels by the extracted weights, the importance ranking of EEG channels for each subject can be described as
where chj is the channel name with the j-th largest wch. Finally, the optimal channel subset for subject i can be obtained from the ranking Ri, i.e.,
where Nc is the number of channels in the channel subset. Researchers can freely determine the size of the channel subset Nc according to their specific needs.
3. Results
3.1. Hyperparameter optimization
Structural hyperparameters of the ECA-DeepNet architecture have been specified and presented in Table 1. For the hyperparameter optimization, the open-source framework Optuna was employed in this paper (Akiba et al., 2019). This framework uses the Tree-structured Parzen Estimator (TPE) to progressively reduce the parameter search space until it finds the optimal value. The hyperparameter search space in the network is shown in Table 2.
This study aims to find the optimal subject-specific classification model. Therefore, each subject obtained a corresponding set of best hyperparameters with and without channel selection during the hyperparameter search. When searching, the training set of BCIC IV 2a dataset was re-split into a training set and a validation set in a radio of 8:2. The best-performing set of hyperparameters was determined based on the highest accuracy on the validation set. Additionally, experiments took place on NVIDIA TITAN V GPUs. To ensure the replicability of results across all experiments, a consistent random seed of 20200220 was employed. Table 3 represents the results of best hyperparameters tuning by Optuna for each subject in BCIC IV 2a dataset.
3.2. Performance of the proposed channel selection method
The proposed method incorporates channel attention learning process and classification during model training through the CA layer. Due to this unique feature, the method can be classified as an embedded technique. To evaluate the channel selection ability of the proposed method, it was compared with two other state-of-the-art embedded techniques: Gumbel-softmax (GS) layer and the automatic channel selection (ACS) layer (Strypsteen and Bertrand, 2021; Zhang et al., 2021).
The network architecture was identical to that of ECA-DeepNet, except for replacing the CA layer with the GS or ACS layer for comparison. The ACS layer is based on another attention module, the SE module (Hu et al., 2018), which has been proven effective and commonly used in many fields (Park et al., 2020; Liu et al., 2021; Zhang and Zhang, 2022). It can be used for channel selection comparison following the steps proposed in this paper. Considering the potential adverse effects between the ECA and SE modules, network with all ECA modules replaced with SE modules was also compared, and it is named All-SE.
When the CA layer is used for channel selection, the relationship between the number of input channels, average classification accuracy, and prediction time of the model is illustrated in Figure 4. As the number of channels increases, the prediction time and classification accuracy were gradually increasing. The average accuracy curve started to flatten when more than eight channels were involved and the increase in accuracy became slow when more than 14 channels were involved. The aim of this study is to maintain the high accuracy while minimizing the number of input EEG channels. Taking into account the trade-off between real-time requirements and accuracy in MI-BCI, this study takes eight channels as an illustrative case to analyze the channel selection method from all 22 channels and compare different methods.
Figure 4. Relationship between the number of input channels, average classification accuracy, and prediction time of the model with the ECA-DeepNet method.
Table 4 demonstrates the classification accuracy for each subject after selecting eight channels by different channel selection methods. In Table 4, the standard deviation (STD) of accuracy across nine subjects was calculated to evaluate the method's robustness across different subjects. Results show that the CA layer achieved optimal channel selection performance in most subjects, with its average eight-channel accuracy of 69.52%, indicating a 13.81%, 5.46% and 3.17% improvement compared to the GS layer (55.71%), All-SE (64.06%), and the ACS layer (66.35%). The proposed method showed a large improvement in the accuracy of subject No. 4 (up to 7.17%) and No. 7 (up to 12.97%).
Table 4. Classification accuracies (%) for eight channels with the GS layer, All-SE, ACS or CA layer.
In addition, a 5-fold cross-validation on the training set was also carried out to further assess the robustness and stability of the proposed method. The results of cross-validation are consistent with those of the hold-out method, demonstrating that the CA layer exhibited superior channel selection capabilities compared to other methods.
Figure 5 shows changes in average accuracy before and after using different channel selection methods. The average accuracy decreased by 18.46% (GS layer), and 7.82% (ACS layer) after selecting eight channels from 22 channels. The CA layer performed the best, reducing the number of channels by 63.64% at the cost of a 4.65% decrease in classification performance.
Figure 5. Classification accuracies (%) comparison for each subject in BCIC IV 2a dataset from 22 to eight channels using different channel selection methods.
To further evaluate the channel selection capability of the proposed method for different number of channels, its average performance across all subjects was compared with other channel selection methods for different channel subset size. Specifically, comparisons were made for subsets of 10, 12, 14, and 16 channels. These specific channel subset sizes were chosen because accuracy significantly degraded when the number of channels was less than eight, while exceeding 16 channels may impact real-time processing speed, as shown in Figure 4. The hold-out and training set cross-validation results of these comparisons are presented in Table 5. The results indicate that the CA layer consistently demonstrated optimal or near-optimal channel selection capabilities in all scenarios, substantiating the effectiveness and practicality of the proposed method for channel selection in MI-BCI.
Table 5. Average classification accuracies (%) across all subjects for 10, 12, 14, and 16 channels with the GS layer, All-SE, ACS or CA layer.
3.3. Position of the selected channels
When the number of electrodes (or channels) is determined, which channels should be selected is a long-standing issue when decoding EEG signals. Figure 6 illustrates the electrode distribution of the eight-channel subset obtained using the proposed method for each subject on BCIC IV 2a dataset.
Figure 6. Position of each subject's personalized channel subset in 10–20 system for BCIC IV 2a dataset. The channels marked in green are the selected channels.
The optimal eight-channel subsets varied among each subject. Subjects No.3, No.7, and No.9 exhibited a tendency toward the parietal region (near electrode Pz), responsible for processing somatosensory information, in their eight selected channels. On the other hand, subject No.1 showed a preference for the frontal region (near electrode Fz), associated with motor control and execution, in his eight selected channels. For the remaining subjects, the electrodes were distributed across both parietal and frontal lobe regions. This indicates that both the parietal and frontal regions, in addition to the motor cortex, play crucial roles in the MI process. Furthermore, the eight-channel subset for most subjects were concentrated near the central sulcus region (electrodes C3, C4, Cz), which is considered to have the most evident ERD/ERS phenomena and is most commonly used for MI classification.
3.4. Performance of ECA modules
To evaluate the performance of ECA modules on classification accuracy, comparative experiment was taken among EEGNet (Lawhern et al., 2018), which is also a highly-cited model, DeepNet, ECA-DeepNet, and SE-DeepNet. All 22 channels and eight channels selected by proposed method were used for the comparison, and the hold-out and training set cross-validation results are shown in Tables 6, 7.
Table 6. Classification accuracy (%) with all 22 channels for EEGNet, DeepNet, SE-DeepNet, ECA-DeepNet methods.
Table 7. Classification accuracy (%) with eight channels for EEGNet, DeepNet, SE-DeepNet, and ECA-DeepNet methods.
It can be found that incorporating ECA modules into the network effectively improved the classification accuracy, and it outperformed the SE module in eight out of nine subjects with both 22 channels and eight channels. The average accuracies of ECA-DeepNet, SE-DeepNet, DeepNet, and EEGNet were 75.76, 72.58, 73.89, and 72.06% respectively with 22 channels and 69.52, 67.03, 64.79, and 62.44%, respectively, with eight channels. The results of cross-validation align with this finding, demonstrating that ECA-DeepNet achieved optimal classification performance in most cases.
The most influential factor leading to the difference between the ECA module and SE module is that the ECA module avoids the side effects of dimensionality reduction on channel attention prediction. The main difference between the two modules is that: To limit model complexity, the SE module adjusts the dimensionality of the two fully connected layers, i.e., by projecting channel features into a low-dimensional space and then mapping them back. At the same time, the ECA module chooses to implement it with 1D fast convolution of kernel size k or an equivalently dimensioned fully connected layer, maintaining the same channel dimensionality (when used as the CA layer). The dimensionality reduction operation in the SE module destroys the direct correspondence between a channel and its weights, directly leading to SE modules not achieving the same performance as ECA modules.
To evaluate the importance of the CA layer and ECA module in ECA-DeepNet, the performance of ECA-DeepNet without the CA layer was compared to both DeepNet and the complete ECA-DeepNet. The comparison results are shown in Tables 6, 7. It can be observed that removing either the CA layer or the ECA modules between the convolutional layers from ECA-DeepNet resulted in a decrease in accuracy. Removing the CA layer led to a decrease in accuracy by 1.59% with 22 channels and 1.93% with eight channels, while removing all ECA modules resulted in a decrease in accuracy by 1.87% with 22 channels and 4.73% with eight channels. The results of cross-validation align closely with this conclusion. The removal of the ECA module exhibited a more significant impact with eight channels, possibly attributed to these eight channels being selected by the ECA-DeepNet network. The characteristic of the embedded channel selection method being trained alongside the classifier determines that changes in the model structure will result in a misalignment between the selected channels and the new model. This misalignment had a relatively minor effect when only the CA layer was removed, but it introduced a substantial adverse impact when all ECA modules were eliminated.
4. Discussion
In this study, a novel embedded channel selection method based on the ECA module is introduced for MI-BCI. The publicly available BCIC IV 2a dataset was employed to compare and evaluate the performance of this method through the hold-out validation and training set cross-validation. The experimental results reveal that, compared to the other two state-of-the-art techniques, the proposed method achieved superior performance in selecting channel subsets (Tables 4, 5). The ECA module, functioning as an attention mechanism, not only serves for channel selection but also contributes to the further enhancement of network performance (Tables 6, 7).
A long-standing challenge pertains to the selection of channels for the decoding of EEG signals when the number of channels is pre-determined. Figure 6 illustrates that even among healthy subjects, the optimal subset of eight channels can exhibit significant variation. While the majority of these channels tended to cluster around the central sulcus, for some individuals, the distribution of eight channels extended to the frontal lobe region, while others showcased distribution within the parietal lobe region, or a combination across both regions. This observation indicates the crucial role played by both the parietal and frontal lobes in the process of MI-EEG decoding. This result is consistent with anatomy and previous studies (Pfurtscheller and da Silva, 1999; Hetu et al., 2013; Park and Chung, 2020), providing the theoretical basis and interpretability to the channel subsets obtained through the proposed method in this paper.
In this study, one of the chosen comparative methods was the SE module (Hu et al., 2018), which is also an channel attention module. The performance of the SE module fell short of the ECA module in most cases (Tables 4–7). This disparity primarily arises from the fact that the SE module introduces dimensionality reduction when forming the bottleneck-like structure. This reduction operation may result in loss of feature information during the process of channel attention learning and disrupt the direct correspondence between channels and attention weights. In contrast, the ECA module usually employs 1D convolution operations, allowing each channel to aggregate information from surrounding channels. When it serves as the CA layer, a fully connected layer that does not alter the channel dimension is employed to learn channel attention. These designs enable a more direct capture of interdependencies between channels.
EEG-based MI classification is a prominent research direction. To enhance accuracy, researchers employ an increasing number of electrodes to acquire more comprehensive information. However, different channels contribute to the classification process in distinct ways, and redundant channels may arise. Consequently, an appropriate channel subset through channel selection becomes highly necessary. This paper proposes a novel channel selection method that integrates the ECA module with a CNN. During the training process, the module automatically learns the importance of individual channels based on feature interdependencies and improves performance as well. Based on the extracted channel weights after training, a ranking of channel importance is established. Researchers can select appropriate channel subsets for different subjects based on practical accuracy and hardware requirements from the ranking. Using the BCIC IV 2a dataset, the proposed method was compared with two state-of-the-art embedded channel selection methods, namely ACS Layer and GS Layer. The proposed method achieved an average accuracy of 69.52% with eight channels, outperforming the other two algorithms. The selected eight channels align with prior research and anatomical knowledge. Furthermore, to evaluate the impact of integrating the ECA module into a CNN, a comparison of classification performance was conducted between EEGNet, the original DeepNet, the proposed ECA-DeepNet, and SE-DeepNet. The ECA-DeepNet achieved an accuracy of 75.76% with 22 channels and 69.52% with eight selected channels, exhibiting a 1.93 and 4.73% accuracy improvement over the original DeepNet. The experimental results demonstrate that the ECA module not only assists in channel selection but also improves classification performance. Therefore, this paper introduces a feasible approach for channel selection in EEG-based MI-BCIs.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Author contributions
LT: Conceptualization, Funding acquisition, Methodology, Validation, Writing—original draft, Writing—review and editing. YQ: Conceptualization, Formal analysis, Methodology, Validation, Writing—original draft, Writing—review and editing. LP: Conceptualization, Formal analysis, Funding acquisition, Methodology, Writing—review and editing. CW: Formal analysis, Validation, Visualization, Writing—original draft. Z-GH: Funding acquisition, Project administration, Supervision, Writing—review and editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported in part by the National Key Research and Development Program of China under Grant 2022YFC3601200, in part by the National Natural Science Foundation of China under Grants 62203441 and U21A20479, and in part by the Beijing Natural Science Foundation under Grant 4232053 and L222013.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Abdullah Faye, I., and Islam, M. R. (2022). EEG channel selection techniques in motor imagery applications: a review and new perspectives. Bioengineering 9, 32. doi: 10.3390/bioengineering9120726
Abiri, R., Borhani, S., Sellers, E. W., Jiang, Y., and Zhao, X. P. (2019). A comprehensive review of EEG-based brain-computer interface paradigms. J. Neural Eng. 16, 21. doi: 10.1088/1741-2552/aaf12e
Aggarwal, S., and Chugh, N. (2022). Review of machine learning techniques for EEG based brain computer interface. Arch. Comp. Methods Eng. 29, 3001–3020. doi: 10.1007/s11831-021-09684-6
Akiba, T., Sano, S., Yanase, T., Ohta, T., Koyama, M., and Assoc Comp, M. (2019). “Optuna: a next-generation hyperparameter optimization framework,” in 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD) (New York, NY: Assoc Computing Machinery), 2623–2631.
Arvaneh, M., Guan, C. T., Ang, K. K., and Quek, C. (2011). Optimizing the channel selection and classification accuracy in EEG-based BCI. IEEE Transact. Biomed. Eng. 58, 1865–1873. doi: 10.1109/TBME.2011.2131142
Baig, M. Z., Aslaml, N., and Shum, H. P. H. (2020). Filtering techniques for channel selection in motor imagery EEG applications: a survey. Artif. Intell. Rev. 53, 1207–1232. doi: 10.1007/s10462-019-09694-8
Berger, A., Horst, F., Muller, S., Steinberg, F., and Doppelmayr, M. (2019). Current state and future prospects of EEG and fNIRS in robot-assisted gait rehabilitation: a brief review. Front. Hum. Neurosci. 13, 17. doi: 10.3389/fnhum.2019.00172
Brunner, C., Leeb, R., Müller-Putz, G., Schlögl, A., and Pfurtscheller, G. (2008). BCI Competition 2008—Graz Data Set A. The Republic of Austria: Institute for Knowledge Discovery, Graz University of Technology, 1–6.
Chaisaen, R., Autthasan, P., Mingchinda, N., Leelaarporn, P., Kunaseth, N., Tammajarung, S., et al. (2020). Decoding EEG rhythms during action observation, motor imagery, and execution for standing and sitting. IEEE Sens. J. 20, 13776–13786. doi: 10.1109/JSEN.2020.3005968
Clevert, D.-A., Unterthiner, T., and Hochreiter, S. (2016). “Fast and accurate deep network learning by exponential linear units (elus),” in 4th International Conference on Learning Representations, ICLR 2016, May 2, 2016 - May 4, 2016, 4th International Conference on Learning Representations, ICLR 2016 - Conference Track Proceedings. International Conference on Learning Representations, ICLR (San Juan).
Dai, G. H., Zhou, J., Huang, J. H., and Wang, N. (2020). Hs-CNN: a CNN with hybrid convolution scale for EEG motor imagery classification. J. Neural Eng. 17, 11. doi: 10.1088/1741-2552/ab405f
Das Chakladar, D., Dey, S., Roy, P. P., and Dogra, D. P. (2020). EEG-based mental workload estimation using deep BLSTM-LSTM network and evolutionary algorithm. Biomed. Signal Process. Control 60, 10. doi: 10.1016/j.bspc.2020.101989
Fadel, W., Kollod, C., Wandow, M., Ibrahim, Y., and Ulbert, I. (2020). Multi-class classification of motor imagery EEG signals using image-based deep recurrent convolutional neural network. IEEE 193–196. doi: 10.1109/BCI48061.2020.9061622
Gu, X. T., Cao, Z. H., Jolfaei, A., Xu, P., Wu, D. R., Jung, T. P., et al. (2021). EEG-based brain-computer interfaces (BCIS): a survey of recent studies on signal sensing technologies and computational intelligence approaches and their applications. IEEE ACM Transact. Comp. Biol. Bioinformat. 18, 1645–1666. doi: 10.1109/TCBB.2021.3052811
Herrera-Vega, J., Trevino-Palacios, C. G., and Orihuela-Espina, F. (2017). Neuroimaging with functional near infrared spectroscopy: from formation to interpretation. Infrared Phys. Technol. 85, 225–237. doi: 10.1016/j.infrared.2017.06.011
Hetu, S., Gregoire, M., Saimpont, A., Coll, M. P., Eugene, F., Michon, P. E., et al. (2013). The neural network of motor imagery: an ale meta-analysis. Neurosci. Biobehav. Rev. 37, 930–949. doi: 10.1016/j.neubiorev.2013.03.017
Hu, J., Shen, L., and Sun, G. (2018). “Squeeze-and-excitation networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Salt Lake City, UT), 7132–7141.
Khan, M. A., Das, R., Iversen, H. K., and Puthusserypady, S. (2020). Review on motor imagery based bci systems for upper limb post-stroke neurorehabilitation: from designing to application. Comput. Biol. Med. 123, 17. doi: 10.1016/j.compbiomed.2020.103843
Lashgari, E., Liang, D. H., and Maoz, U. (2020). Data augmentation for deep-learning-based electroencephalography. J. Neurosci. Methods 346, 26. doi: 10.1016/j.jneumeth.2020.108885
Lawhern, V. J., Solon, A. J., Waytowich, N. R., Gordon, S. M., Hung, C. P., and Lance, B. J. (2018). EEGNet: a compact convolutional neural network for EEG-based brain-computer interfaces. J. Neural Eng. 15:056013. doi: 10.1088/1741-2552/aace8c
Lazarou, I., Nikolopoulos, S., Petrantonakis, P. C., Kompatsiaris, I., and Tsolaki, M. (2018). EEG-based brain-computer interfaces for communication and rehabilitation of people with motor impairment: a novel approach of the 21st century. Front. Hum. Neurosci. 12, 14. doi: 10.3389/fnhum.2018.00014
Liu, F., Zhou, X., Yan, X. H., Lu, Y. L., and Wang, S. D. (2021). Image steganalysis via diverse filters and squeeze-and-excitation convolutional neural network. Mathematics 9, 13. doi: 10.3390/math9020189
Lu, N., Li, T. F., Ren, X. D., and Miao, H. Y. (2017). A deep learning scheme for motor imagery classification based on restricted boltzmann machines. IEEE Transact. Neural Syst. Rehabil. Eng. 25, 566–576. doi: 10.1109/TNSRE.2016.2601240
Palumbo, A., Gramigna, V., Calabrese, B., and Ielpo, N. (2021). Motor-imagery eeg-based bcis in wheelchair movement and control: a systematic literature review. Sensors 21:6285. doi: 10.3390/s21186285
Park, J., Kim, J. K., Jung, S., Gil, Y., Choi, J. I., and Son, H. S. (2020). ECG-signal multi-classification model based on squeeze-and-excitation residual neural networks. Appl. Sci. 10, 8. doi: 10.3390/app10186495
Park, Y., and Chung, W. (2020). Optimal channel selection using correlation coefficient for csp based EEG classification. IEEE Access 8, 111514–111521. doi: 10.1109/ACCESS.2020.3003056
Pfurtscheller, G., and da Silva, F. H. L. (1999). Event-related EEG/MEG synchronization and desynchronization: basic principles. Clin. Neurophysiol. 110, 1842–1857. doi: 10.1016/S1388-2457(99)00141-8
Saha, S., Mamun, K. A., Ahmed, K., Mostafa, R., Naik, G. R., Darvishi, S., et al. (2021). Progress in brain computer interface: challenges and opportunities. Front. Syst. Neurosci. 15, 20. doi: 10.3389/fnsys.2021.578875
Saibene, A., Caglioni, M., Corchs, S., and Gasparini, F. (2023). EEG-based BCIS on motor imagery paradigm using wearable technologies: a systematic review. Sensors 23:2798. doi: 10.3390/s23052798
Schirrmeister, R. T., Springenberg, J. T., Fiederer, L. D. J., Glasstetter, M., Eggensperger, K., Tangermann, M., et al. (2017). Deep learning with convolutional neural networks for EEG decoding and visualization. Hum. Brain Mapp. 38, 5391–5420. doi: 10.1002/hbm.23730
Shi, Y. X., Li, Y. H., and Koike, Y. (2023). Sparse logistic regression-based EEG channel optimization algorithm for improved universality across participants. Bioengineering 10:664. doi: 10.3390/bioengineering10060664
Strypsteen, T., and Bertrand, A. (2021). End-to-end learnable EEG channel selection for deep neural networks with gumbel-softmax. J. Neural Eng. 18, 12. doi: 10.1088/1741-2552/ac115d
Tam, W. K., Ke, Z., Tong, K. Y., and Ieee (2011). “Performance of common spatial pattern under a smaller set of EEG electrodes in brain-computer interface on chronic stroke patients: a multi-session dataset study,” in 33rd Annual International Conference of the IEEE Engineering-in-Medicine-and-Biology-Society (EMBS), IEEE Engineering in Medicine and Biology Society Conference Proceedings (New York, NY: IEEE), 6344–6347.
Tang, C., Gao, T. Y., Li, Y. H., and Chen, B. D. (2022). EEG channel selection based on sequential backward floating search for motor imagery classification. Front. Neurosci. 16, 1045851. doi: 10.3389/fnins.2022.1045851
Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., and Hu, Q. (2020). “ECA-net: efficient channel attention for deep convolutional neural networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (Seattle, WA), 11534–11542.
Zhang, H., Zhao, X., Wu, Z. X., Sun, B., and Li, T. (2021). Motor imagery recognition with automatic eeg channel selection and deep learning. J. Neural Eng. 18, 12. doi: 10.1088/1741-2552/abca16
Keywords: brain-computer interface, motor imagery, channel selection, deep learning, attention mechanism
Citation: Tong L, Qian Y, Peng L, Wang C and Hou Z-G (2023) A learnable EEG channel selection method for MI-BCI using efficient channel attention. Front. Neurosci. 17:1276067. doi: 10.3389/fnins.2023.1276067
Received: 11 August 2023; Accepted: 05 October 2023;
Published: 20 October 2023.
Edited by:
Zhao Lv, Anhui University, ChinaReviewed by:
Hao Jia, University of Vic – Central University of Catalonia, SpainYuanhao Li, Tokyo Institute of Technology, Japan
Copyright © 2023 Tong, Qian, Peng, Wang and Hou. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Chen Wang, d2FuZ2NoZW4yMDE2JiN4MDAwNDA7aWEuYWMuY24=; Liang Peng, bGlhbmcucGVuZyYjeDAwMDQwO2lhLmFjLmNu