Skip to main content

ORIGINAL RESEARCH article

Front. Comput. Neurosci., 25 January 2024

Positional multi-length and mutual-attention network for epileptic seizure classification

\r\nGuokai ZhangGuokai Zhang1Aiming ZhangAiming Zhang1Huan LiuHuan Liu2Jihao Luo
Jihao Luo3*Jianqing Chen
Jianqing Chen4*
  • 1School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai, China
  • 2Department of Hematology, Affiliated Qingdao Central Hospital of Qingdao University, Qingdao Cancer Hospital, Qingdao, China
  • 3School of Computing, National University of Singapore, Singapore, Singapore
  • 4Department of Otolaryngology, Head and Neck Surgery, Shanghai Ninth People's Hospital, Affiliated to Shanghai Jiaotong University School of Medicine, Shanghai, China

The automatic classification of epilepsy electroencephalogram (EEG) signals plays a crucial role in diagnosing neurological diseases. Although promising results have been achieved by deep learning methods in this task, capturing the minute abnormal characteristics, contextual information, and long dependencies of EEG signals remains a challenge. To address this challenge, a positional multi-length and mutual-attention (PMM) network is proposed for the automatic classification of epilepsy EEG signals. The PMM network incorporates a positional feature encoding process that extracts minute abnormal characteristics from the EEG signal and utilizes a multi-length feature learning process with a hierarchy residual dilated LSTM (RDLSTM) to capture long contextual dependencies. Furthermore, a mutual-attention feature reinforcement process is employed to learn the global and relative feature dependencies and enhance the discriminative abilities of the network. To validate the effectiveness PMM network, we conduct extensive experiments on the public dataset and the experimental results demonstrate the superior performance of the PMM network compared to state-of-the-art methods.

1 Introduction

Epilepsy is a prevalent neurological disease worldwide, affecting individuals' cognitive abilities and presenting risks of sudden falls or fatality (Rajinikanth et al., 2022). To mitigate epilepsy risks, the analysis of Electroencephalography (EEG) signals is the most effective approach to identify real-time neural disorder activity. However, EEG events often exhibit subtle amplitude variations, and manual detection of EEG signals is time-consuming, prone to errors, and requires specialized expertise. Thus, the significance of automatic EEG diagnosis lies in its capacity to analyze EEG signals with efficiency and accuracy, facilitating the timely detection of epilepsy (Xin et al., 2022; Wu et al., 2023).

Early studies on automatic EEG diagnosis focused on using hand-engineered low-level features, such as spectral, temporal, low-frequency, and high-frequency features, to achieve automatic classification of EEG signals (Liu et al., 2022). For instance, Lemm et al. (2005) applied spatio-spectral filters to mitigate the impact of noisy, non-stationary, and contaminated information in EEG signals, thereby improving classification performance. Meng et al. (2014) proposed a novel approach that learned spatial and spectral features and optimized the loss function by calculating the mutual information between the learned spectral features and class labels. Additionally, to assess the impact of different frequency sub-bands on EEG classification accuracy (Tsipouras, 2019), multiple sub-bands were combined as feature vectors. In another work, Qi et al. (2015) introduced regularized spatio-temporal filtering to classify EEG signals using supervised optimization algorithms. Jrad (2016) utilized high-frequency oscillations to extract relevant features, which were then input into a support vector machine for classifying different EEG events. Furthermore, Gao et al. (2019) developed a multi-scale information analysis model that utilized high-frequency EEG oscillations to recognize emotional states. Despite the success of these approaches, the subjective selection of hand-engineered features typically requires domain knowledge and may not capture the full range of characteristics present in input EEG signals.

Recently, with the great success of the convolutional neural network (CNN) on a broad array of medical image analysis, a large body of work in this area has been considered (Chen et al., 2021, 2023). Compared with the traditional hand-engineered methods, the CNN-based ones have the advantage of extracting more complicated and discriminative characteristics from the medical image. For instance, Zheng and Lu (2015) adopted the deep belief networks (DBNs) as the main detection architecture to train differential features for the automatic detection of the seizure. Regarding temporal features, Kasabov and Capecci (2015) designed a spiking neural network architecture that extracted spatio-temporal features for detecting and interpreting EEG signals. In Liu M. et al. (2020), it used the pre-trained CNN models to extract deep features and then adopted the cartesian K-means algorithm to conduct the semi-supervised learning on the EEG data. Furthermore, the unsupervised learning method (Chai et al., 2016) which combined the auto-encoder network with a subspace alignment solution into a unified framework was developed for analyzing the EEG data. Some other works, such as Qiu et al. (2018) and Liu J. et al. (2020) utilized the sparse autoencoder with different classifiers to jointly detect the seizure signal. Moreover, to improve the performance of the classification model, Yuan et al. (2018a) proposed a multi-view CNN model which aimed to learn the brain seizure from input multi-channel signals. Similarly, in Yuan et al. (2018b), it further developed a novel channel-aware attention network for multi-channel EEG seizure detection by using CNNs. Hossain et al. (2019) proposed a model to extract the spectral, temporal features and then input them to the classifier for EEG seizure classification. For learning the multi-scale features from the EEG, Zhang et al. (2020) designed a multi-scale non-local (MNL) network with two special layers to achieve promising classification results of the seizure. Additionally, some other works (Aliyu and Lim, 2021; Hussain et al., 2021; Saichand, 2021) adopted the long short-term memory (LSTM) to overcome the vanishing gradient problem of the recurrent neural network and boost the feature extraction ability of the EEG signal data.

Despite the promising results shown by CNN-based methods for EEG signal classification, three major challenges still need to be addressed. Firstly, seizures in EEG signals often exhibit subtle abnormal characteristics that can pose challenges for feature extraction, potentially impacting the performance of classification models. Secondly, the extraction of long contextual dependencies is crucial for effective EEG signal classification, but the use of LSTM for this purpose is impeded by limited receptive fields, which compromises their ability to capture necessary contextual information. Lastly, it is worth noting that previous works have paid relatively less attention to the incorporation of global relative dependencies in EEG signal analysis, which could offer valuable discriminative information crucial for improving classification accuracy. To tackle these challenges, we propose a novel approach called the positional multi-length and mutual-attention (PMM) network. The PMM network comprises three main processes: positional feature encoding, multi-length feature learning, and mutual-attention feature reinforcement. In the positional feature encoding process, minute abnormal characteristics from the shallow layers of the network are captured through the utilization of residual positional attention. This facilitates the PMM network in focusing on and extracting crucial information associated with those characteristics. The multi-length feature learning process employs a stacking of hierarchical residual dilated (RD) LSTMs to acquire long contextual dependencies within the EEG signal. By doing so, the network becomes adept at capturing temporal patterns across various time scales and effectively modeling the relationships between distant time steps. To further fortify the features, a mutual-attention feature reinforcement process is introduced. This process delves into both the global discriminative and relative dependencies present in the EEG signal. It selectively enhances informative features while simultaneously suppressing irrelevant ones, thereby enhancing the overall discriminative power of the network. Incorporating these three processes into the PMM network enables it to capture minute abnormal characteristics, long contextual dependencies, and global discriminative and relative dependencies simultaneously, resulting in a significant improvement in classifying EEG signals. Overall, the main contributions of this paper can be summarized as follows:

(1) A novel PMM network is proposed for the automatic classification of epilepsy seizures from EEG signals, with the incorporation of positional feature encoding to improve the extraction of minute abnormal characteristics from the EEG signals.

(2) In the proposed multi-length feature learning process, hierarchical RDLSTMs are used to capture long contextual dependencies from the EEG signal. Additionally, mutual-attention feature reinforcement is employed to jointly explore global discriminative features and relative dependencies simultaneously.

(3) Extensive experiments are conducted on the publicly available dataset. The results of the comparative analysis demonstrate that competitive performance is achieved by our proposed PMM network when compared to other state-of-the-art methods.

The remainder of the paper is organized as follows. Section 2 provides an introduction to the main method used in our proposed network. In Section 3, we provide a detailed description of the experimental data utilized in our study. Section 4 covers the implementation details, evaluation metrics, and a series of experiments conducted to evaluate the performance of our proposed approach. Finally, in Section 6, we summarize the findings of our study and provide a conclusion.

2 Method

As depicted in Figure 1, the input EEG signal undergoes an initial pre-processing step to obtain the processed signal. This processed signal is then fed into the positional feature encoding module, which extracts subtle abnormal characteristics from the shallow layers. Subsequently, multi-length feature learning and mutual-attention feature reinforcement are employed to enhance the classification of the PMM network. Finally, the reinforced feature is delivered to the dense layer with the softmax activation function to generate the predicted result. To provide a more specific overview, the main architecture of the proposed PMM network is illustrated in Figure 2. Initially, the input processed signal undergoes a positional feature encoding process, employing multiple positional feature encoding blocks (PFEBlocks) to capture minute abnormal features. Next, a multi-length feature learning process utilizing stacked RDLSTMs is employed to capture long contextual dependencies from the EEG signal. Furthermore, the network includes a mutual-attention feature reinforcement mechanism, which enables the extraction of both global discriminative features and relative dependencies, enhancing the network's overall capability in these aspects. In the following subsections, we will provide more detailed descriptions of positional feature encoding, multi-length feature learning, and mutual-attention feature reinforcement.

Figure 1
www.frontiersin.org

Figure 1. The PMM network systematically analyzes EEG signals through pre-process. The processed signal then undergoes positional feature encoding, multi-length feature learning, and mutual-attention feature reinforcement. Finally, the reinforced features are processed in a dense layer with softmax activation to generate the predicted result.

Figure 2
www.frontiersin.org

Figure 2. Overview of the PMM network, the PFEBlocks capture minute abnormal features, while the stacked RDLSTMs capture long contextual dependencies from the processed EEG signal. Additionally, the mutual-attention feature reinforcement further enhances the network's capability by extracting global discriminative features and relative dependencies.

2.1 Positional feature encoding

In the shallow layers of the network, the extracted feature map contains crucial details of the EEG signal that are vital for accurate EEG classification. Inspired by the structure of the residual block (Figure 3A), we incorporate a positional feature encoding block (Figure 3B) during feature encoding to automatically extract informative detail representations from the EEG signal. Considering the input of the positional feature encoding as Fe, it first passes through the 1D convolution layer, which can be defined as:

Xo=RW*Fe+b    (1)

where Xo is the output feature vector, R denotes the receptive field, W and b is the weighted parameter and bias, respectively. After that, a randomized leaky rectified linear unit (RReLU) nonlinear activation function is employed, which is formulated as:

y={xif x0axif x<0    (2)

where x is the input value, a represents a random number gained from the uniform distribution U(p, q), and it is given as:

a~U(p,q),p<q  and  p,q[0,1)    (3)

Next, an extra 1D convolution layer is adopted to further refine the output feature from the RReLU activation function, resulting in Xo. In contrast to the traditional residual block, the proposed positional feature encoding block applies a sigmoid operation on Fe to obtain the position weight matrix Wpos. The formulation of Wpos is defined as follows:

Wpos=eFeeFe+1    (4)

Afterward, we multiply Wpos with Xo, and then add it to Fe by a residual connection, therefore the final output feature map of Fo is formulated as:

Fo=Wpos*Xo+Fe    (5)

Subsequently, the positional feature encoding block is followed by a 1D max-pooling layer, which downsamples the resolution of Fo, and the resulting feature map is then further refined and enhanced through multi-length feature learning.

Figure 3
www.frontiersin.org

Figure 3. The structure of residual block (A) and positional feature encoding block (B).

2.2 Multi-length feature learning

To leverage the valuable information provided by the dependencies among multi-length features, the learned features obtained from the positional feature encoding are fed into the multi-length feature learning process. To be more specific, we denote the output features from the positional feature encoding process as Fpos, for learning dependencies of the input signal features, we use the LSTM as the primary feature extraction unit for learning high-level representations, as illustrated in Figure 4. Notably, considering the shortcoming of LSTM of disappearing and losing the information of cell state (Schoene et al., 2020), we further add the residual dilation (Chang et al., 2017) to the LSTM for learning the multi-length and long sequence dependencies as shown in Figure 5. Here, denote the cell state, hidden state, and input of LSTM at time t as ct, ht, xt, respectively. Thus, the output of the block input zt is calculated as:

zt=g(Wzxt+Rzht-1+bz)    (6)

where Wz, Rz, bz is the input weight, recurrent weight, and bias weight, respectively. The function of g(·) is the tanh activation function.

Figure 4
www.frontiersin.org

Figure 4. The structure of LSTM block.

Figure 5
www.frontiersin.org

Figure 5. The structure of RDLSTM block.

Then, the input gate, forget gate and output gate could be formulated as:

it=σ(Wixt+Riht-1+bi)    (7)
ct=ftct-1+itzt    (8)
ft=σ(Wfxt+Rfht-1+bf)    (9)
ot=σ(Woxt+Roht-1+bo)    (10)
ht=g(ct)ot    (11)

here, the Wi, Wf, Wo and bi, bf, bo are the corresponding inputs and bias weights of LSTM. The σ(·) and ⊙ represent the sigmoid and element-wise multiplication, respectively. In the RDLSTM, instead of using the previous cell state ct−1 and hidden state ht−1, it takes in the cell state ct−d and hidden state ht−d, where the dilation rate d exponentially increases the receptive field of the LSTM. By incorporating these distant past states, the RDLSTM can capture long dependencies from the EEG signals, allowing for a more comprehensive understanding of the input sequence. Mathematically, we denote zt, ct, ht, it, ft, ot as the outputs of block input, cell state, hidden state, input gate, forget gate, and output gate, respectively. Then, the dilated LSTM could be defined as:

zt=g(Wzht(l-1)+Rzht-d+bz)    (12)
ct=ftct-d+itzt    (13)
ht=ot(g(ct)+ht(l-1))    (14)
it=σ(Wiht(l-1)+Riht-d+bi)    (15)
ft=σ(Wfht(l-1)+Rfht-d+bf)    (16)
ot=σ(Woht(l-1)+Roht-d+bo)    (17)

where Wz, Wi, Wf, Wo are the input weights, Rz, Ri, Rf, Ro denote the recurrent weights, and bz, bi, bf, bo are the bias. The ht(l-1) represents the hidden state at the (l − 1)-th layer and is used to create a shortcut connection with the current LSTM cell to mitigate the issue of gradient vanishing. Additionally, hierarchically stacking the RDLSTMs enables the network to capture effective long dependencies among multi-length features across different layers. Therefore, various dilated rates of 1, 2, 4 are utilized to increase the receptive field of the network exponentially, enabling it to capture both local and global information. Following the processing in the RDLSTMs, the learned features are fed into the mutual-attention feature reinforcement process to extract global context information and further enhance the network's understanding of the input feature.

2.3 Mutual-attention feature reinforcement

Previous research has demonstrated that attention-based learning is effective in encoding discriminative features and capturing global dependencies (Vaswani et al., 2017; Zhang et al., 2020). Building on these findings, we propose to enhance the feature learning capability by incorporating a mutual-attention feature reinforcement after the multi-length feature learning process (as shown in Figure 6). Formally, let the output feature maps of the RDLSTM with dilation rates q, k, v be denoted as Fq, Fk, Fv (qkv). In order to standardize the features to the same value domain, a batch normalization layer N(·). Ioffe and Szegedy (2015) is initially applied to Fq, Fk, Fv, yielding Fq, Fk, Fv. Subsequently, these features with different value domains are separately passed through three linear layers and combined as inputs to the mutual-attention module. The formulation of this module can be expressed as:

Q=W(query)Fq+b(query)    (18)
K=W(key)Fk+b(key)    (19)
V=W(value)Fv+b(value)    (20)

where Q, K, V is the corresponding query, key, and value of Fq, Fk, Fv, Wq(query), Wk(key), and Wv(value) is the respective learned parameters of the linear operation, bq(query), bk(key), and bv(value) is the bias, separately. Then, the mutual scaled-attention Aqk,Aqv,Avk could be calculated by:

Aqk=σ(QKS)V    (21)
Aqv=σ(QVS)K    (22)
Avk=σ(VKS)Q    (23)

where S is the scaling parameter, ⊺ denotes the transpose, and the softmax function is performed on the gained attention values to normalize the attention values into probability distributions, which is defined as:

σ(z)d=ezjj=1ezj    (24)

Furthermore, to exploit more discriminative and global representations, we employ the multi-head attention (Vaswani et al., 2017) which calculates the mutual-attention operations for T times. Thus, the output of multi-head attention Ad(multi) is formulated as:

Aqk(multi)=Wqk(multi)Concat(Aqk,1,Aqk,2,,Aqk,T)    (25)
Aqv(multi)=Wqv(multi)Concat(Aqv,1,Aqv,2,,Aqv,T)    (26)
Avk(multi)=Wvk(multi)Concat(Avk,1,Avk,2,,Avk,T)    (27)

where Wqk(multi), Wqv(multi), Wvk(multi) are the weight matrices of the linear combination.

Figure 6
www.frontiersin.org

Figure 6. The structure of the mutual-attention feature reinforcement.

The operation of Concat(·) represents the concatenation of the input features. Finally, the output from mutual-attention feature reinforcement Fat is given as:

Fat=Conv(Concat(Aqk(multi),Aqv(multi),Avk(multi)))    (28)

2.4 Classification of EEG data

After the feature extraction process, the extracted features are passed into a softmax layer to generate prediction probabilities for different classes. Mathematically, the training data can be represented as {(s(1), y(1)), (s(2), y(2)), ⋯ , (s(N), y(N))}, where sN ∈ ℝC denotes the input features, y(N) ∈ {1, 2, ⋯ , C} represents the class label, and C is the total number of class labels. Therefore, the mapping function of softmax could be given as:

hθ(s)=(P(y=1s;θ)P(y=2s;θ)P(y=Cs;θ))=1j=1Kexp(θjTs)(exp(θ1Ts)exp(θ2Ts)exp(θCTs))    (29)

where θ is the learned parameters of the softmax. To optimize the network, we use the cross-entropy J(·) as the loss function, which can be defined as:

J(w)=-1Nc=1C[yclogy^c+(1-y^c)log(1-y^c)]    (30)

where yc is the true class label, and y^c denotes the predicted class label. Overall, the whole process of the proposed PMM Network could be illustrated in Algorithm 1.

Algorithm 1
www.frontiersin.org

Algorithm 1. Epileptic Seizure Classification of PMM Network.

3 Data description

We evaluate our approach using the Bonn EEG dataset, initially reported in Andrzejak (2001). This dataset consists of five subsets: Set A, B, C, D, and E. Each subset contains 100 EEG channels and has a duration of 23.6 seconds. Subsets A and B were collected from healthy subjects, with recordings taken during both eyes open and closed conditions. Subsets C, D, and E correspond to different locations in epileptic subjects. Subset C represents recordings from the hippocampal formation, Subset D records the epileptogenic zone, and Subset E captures signals during seizure activity. It is important to note that the signals in subset C and D were recorded during seizure-free intervals, while subset E was captured during seizure activity. For simplicity, the eye movements in dataset A and B were not considered in our evaluation.

Moreover, the UCI-EEG Recognition dataset (Wu and Fokoue, 2017) is also used for the detection of epileptic seizures. It consists of five distinct groups, each consisting of 100 single-channel EEG signals. Each EEG file corresponds to a 23.6-s recording of brain activity, which is sampled into 4097 data points. Therefore, the dataset comprises a total of 500 subjects, with each subject's data containing 4,097 data points. Additionally, the EEG samples in this dataset are further divided into 23 data chunks, with each chunk containing 178 data points. Overall, the dataset contains 11,500 time-series EEG signal data samples from the 500 subjects. In the EEG recognition dataset, Class 1 represents the state of epileptic seizure, while Classes 2–5 represent normal healthy states. This dataset facilitates a binary classification task aimed at distinguishing between the combined normal states (Classes 2–5) and the seizure condition (Class 1).

4 Results

4.1 Implementation details

The PMM network is implemented using the PyTorch deep learning framework, and cross-entropy is adopted as the loss function. The optimizer used is Adam, which helps in the convergence of the network. The initial learning rate is set at 0.0003 and is decayed by a factor of 0.001 after each epoch. To accelerate the training process, a GTX 1080 GPU is used. Additionally, 10-fold cross-validation is carried out to assess the performance of the model.

4.2 Evaluation metrics

The performance of the experiment is evaluated using several commonly used performance metrics, including accuracy, precision, sensitivity, specificity, and F1-score.

Accuracy is defined as the ratio of the number of correctly predicted samples to the total number of predicted samples, which is defined as

Accuracy=TP+TNTP+TN+FP+FN    (31)

Precision refers to the ratio of the number of correctly predicted positive samples to the total number of predicted positive samples, which is given as

Precision=TPTP+FP    (32)

Sensitivity measures the proportion of positives that are correctly identified, which is defined as

Sensitivity=TPTP+FN    (33)

Specificity measures the proportion of negatives that are correctly identified, which is defined as

Specif icity=TNFP+TN    (34)

F1 score is the harmonic mean of precision and sensitivity, which is defined as

F1-score=2×Precision×SensitivityPrecision+Sensitivity    (35)

Among all the equations presented above, the term TP (true positives) represents the number of EEG data samples that are abnormal and correctly identified as abnormal. Similarly, TN (true negatives) represents the number of EEG data samples that are normal and correctly identified as normal. FP (false positives) refers to the number of normal EEG data samples that are incorrectly predicted as abnormal, and FN (false negatives) refers to the number of abnormal EEG data samples that are incorrectly predicted as normal.

To ensure a comprehensive evaluation of the system, a 10-fold cross-validation approach is applied. During each iteration, one fold is used for testing the model, while the remaining nine folds are used for training. This process is repeated ten times, with each fold used as the test set once. The average values of accuracy, sensitivity, and specificity are then collected from the ten-fold cross-validation, providing an average performance measurement of the system across different categories of data.

4.3 The performance of double classes classification

In this section, the performance of the double class classification is evaluated on Bonn dataset. Table 1 compares the performance of different combinations of double classes, including A-E, B-E, C-E, D-E, AB-E, AC-E, AD-E, BC-E, BD-E, CD-E, ABC-E, ABD-E, BCD-E, and ABCD-E on Bonn dataset. Among these combinations, the highest performance is achieved in the A-E class classification with an accuracy of 99.95%, while the most challenging classification task is D-E with an accuracy of 98.79%. The experimental results reveal a significant difference between the eyes open classes and the seizure epileptic classes, resulting in a higher classification performance. Furthermore, these results demonstrate that our proposed PMM network performs well on different double class classification tasks.

Table 1
www.frontiersin.org

Table 1. The overall performance of double classes classification on Bonn dataset.

4.4 The performance of multiple classes classification

We further evaluate the performance of the proposed PMM network on multiple class classification using Bonn dataset. We compare the combinations of classes including A-C-E, A-D-E, B-C-E, B-D-E, AB-CD-E, and A-B-C-D-E separately, and present the results in Table 2. The results clearly indicate that the B-D-E combination achieves the best performance, with an accuracy of 98.73%, sensitivity of 97.22%, specificity of 98.61%, precision of 97.32%, and F1-score of 97.52%. On the other hand, the A-B-C-D-E combination, consisting of five classes, shows the lowest performance. This showcases the increasing difficulty of multiple class classification tasks as the number of classes increases.

Table 2
www.frontiersin.org

Table 2. The overall performance of multiple classes classification on Bonn dataset.

4.5 The effectiveness of different components

In this section, we conduct extensive experiments to validate the effectiveness of each proposed component in the A-B-C-D-E classes combination classification task using Bonn dataset. We refer to the positional feature encoding block, multi-length feature learning, mutual-attention feature reinforcement, and residual dilation with LSTM as PFEBlock, MFL, MFR, and RDLSTM, respectively. The model without any proposed module is defined as “Original”. Table 3 illustrates the results of these experiments. It can be observed that integrating any of the processes, i.e., PFEBlock, MFL, or MFR, leads to improved classification performance compared to RDLSTM. This confirms the effectiveness of each proposed process in enhancing the overall classification performance. Additionally, we find that adding the MFL could gain the best performance, which further demonstrate that the dependencies among multi-length features play vital importance in this task.

Table 3
www.frontiersin.org

Table 3. The effectiveness of different components.

4.6 Influence of different dilated rates

The optimal dilated rates in the Dilated LSTM network play a crucial role in achieving improved performance. Through our experiments, we conducte tests to determine the best rates based on various metrics. The results, as shown in Table 4, indicate that larger dilated rates lead to enhanced performance when using a single dilated rate or two different rates. The higher dilated rates offer better performance because they allow the network to capture a broader range of information from the input data. By increasing the dilation rate, the network can expand its receptive field and consider a wider context, resulting in more accurate and informed predictions. Comparing the dilated rate sequences “1, 2, 4, 8” and “1, 2, 4”, we find that the latter sequence achieves superior performance. This is because the rates “1, 2, 4” strike a balance between capturing local patterns and incorporating global relationships within the data. On the other hand, including the rate “8” in the first sequence potentially introduces noise or redundant information, which may degrade the model's performance.

Table 4
www.frontiersin.org

Table 4. The influence of different dilated rates on Bonn dataset.

4.7 Compare with other classification methods

To evaluate the performance of the proposed network, we first compare it with other classification methods on Bonn dataset, especially those based on CNN, in various classification tasks. For consistency, we use fixed combinations of EEG classes, including ABCD-E, AB-CD-E, A-B-C-D-E, A-E, AC-E, C-E, A-D-E, D-E, A-D, B-E, and B-C-D. Tables 5, 6 present the results of the comparison. It can be observed that our proposed method achieves competitive performance in most classification tasks when compared to other methods, particularly in the double classes classification task. This demonstrates the effectiveness of our proposed method in handling and accurately classifying EEG signals for various classification tasks. In addition, we conduct experiments on binary classification using the UCI-EEG dataset, and the experimental results are illustrated in Table 7. The results demonstrate that our proposed method achieved the highest classification accuracy compared to other classification methods.

Table 5
www.frontiersin.org

Table 5. Comparison with other methods (a).

Table 6
www.frontiersin.org

Table 6. Comparison with other methods (b).

Table 7
www.frontiersin.org

Table 7. Comparison with other methods on UCI-EEG dataset.

5 Discussion

From a technical perspective, our paper introduces the PMM network as a potentional solution to address the challenges associated with learning minute abnormal characteristics and modeling long dependencies in EEG signals. We employ positional feature encoding to enhance the network's detection of subtle abnormalities, leveraging temporal position information. Additionally, our proposed multi-length feature learning enables the network to extract features at different scales, capturing short-term and long-term dependencies in the EEG signals. Moreover, incorporating the mutual-attention feature reinforcement mechanism enhances the network's ability to identify relevant spatial and temporal dependencies, allowing it to distinguish abnormal patterns from background activity more effectively. These advancements collectively contribute to the PMM network's potential in clinical applications and EEG signal analysis by providing a more comprehensive and accurate approach for capturing small abnormal characteristics, modeling long dependencies, and improving attention mechanisms.

From a clinical perspective, our proposed PMM network offers improved advancements in EEG signal analysis that have the potential to benefit clinical practice. It effectively captures minute abnormal characteristics often associated with neurological disorders, allowing for precise identification even within complex EEG patterns. By modeling long dependencies and incorporating multi-length feature learning, the network provides a comprehensive understanding of the underlying abnormal processes that evolve over time. The mutual-attention feature reinforcement mechanism further enhances specificity in detecting abnormal patterns, which is crucial for accurate diagnosis and informed decision-making in patient management. While these improvements hold promise for clinical practice, it is important to note that further research and evaluation are needed. Extensive testing with diverse datasets is necessary to validate the network's performance across various EEG classification tasks encountered in real-world clinical settings. Such validation is crucial before the network's potential can be fully realized and integrated into routine clinical workflows.

6 Conclusion

In this paper, we introduced the PMM network to address the challenges related to learning minute abnormal characteristics and long dependencies in EEG signals. Our proposed approach effectively captures the minute abnormal characteristics through positional feature encoding and improves the modeling of long dependencies with multi-length feature learning and mutual-attention feature reinforcement. Experimental evaluations on the publicly available dataset demonstrated that the PMM network achieves competitive performance compared to other state-of-the-art methods. One limitation of this study is that the proposed network was only evaluated on the limited dataset, which may not cover the full spectrum of EEG classification tasks. Therefore, in future work, we aim to extend our network to more diverse time-series datasets to further validate its effectiveness and generalizability. Moreover, while this approach proves beneficial for capturing subtle abnormal characteristics in EEG signals, it may be sensitive to variations in signal alignment and time-dependent patterns. Different EEG recording setups or variations in patient-specific factors could introduce spatial and temporal misalignments, potentially affecting the network's performance. Therefore, future research should focus on developing more robust techniques for positional feature encoding that can adapt to different recording setups and account for these variations, ensuring the network's stability and reliability across diverse EEG datasets. By addressing this limitation, we can enhance the network's applicability and strengthen its performance in real-world clinical scenarios.

Data availability statement

Publicly available datasets were analyzed in this study. This data can be found here: https://www.upf.edu/web/ntsa/downloads.

Author contributions

GZ: Writing—original draft, Writing—review & editing. AZ: Writing—review & editing. HL: Writing—review & editing. JL: Writing—review & editing. JC: Writing—review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported by the General Program of National Natural Science Foundation of China (NSFC) under Grant 62102259, the Shanghai Sailing Program (21YF1431600), and the Interdisciplinary Program of Shanghai Jiao Tong University (YG2019QNB12).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Ahmad, I., Yao, C., Li, L., et al. (2024). An efficient feature selection and explainable classification method for EEG-based epileptic seizure detection. J. Inform. Commun. Technol. 80, 103654. doi: 10.1016/j.jisa.2023.103654

Crossref Full Text | Google Scholar

Al-Hadeethi, H., Abdulla, S., Diykh, M., and Green, J. H. (2021). Determinant of covariance matrix model coupled with adaboost classification algorithm for EEG seizure detection. Diagnostics 12, 74. doi: 10.3390/diagnostics12010074

PubMed Abstract | Crossref Full Text | Google Scholar

Aliyu, I., and Lim, C. G. (2021). Selection of optimal wavelet features for epileptic EEG signal classification with LSTM. Neural Comp. Appl. 2021, 1–21. doi: 10.1007/s00521-020-05666-0

Crossref Full Text | Google Scholar

Andrzejak, R. G., Lehnertz, K., Mormann, F., Rieke, C., David, P., and Elger, C. E. (2001). Indications of nonlinear deterministic and finite-dimensional structures in time series of brain electrical activity: dependence on recording region and brain state. Phys. Rev. E 64, 061907. doi: 10.1103/PhysRevE.64.061907

PubMed Abstract | Crossref Full Text | Google Scholar

Chai, X., Wang, Q., Zhao, Y., Liu, X., Bai, O., and Li, Y. (2016). Unsupervised domain adaptation techniques based on auto-encoder for non-stationary EEG-based emotion recognition. Comp. Biol. Med. 79, 205–214. doi: 10.1016/j.compbiomed.2016.10.019

PubMed Abstract | Crossref Full Text | Google Scholar

Chang, S., Zhang, Y., Han, W., Yu, M., Guo, X., Tan, W., et al. (2017). Dilated recurrent neural networks. Adv. Neural Inf. Process Syst. 30, 77–87.

Google Scholar

Chen, J., Guo, Z., Xu, X., Zhang, L., Teng, Y., Chen, Y., et al. (2023). Robust deep learning framework based on spectrograms for heart sound classification. IEEE/ACM Trans Comput Biol Bioinform. 22, 433. doi: 10.1109/TCBB.2023.3247433

PubMed Abstract | Crossref Full Text | Google Scholar

Chen, J., Sun, S., Zhang, L., Yang, B., and Wang, W. (2021). Compressed sensing framework for heart sound acquisition in internet of medical things. IEEE Trans. Indust. Informat. 18, 2000–2009. doi: 10.1109/TII.2021.3088465

Crossref Full Text | Google Scholar

Das, P., Manikandan, M. S., and Ramkumar, B. (2018). “Detection of epileptic seizure event in EEG signals using variational mode decomposition and mode spectral entropy,” in 2018 IEEE 13th International Conference on Industrial and Information Systems (ICIIS). Rupnagar: IEEE, 42–47. doi: 10.1109/ICIINFS.2018.8721426

Crossref Full Text | Google Scholar

Deng, Z., Xu, P., Xie, L., Choi, K.-S., and Wang, S. (2018). Transductive joint-knowledge-transfer TSK FS for recognition of epileptic EEG signals. IEEE Trans. Neural Syst.d Rehabilit. Eng. 26, 1481–1494. doi: 10.1109/TNSRE.2018.2850308

PubMed Abstract | Crossref Full Text | Google Scholar

Gao, Z., Cui, X., Wan, W., et al. (2019). Recognition of emotional states using multiscale information analysis of high frequency EEG oscillations. Entropy 21, 609. doi: 10.3390/e21060609

PubMed Abstract | Crossref Full Text | Google Scholar

Hossain, M. S., Amin, S. U., Alsulaiman, M., and Muhammad, G. (2019). Applying deep learning for epilepsy seizure detection and brain mapping visualization. ACM Trans. Multimedia Comp. Commun. Applicat. (TOMM). 15, 1–17. doi: 10.1145/3241056

Crossref Full Text | Google Scholar

Hussain, W., Sadiq, M. T., Siuly, S., and Rehman, A. (2021). Epileptic seizure detection using 1 D-convolutional long short-term memory neural networks. Appl. Acoust. 177, 107941. doi: 10.1016/j.apacoust.2021.107941

Crossref Full Text | Google Scholar

Ioffe, S., and Szegedy, C. (2015). “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” in International Conference on Machine Learning (New York, NY: PMLR), 448–456.

Google Scholar

Jrad, N., Kachenoura, A., Merlet, I., Bartolomei, F., Nica, A., and Biraben, A. (2016). Automatic detection and classification of high-frequency oscillations in depth-EEG signals. IEEE Trans. Biomed. Eng, 64, 2230–2240. doi: 10.1109/TBME.2016.2633391

PubMed Abstract | Crossref Full Text | Google Scholar

Kasabov, N., and Capecci, E. (2015). Spiking neural network methodology for modelling, classification and understanding of EEG spatio-temporal data measuring cognitive processes. Inform. Sci. 294, 565–575. doi: 10.1016/j.ins.2014.06.028

Crossref Full Text | Google Scholar

Kavitha, K. V. N., Ashok, S., Imoize, A. L., Ojo, S., Selvan, K., Ahanger, T., et al. (2022). On the use of wavelet domain and machine learning for the analysis of epileptic seizure detection from EEG signals. J. Healthc Eng. 2022, 8928021. doi: 10.1155/2022/8928021

PubMed Abstract | Crossref Full Text | Google Scholar

Lemm, S., Blankertz, B., Curio, G., and Muller, K. (2005). Spatio-spectral filters for improving the classification of single trial EEG. IEEE Trans. Biomed. Eng. 52, 1541–1548. doi: 10.1109/TBME.2005.851521

PubMed Abstract | Crossref Full Text | Google Scholar

Li, Y., Liu, Y., Cui, W.-G., Guo, Y.-Z., Huang, H., and Hu, Z.-Y. (2020). Epileptic seizure detection in EEG signals using a unified temporal-spectral squeeze-and-excitation network. IEEE Trans. Neural Syst. Rehabilitat. Eng. 28, 782–794. doi: 10.1109/TNSRE.2020.2973434

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, J., Wu, G., Luo, Y., Qiu, S., Yang, S., Li, W., et al. (2020). EEG-based Emotion Classification Using Deep Neural Network and Sparse Autoencoder. Front. Syst. Neurosci. 14, 43. doi: 10.3389/fnsys.2020.00043

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, M., Zhou, M., Zhang, T., et al. (2020). Semi-supervised learning quantization algorithm with deep features for motor imagery EEG Recognition in smart healthcare application. Appl. Soft Comp. 89, 106071. doi: 10.1016/j.asoc.2020.106071

Crossref Full Text | Google Scholar

Liu, Q., Wang, Z., Dong, H., and Jiang, C. (2022). Remote estimation for energy harvesting systems under multiplicative noises: a binary encoding scheme with probabilistic bit flips. IEEE Trans. Automatic Control. 68, 3170540. doi: 10.1109/TAC.2022.3170540

Crossref Full Text | Google Scholar

Liu, Z., Zhu, B., Hu, M., Deng, Z., and Zhang, J. (2023). Revised tunable q-factor wavelet transform for EEG-based epileptic seizure detection. IEEE Trans. Neural Syst. Rehabilit. Eng. 31, 1707–1720. doi: 10.1109/TNSRE.2023.3257306

PubMed Abstract | Crossref Full Text | Google Scholar

Meng, J., Yao, L., Sheng, X., Zhang, D., and Zhu, X. (2014). Simultaneously optimizing spatial spectral features based on mutual information for EEG classification. Trans. Biomed. Eng. 62, 227–240. doi: 10.1109/TBME.2014.2345458

PubMed Abstract | Crossref Full Text | Google Scholar

Patro, K. K., Prakash, A. J., Sahoo, J. P., Routray, S., Baihan, A., Samee, N., et al. (2023). SMARTSeiz: deep learning with attention mechanism for accurate seizure recognition in iot healthcare devices. IEEE J Biomed Health Inform. 6, 3336935. doi: 10.1109/JBHI.2023.3336935

PubMed Abstract | Crossref Full Text | Google Scholar

Prakash, V., and Kumar, D. A. (2023). Modified gated recurrent unit approach for epileptic electroencephalography classification. J. Inform. Commun. Technol. 22, 587–617. doi: 10.32890/jict2023.22.4.3

Crossref Full Text | Google Scholar

Qi, F., Li, Y., and Wu, W. R. (2015). A novel algorithm for spatio-temporal filtering and classification of single-trial EEG. IEEE Trans. Neural Networks Learning Syst. 26, 3070–3082. doi: 10.1109/TNNLS.2015.2402694

PubMed Abstract | Crossref Full Text | Google Scholar

Qiu, X., Yan, F., and Liu, H. A. (2023). difference attention ResNet-LSTM network for epileptic seizure detection using EEG signal. Biomed. Signal Proc. Control 83, 104652. doi: 10.1016/j.bspc.2023.104652

Crossref Full Text | Google Scholar

Qiu, Y., Zhou, W., Yu, N., and Du, P. (2018). Denoising sparse autoencoder-based ictal EEG classification. IEEE Trans. Neural Syst. Rehabilit. Eng. 26, 1717–1726. doi: 10.1109/TNSRE.2018.2864306

PubMed Abstract | Crossref Full Text | Google Scholar

Rajinikanth, V., Kadry, S., Taniar, D., Kamaland, K., Elaziz, M. A., and Palani, K. (2022). Detecting epilepsy in EEG signals using synchro-extracting-transform (SET) supported classification technique. J. Ambient Intellig. Human. Comp. 2022, 1–19. doi: 10.1007/s12652-021-03676-x

Crossref Full Text | Google Scholar

Riaz, F., Hassan, A., Rehman, S., Niazi, I. K., and Dremstrup, K. (2015). EMD-based temporal and spectral features for the classification of EEG signals using supervised learning. IEEE Trans. Neural Syst. Rehabilitat. Eng 24, 28–35. doi: 10.1109/TNSRE.2015.2441835

PubMed Abstract | Crossref Full Text | Google Scholar

Rohan, T. I., Yusuf, M. S. U., Islam, M., and Roy, S. (2020). Efficient approach to detect epileptic seizure using machine learning models for modern healthcare system. IEEE. 2020, 1783–1786. doi: 10.1109/TENSYMP50017.2020.9230731

Crossref Full Text | Google Scholar

Saichand, N. V. (2021). Epileptic seizure detection using novel multilayer LSTM discriminant network and dynamic mode Koopman decomposition. Biomed. Signal Proc. Control 68, 102723. doi: 10.1016/j.bspc.2021.102723

Crossref Full Text | Google Scholar

Schoene, A. M., Turner, A., and Dethlefs, N. (2020). “Bidirectional dilated LSTM with attention for fine-grained emotion classification in tweets,” in Proceedings of the AAAI-20 Workshop on Affective Content Analysis (New York, USA: AAAI).

Google Scholar

Shoeibi, A., Ghassemi, N., Khodatars, M., et al. (2022). Detection of epileptic seizures on EEG signals using ANFIS classifier, autoencoders and fuzzy entropies. Biomed. Signal Proc. Control 73, 103417. doi: 10.1016/j.bspc.2021.103417

Crossref Full Text | Google Scholar

Siuly, S., Alcin, O. F., Bajaj, V., et al. (2019). Exploring Hermite transformation in brain signal analysis for the detection of epileptic seizure. IET Sci. Measur. Technol.13, 35–41. doi: 10.1049/iet-smt.2018.5358

Crossref Full Text | Google Scholar

Tsipouras, M. G. (2019). Spectral information of EEG signals with respect to epilepsy classification. EURASIP 2019, 1–17. doi: 10.1186/s13634-019-0606-8

Crossref Full Text | Google Scholar

Tuncer, T., Dogan, S., and Akbal, E. A. (2019). novel local senary pattern based epilepsy diagnosis system using EEG signals. Aust. Phys. Eng. Sci. Med. 42, 939–948. doi: 10.1007/s13246-019-00794-x

PubMed Abstract | Crossref Full Text | Google Scholar

Türk, Ö., and Özerdem, M. S. (2019). Epilepsy detection by using scalogram based convolutional neural network from EEG signals. Brain Sci. 9, 115. doi: 10.3390/brainsci9050115

PubMed Abstract | Crossref Full Text | Google Scholar

Varli, M., and Yilmaz, H. (2023). Multiple classification of EEG signals and epileptic seizure diagnosis with combined deep learning. J. Comp. Sci. 67, 101943. doi: 10.1016/j.jocs.2023.101943

Crossref Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). “Attention is all you need,” in Advances in Neural Information Processing Systems, 5998–6008.

Google Scholar

Wang, J., Gao, R., Zheng, H., Zhu, H., and Shi, C-J. R. (2023). SSGCNet: a sparse spectra graph convolutional network for epileptic EEG signal classification. IEEE Trans. Neural. Netw. Learn. Syst. 16, 3252569. doi: 10.1109/TNNLS.2023.3252569

PubMed Abstract | Crossref Full Text | Google Scholar

Wu, D., Shi, Y., Wang, Z., Yang, J., and Sawan, M. (2023). C 2 SP-Net: joint compression and classification network for epilepsy seizure prediction. IEEE Trans. Neural Syst. Rehabilitat. Eng. 31, 841–850. doi: 10.1109/TNSRE.2023.3235390

PubMed Abstract | Crossref Full Text | Google Scholar

Wu, Q., and Fokoue, E. (2017). “Epileptic seizure recognition,” in UCI Machine Learning Repository.

Google Scholar

Xin, Q., Hu, S., Liu, S., Zhao, L., and Zhang, Y.-D. (2022). An attention-based wavelet convolution neural network for epilepsy EEG classification. IEEE Trans. Neural Syst. Rehabilit. Eng. 30, 957–966. doi: 10.1109/TNSRE.2022.3166181

PubMed Abstract | Crossref Full Text | Google Scholar

Yuan, Y., Xun, G., Jia, K., and Zhang, A. (2018a). A multi-view deep learning framework for EEG seizure detection. IEEE J. Biomed. Health Informat. 23, 83–94. doi: 10.1109/JBHI.2018.2871678

PubMed Abstract | Crossref Full Text | Google Scholar

Yuan, Y., Xun, G., Ma, F., Suo, Q., Xue, H., Jia, K., et al. (2018b). “A novel channel-aware attention framework for multi-channel eeg seizure detection via multi-view deep learning,” in 2018 IEEE EMBS International Conference on Biomedical and Health Informatics (BHI) (Las Vegas, NV: IEEE), 206–209. doi: 10.1109/BHI.2018.8333405

Crossref Full Text | Google Scholar

Zhang, G., Le Yang, B., Li, B., Lu, Y., Liu, Q., Zhao, W., et al. (2020). MNL-network: a multi-scale non-local network for epilepsy detection from EEG signals. Front. Neurosci. 14, 870. doi: 10.3389/fnins.2020.00870

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, T., and Chen, W. L. M. D. (2016). based features for the automatic seizure detection of EEG signals using SVM. IEEE Trans. Neural Syst. Rehabilitat. Eng. 25, 1100–1108. doi: 10.1109/TNSRE.2016.2611601

PubMed Abstract | Crossref Full Text | Google Scholar

Zheng, W. L., and Lu, B. L. (2015). Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Auton. Mental Dev. 7, 162–175. doi: 10.1109/TAMD.2015.2431497

Crossref Full Text | Google Scholar

Keywords: EEG signal, feature encoding, multi-length, feature reinforcement, deep learning

Citation: Zhang G, Zhang A, Liu H, Luo J and Chen J (2024) Positional multi-length and mutual-attention network for epileptic seizure classification. Front. Comput. Neurosci. 18:1358780. doi: 10.3389/fncom.2024.1358780

Received: 20 December 2023; Accepted: 05 January 2024;
Published: 25 January 2024.

Edited by:

Anguo Zhang, University of Macau, China

Reviewed by:

Xin Deng, Chongqing University of Posts and Telecommunications, China
Weibo Liu, Brunel University London, United Kingdom

Copyright © 2024 Zhang, Zhang, Liu, Luo and Chen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jihao Luo, amloYW8ubHVvJiN4MDAwNDA7dS5udXMuZWR1; Jianqing Chen, UmNoZW4uY2hyaXN0b3BoZSYjeDAwMDQwO3lhaG9vLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.