Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 10 May 2022
Sec. Emotion Science

Transfer Discriminative Dictionary Pair Learning Approach for Across-Subject EEG Emotion Classification

  • 1HUA LOOKENG Honors College, Changzhou University, Changzhou, China
  • 2School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou, China

Electroencephalogram (EEG) signals are not easily camouflaged, portable, and noninvasive. It is widely used in emotion recognition. However, due to the existence of individual differences, there will be certain differences in the data distribution of EEG signals in the same emotional state of different subjects. To obtain a model that performs well in classifying new subjects, traditional emotion recognition approaches need to collect a large number of labeled data of new subjects, which is often unrealistic. In this study, a transfer discriminative dictionary pair learning (TDDPL) approach is proposed for across-subject EEG emotion classification. The TDDPL approach projects data from different subjects into the domain-invariant subspace, and builds a transfer dictionary pair learning based on the maximum mean discrepancy (MMD) strategy. In the subspace, TDDPL learns shared synthesis and analysis dictionaries to build a bridge of discriminative knowledge from source domain (SD) to target domain (TD). By minimizing the reconstruction error and the inter-class separation term for each sub-dictionary, the learned synthesis dictionary is discriminative and the learned low-rank coding is sparse. Finally, a discriminative classifier in the TD is constructed on the classifier parameter, analysis dictionary and projection matrix, without the calculation of coding coefficients. The effectiveness of the TDDPL approach is verified on SEED and SEED IV datasets.

Introduction

Emotion is an advanced cognitive activity of human beings and plays an increasingly important role in human rational thinking, decision-making, perception and learning. Since human emotional state is a complex psychological and physiological process, the study of human emotion is a complex cognitive process. At present, emotion classification is an important research direction in emotional computing and artificial intelligence. Emotion classification has been widely used in intelligent transportation, distance education, and human-computer interaction. However, as an interdisciplinary research field, there are still many challenges in emotion classification based on artificial intelligence technology. Early researchers usually used facial expressions, audio and peripheral physiological signals (body temperature, blood pressure, pulse, respiration, etc.) to indirectly study emotions (Shu et al., 2018). However, such information is easy to disguise and not objective, it is easy for users to deceive the detection device by subjectively controlling the external expression of emotions, making it difficult to objectively and accurately describe the user's emotional state. With the development of cognitive science, the relationship between the specific location of the cerebral cortex and emotion is gradually recognized, so it is a good choice to directly study the activity of the cerebral cortex to study the emotional state. Researchers can use the existing knowledge of signal processing to directly process physiological signal data, which also greatly promotes the development of emotion recognition. Studies have shown that EEG signals can be used as an informative feature of emotional state. EEG signals have the advantages of being difficult to camouflage, portable and non-invasive acquisition, high temporal resolution, and can directly reflect brain activity (Ramakrishnan and Panachakel, 2021).

Sorkhabi (2014) used continuous wavelet transform (CWT) algorithm to divide EEG signals into 5 frequency bands. The CWT algorithm extracted variable-sized window features and obtained high-frequency information in shorter regions. Based on this, the researchers detected the activity of EEG signals within a 1-second time window. The subjects' self-rating values in terms of valence dimension and arousal degree were found to have a stable correlation with their EEG power and power spectrum entropy, and the high frequency band features had a more accurate classification effect than the low frequency band features. Atkinson and Campos (2016) proposed an emotion classification approach based on the combination of minimum redundancy-maximum correlation feature selection and kernel classifier. The significant benefit of this approach is that it incorporated the feature selection task of EEG signals into the classification task, identifying a wider range of emotion recognition approaches while using a multi-label classifier. Zhang and Lee, 2009 proposed an emotion understanding system. The system used the asymmetry of the prefrontal lobe of the brain as a feature and uses support vector machine (SVM) as a classifier, which can distinguish two emotional states with a recognition rate of about 73%. Zhuang et al. (2017) used the empirical mode decomposition (EMD) approach to automatically decompose the EEG signal into multiple intrinsic mode functions (IMFs), and achieved a recognition result of 70.41%. Mert and Akan (2018) used the multivariate extended version of the EMD approach to extract the multi-channel IMFs of EEG signal.

Neurobiological studies have shown that the human nervous system adopts a sparse representation strategy to receive and process external stimuli, and only needs to activate a small number of neurons in the cerebral cortex to complete the reception of information (Beyeler et al., 2019). Dictionary learning is a machine learning method based on this idea. In dictionary learning, data samples can be sparsely approximated by linear combinations of basis signals in the dictionary. The core idea is to learn an optimal dictionary under the certain constraint, and obtain the optimal sparse representation of the sample on the learned dictionary (Zhang et al., 2018). For EEG signals, we can also use sparse representation to avoid a lot of redundant information. Sheykhivand et al. (2020) took raw EEG signals directly as the model input and used a dictionary-learning-based sparse representation classifier. The classifier showed good performance without involving feature extraction and selection. Gu et al. (2021) mapped EEG signals in different frequency bands to subspaces, and learned a shared dictionary of multiple frequency bands, which can learn more discriminative knowledge hidden frequency bands. On the basis of this research, Zhu et al. (2022) proposed a new multi-band dictionary learning-based EEG emotion analysis model. They divided the projection matrix of each frequency band into two parts: common component and private component. The common component was used to mine the shared knowledge of different frequency bands, and the private component was used to mine the unique knowledge of each frequency band. Simultaneously, the shared dictionary on the multi-band signal in the subspace established the connection between multi-bands. Barthélemy et al. (2013) proposed a dictionary learning that considers both inter-channels links and shift-invariance, which improved the representation ability and flexibility of the dictionary. Although traditional EEG-based emotion classification approaches have been effectively verified and widely used, they are all based on subject-independent general models and do not consider individual differences. Studies have shown that subjects of different genders, ages, races, and health conditions have great differences in emotional expression, and sometimes this difference is even greater than the change in emotion (Chaplin, 2015). Therefore, the generalization performance of traditional emotion classification model is poor.

Traditional emotion classification model assumes that EEG signals from different subjects (training and test data) have the same feature space distribution. Transfer learning can relax this restriction. In target task modeling, transfer learning can use knowledge from other domain (source domain, SD) to help target domain (TD) training and modeling (Lin and Tzyy-Ping, 2017). For example, knowledge or patterns learned from other subjects can be applied to a new subject through transfer learning. By mining the information shared between different subjects, a model that adapts to the target subject's data distribution is finally constructed. In this study, a transfer discriminative dictionary pair learning approach (TDDPL) is developed for across-subject EEG emotion classification. The core idea of TDDPL is to find the domain-invariant subspace through the projection matrix, and learn the shared dictionary pair based on the paired dictionary learning framework in the subspace. By leveraging SD discriminative information, a shared synthesis dictionary and analysis dictionary are used to learn more discriminative domain-invariant low-rank coding to improve the performance of TD model. Our experiments in SEED and SEED IV datasets show that TDDPL achieves very competitive accuracy with state-of-the-art transfer learning approaches.

Specifically, the contribution of this work is fourfold. (1) By adopting the maximum mean discrepancy (MMD) of low-rank encoding to mitigate the distribution difference, EEG data from different subjects are projected into the domain-invariant subspace, and a bridge is built between SD and TD through transfer dictionary pair learning approach. (2) By minimizing the reconstruction error and inter-class separation of each synthesis sub-dictionary, the reconstruction between each sub-dictionary and heterogeneous low-rank coding is approximately an empty set, which can improve the discriminative ability of synthesis dictionary. (3) By minimizing the reconstruction error and the inter-class separation of each analysis sub-dictionary, the learned low-rank encoding has good sparsity. (4) The discriminative classifier in the TD is trained on the classifier parameter, analysis dictionary and projection matrix, the classifier can be directly used for the classification of test data, avoiding additional time-consuming coding reconstruction.

The remainder of this paper is organized as follows. Section Background introduces the study background. Section Domain Adaptation Sparse Representation Classifier the proposed approach. Section Experiment reports the experimental results. Section Conclusion draws conclusions.

Background

Transfer Leaning in EEG Emotion Classification

There are two general classification models for EEG emotion classification, one is to build an emotion classification model that can be used for a long time, and the other is to build an emotion classification model that can be used across subjects. For the same subject, the EEG signal will change over time. Compared with the differences in EEG signals between the same subjects, the differences in EEG signals between different subjects are greater, resulting in poor model generalization performance across subjects (Wan et al., 2021). In addition, in the research of emotion classification, building a general model often needs to acquire a large amount of subject data, which is often difficult to achieve. Therefore, the above two problems are how to mine the shared useful information in the EEG data with large differences, in which how to construct a general model across subject is more challenging.

Transfer learning is a machine learning strategy that uses existing knowledge to solve problems in different but related domains. The essence of transfer learning is the transfer and reuse of knowledge, which is, extracting useful knowledge from one or more SDs to assist in tasks in TD (Li et al., 2020). For example, Zanini et al. (2018) proposed a transfer learning model using Riemannian geometry to handle across-session and across-subject classification tasks in brain-computer interface (BCI). Using affine transformations of the spatial covariance matrices of data from each session or subject, the authors tackled EEG-based BCI across-subject classification problems, and then calibrated the classifier using data from previous sessions or subject data. Lan et al. (2018) combined domain adaptation approaches such as maximum independent domain adaptation, transfer component analysis, subspace alignment, and information theoretical learning with non-transfer learning approaches. The experimental results showed that the models using domain adaptation techniques outperform the non-transfer learning approaches. Zhang and Wu (2020) proposed a manifold embedded knowledge transfer learning model. This model aligned the covariance matrices of EEG data on a Riemannian manifold, and then performed domain adaptation by minimizing the difference in the joint probability distribution, while preserving the geometry of the original EEG data. Morioka et al. (2015) proposed a dictionary learning approach with strong generalization ability. To be suitable for emotion analysis of multiple individuals, the approach took the EEG signals of the target individual as the calibration data. Because EEG signals are weak and easily affected by noise, transfer learning for EEG emotion classification is still a very challenging research.

EEG Emotion Datasets Used in This Study

The datasets used in this paper are experimented on two public emotion datasets SEED (Zheng and Lu, 2015) and SEED-IV (Zheng et al., 2019). The SEED dataset employed six emotion-labeled movie clips to elicit three emotions in subjects: positive, neutral, and negative. The subject's EEG data was recorded while inducing the subject's emotion, and the emotional label represented by the movie was the subject's EEG label. The mean age of the 15 Chinese subjects participating in the experiment was 23.27. Each subject performed a total of 3 trials, and each trial consisted of 15 trials. Each trial consisted of 5 s prompts, about 4 min of audio stimulation, 45s of self-assessment and 15 s of rest. Similarly, the SEED-IV dataset employed emotion-labeled movie clips to elicit four emotions in subjects: happy, sad, fearful, and neutral. The experiment recorded the subjects' EEG data at the same time, and the emotional label represented by the movie is the EEG label. The 15 healthy subjects who participated in the experiment were between the ages of 20 and 24. The experiment designed three different experiments for each subject, each experiment contained 24 trials (six trials for each emotion), and each experiment used a completely different movie clip. The specification comparison between the SEED and SEED-IV datasets is shown in Table 1.

TABLE 1
www.frontiersin.org

Table 1. Specification comparison between the SEED and SEED-IV datasets.

Synthesis Dictionary and Analysis Dictionary Learning

Let Y=[y1,...,yN]Rd×N be the data matrix. The core idea of synthesis dictionary learning looks for a dictionary DRd×r that can express each sample, where r is number of dictionary atoms. Let ARr×N be the coding coefficient matrix obtained by dictionary D. Synthesis dictionary learning model (Jiang et al., 2013) can be formulated by,

minD,A Y-DAF2+Ap+f1(Y,D,A),    (1)

where ‖·‖p is usually 0 or 1 norm. Y-DAF2 represents the reconstruction error. f1(·) can be some constraint terms on dictionary atoms, training samples and coding coefficients, such as low-rank constraint, label consistency constraint, locality constraint, structured sparsity constraint, and Fisher discriminative constraint (Wang et al., 2017). Synthesis dictionary learns a comprehensive dictionary by solving a reconstruction error minimization problem.

Analysis dictionary learning provides an intuitive explanation (Du et al., 2021). It directly acts dictionary on feature samples into sparse coding space, which is similar to the feature transformation. Analysis dictionary learning model can be formulated by

minP,A PY-AF2+Ap+f2(Y,P,A),    (2)

where PRd×r is the learned synthesis dictionary. f2(·) can be some constraint terms to obtain the stable solution of P and A.

Domain Adaptation Sparse Representation Classifier

The Objective Function of TDDPL

Suppose the training set Y consists of n SD sets Ysi(1 ≤ in) and a TD set Yt,Y=[Ys1,Ys2,...,Ysn,Yt]Rd×N, where YsiRd×Nsi and YtRd×Nt composing of k classes of training samples, N=i=1nNsi+Nt.

Synthesis Sub-dictionary Discriminative Term

The synthetic dictionary learning model learns a synthetic dictionary D to sparsely represent samples with a linear combination of a small number of dictionary atoms. We use the projection matrix to project the data of each domain into the subspace, and use the projected low-dimensional samples to learn a shared synthetic dictionary. Suppose the projection matrices corresponding to the kth class sample on the Ysi and Yt are Ωsi,kRd×m and Ωt,kRd×m, respectively. m is the dimension of the projected subspace. Ωsi,kTYsi,k and Ωt,kTYt represent the kth class projecting samples belonging to SD and TD, respectively. The synthesis sub-dictionary discriminative term minimizes the reconstruction error term of each synthesis sub-dictionary, while minimizing the inter-class separation term. The designed synthesis sub-dictionary discriminative term is written as,

minΩ,D,Ak=1K(i=1n(Ωsi,kTYsi,k-DkAsi,kF2+DkA^si,kF2)       +Ωt,kTYt,k-DkAt,kF2+DkA^t,kF2)    (3)

where Asi,k(At,k) is the coding coefficients of Ysi,k(Yt,k), and the A^si,k(A^t,k) is the complement set of Asi,k(At,k), A^si,kAsi,k=A,A^si,kAsi,k= (A^t,kAt,k=A,A^t,kAt,k=). DkRm×rk is the kth class sub-dictionary, where rk is the number of sub-dictionary atoms.

The first item of Eq. (3) is to ensure the discriminative ability of the synthesis sub-dictionary, and each synthesis sub-dictionary Dk can represent the data well. The reconstruction structure of second item of Eq. (3) is approximately zero to achieve the separation between sub-dictionary classes.

Let ΩkT=[Ωs1,kT,...,Ωsn,kT,Ωt,kT], [[Mathtype-mtef1-eqn-42.mtf]], Ak=[As1,k,...,Asn,k,At,k], Eq. (3) can be combined into the following form,

minΩ,D,Ak=1K(ΩkTYk-DkAkF2+DkA^kF2)    (4)

Analysis Sub-dictionary Discriminative Term

For the Ysi,k,PkΩsi,kTYsi,k represents the coding coefficients of Ωsi,kTYsi,k on the analysis sub-dictionary PkRm×rk. To achieve the discriminative ability of each Pk, the k-th low-dimensional projection data is projected into a non-zero encoding space, i.e., PkΩsi,kTYsi,kAsi,k. Simultaneously, the j-th low-dimensional projection data is projected into approximate zero encoding space, i.e., PkΩsj,kTYsi,j0,kj.

Implement this idea on all SDs and TD, the designed analysis sub-dictionary discriminative term is written as,

minΩ,D,Ak=1K(i=1n(PkΩsi,kTYsi,k-Asi,kF2+PkΩsi,kTYsi,kA^si,kF2)     +PkΩt,kTYt,k-At,kF2+PkΩsi,kTYsi,kA^t,kF2)    (5)

Let Pk=[Ps1,k,...,Psn,k,Pt,k], Eq. (5) can be combined into the following form,

minΩ,D,Ak=1K(PkΩkTYk-AkF2+PkΩkTYkA^kF2)    (6)

Domain Adaptation Term

In the projected subspace, the difference of data distribution between SD and TD is measured on low-rank coding using MMD strategy. The designed domain adaptation term is written as,

minA k=1K1Nksi=1Nksais-1Nkti=1Nktajt22    (7)

Let

Qkij={1NksNks,if yi,yj is from SD1NktNkt,if yi,yj is from  TD-1NksNkt,if yi(yj) is from SD, and yj(yi) is from  TD0, otherwise

Eq. (7) can be written in the matrix form,

minA k=1Ki=1Nksj=1NktaiTajQijk=minA k=1Ktr(AkQkAkT)    (8)

where low-rank matrix A=[As1,...,Asm,At].

Explicit Rank Minimization Term

Imposing low-rank constraints on the coding coefficients can significantly reduce the adverse effects of noise in the samples. A low-rank constraint is imposed on each analysis in this study. To improve computing efficiency of low-rank computation, following (Ding and Fu, 2019), the low-rank coding matrix is expressed as the product of two matrices, AΛΘ, where ΛRm×K and ΘRK×N. The designed explicit rank minimization term is written as,

rank(A)=k=1KAk-ΛkΘkF2    (9)

Discriminative Classifier Term

A linear classifier based on low rank coding is embedded in the TDDPL approach. Let Wk and Hk be the classifier parameter and class label of k-th class data, respectively. For Wk and Hk with the same class, the ideal classification result is WkPkXkHk. According to this idea, the designed discriminative classifier term is written as,

minP,W,Ωk=1KHk-WkPkΩkTYkF2    (10)

Considering Eqs. (4), (6), (8)-(10), the objective function of TDDPL can be written as,

minΩ,D,P,A,Λ,Θ,Wk=1K(ΩkTYk-DkAkF2+DkA^kF2+λ1(PkΩkTYk-AkF2+PkΩkTYkA^kF2)+λ2tr(AkQkAkT)+λ3Ak-ΛkΘkF2+λ4Hk-WkPkΩkTYkF2)+γWF2.    (11)

Obviously, we can obtain the projection matrix Ω, synthesis dictionary D, analysis dictionary P, low-rank coding A (matrices Λ and Θ) by solving the optimization Eq. (11). The projection matrix Ω projects the SD and TD samples into a low-dimensional subspace. According to the MMD strategy, the data distribution differences between different domains are as small as possible. The synthesis dictionary D can better reconstruct the projection sample ΩTY. The multiplication of the analysis dictionary P and the projection sample ΩTY, i.e., PΩTY can obtain an approximate block diagonal coding coefficient matrix with strong discriminative ability. The shared dictionary pair P and D become the bridges between SD and TD. The discriminative knowledge of SD is transferred to TD space to build a discriminative classifier.

Optimization

Eq. (11) is a non-convex optimization problem. However, it is a convex optimization problem when only one variable is optimized while fixing the other variables. Therefore, the optimization problem Eq. (11) is split into several sub-optimization problems here.

Let ΩkTYk=Bk,PkΩkTYk=Ck, we have PkBk = Ck. Eq. (11) can be re-written as,

minΩ,D,P,A,Λ,Θ,B,C,Wk=1K(Bk-DkAkF2+DkA^kF2+λ1(Ck-AkF2     +CkA^kF2)+λ2tr(AkQkAkT)+λ3Ak-ΛkΘkF2     +λ4Hk-WkCkF2+λ5(Bk-ΩkTYkF2     +Ck-PkBkF2))+γWF2,s.t. di2=1,i    (12)

1) Update A, while fixing Ω, D, P, Λ, Θ, B, C, W.

To remove terms which are irrelevant to A, we have,

[ A]=k=1K(Bk-DkAkF2+λ1Ck-AkF2+λ2tr(AkQkAkT)       +λ3Ak-ΛkΘkF2),    (13)

By setting the derivative of Eq. (13) to zero, we obtain the solution of Ak,

Ak=(DkTDk+(λ1+λ3)I+λ2Qk)-1(DkTBk+λ1Ck+λ3ΛkΘk)    (14)

2) Update P, while fixing Ω, D, A, Λ, Θ, B, C, W.

To remove terms which are irrelevant to P, we have,

[P]=k=1K(Ck-PkBkF2),    (15)

By setting the derivative of Eq. (15) to zero, we obtain the solution of P,

Pk=CkBkT(BBkT+θI)-1    (16)

where θI is to prevent singular solution in matrix inversion.

3) Update W, while fixing Ω, D, P, A, Λ, Θ, B, C.

To remove terms which are irrelevant to W, we have,

 [ W ]=k=1K(λ4Hk-WkCkF2)+γWF2    (17)

By setting the derivative of Eq. (17) to zero, we obtain the solution of Wk,

Wk=(λ4HkCkT)(λ4CkCkT+γI)-1    (18)

4) Update D, while fixing Ω, P, A, Λ, Θ, B, C, W.

To remove terms which are irrelevant to D, we have,

 [ D ]=k=1K(Bk-DkAkF2+DkA^kF2),s.t. di221, i    (19)

Eq. (19) can be solved via the dual from, then we obtain,

 [ Dk ]=Bk-DkAkF2+DkA^kF2+i=1nσk,i(di2-1)    (20)

where σk,i is the Lagrange multiplier of the i th equality constraint.

Let ΓkRm×m[[Mathtype-mtef1-eqn-105.mtf]], where (Γk)i,iRm×m,(Γk)i, i = σi, we obtain,

L(Dk,di)=Bk-DkAkF2+DkA^kF2+tr(DkTDkΓk)-tr(Γk)    (21)

By setting the derivative of Eq. (21) to zero, we obtain the solution of Dk,

Dk=BkAkT(AkAkT+A^k(A^k)T+Γk)-1    (22)

5) Update Λand Θ, while fixing Ω, P, A, D, B, C, W.

To remove terms which are irrelevant to Λ and Θ, we have,

[Λk,Θk ]=k=1KAk-ΛkΘkF2    (23)

By setting the derivative of Eq. (23) to zero, we obtain the solution of Λk and Θk,

Λk=AkΘkT(ΘkΘkT)    (24)
Θk=(ΛkΛkT)ΛkTΛk    (25)

where (▪)is the Moore–Penrose pseudoinverse.

6) Update Ω, while fixing P, A, Λ, Θ, D, B, C, W.

To remove terms which are irrelevant to Ω, we have,

[Ω]=k=1KBk-ΩkTYkF2    (26)

To set the derivative of Eq. (26) with respect to Ω to zero, we have,

Ωk=BkYkT(YkYkT+θI)-1    (27)

where θI is to prevent singular solution in matrix inversion.

7) Update and C, while fixing.

To remove terms which are irrelevant to Ck, we have,

[Ck]=k=1K(λ1(CkAkF2+CkA^kF2)+λ1Hk                  WkCkF2+λ5CkPkBkF2),    (28)

We use sub-gradient descent method, and compute the gradient of Eq.(28) with respect to Ck,

[Ck]Ck=λ1(Ck-Ak+CkA^kA^kT)+λ4(WkTWkCk-WkTHk)                  +λ5(Ck-PkBk)    (29)

Ck can be updated by descent method with learning rate α,

Ck=Ck-α[Ck]Ck    (30)

To remove terms which are irrelevant to Bk, we have,

[Bk]=k=1K(BkDkAkF2+λ5(BkΩkTF2                  +CkPkBkF2)),    (31)

By setting the derivative of Eq. (31) to zero, we obtain the solution of Bk,

Bk=((1+λ5)I+λ5PkTPk)-1(DkAk+λ5ΩkTYk+λ5PkTCk)    (32)

Algorithm 1 describes the proposed TDDPL approach.

ALGORITHM 1
www.frontiersin.org

Algorithm 1. Transfer discriminative dictionary pair learning approach.

Testing

After obtaining the classifier parameter W, projection matrix Ω, and analysis sub-dictionary P, we classify the test samples based on W, Ω, and P. For test sample xnew, we use the following formulation to predict its class label,

l(xnew)=argmaxic(WPΩTxnew)i    (33)

where (WPΩTxnew)i is the i-th element of WPΩTxnew.

Experiment

Experimental Setting

For the SEED dataset, EEG features were extracted in non-overlapping 1s time windows for each segment of preprocessed EEG data. For the SEED-IV dataset, EEG features were extracted in non-overlapping 4 s time windows for each segment of preprocessed EEG data. The feature extraction approach used in this paper is the differential entropy (DE) feature (Li et al., 2019). In the experiment, DE features are calculated in 5 frequency bands for each channel. So the total dimension of the extracted features is 62 × 5=310 dimensions. We evaluate the TDDPL approach on two transfer learning strategies: one SD to one TD (o → o), and multiple SDs to one TD (m → o). For the o → o scenario, we choose one subject used as the TD and another subject used as the SD. If z is the number of subjects in the SEED and SEED IV datasets, there are z(z-1) different o → o tasks in total. In the SEED dataset, for the data in the TD, the 30 s data of each trial is randomly selected for training, and the rest of the data is used for testing. In the SEED-IV dataset, for the data in the TD, the 28 s data of each trial is randomly selected for training, and the rest of the data is used for testing. For the m → o scenario, one subject is selected as the TD and all the remaining subjects are used as the SD, so there are z different tasks in the m → o scenario. Due to the large number of training samples in the SD, for the SEED dataset and SEED-IV datasets, we select 1/10 of the SD data for training, and repeat this process 10 times, so that the randomly sampled training data can cover the entire training dataset. The selection of training set and test set of the TD is consistent with the o → o scenario.

We compare the TDDPL approach with the following approaches: 1) Traditional machine learning approaches: sparse representation-based classification (SRC) (Wright et al., 2009) and DPL approach (Ameri et al., 2016). 2) Transfer learning approaches: manifold embedded knowledge transfer (MEKT) (Zhang and Wu, 2020), transfer component analysis (TCA) (Pan et al., 2011), maximum independence domain adaptation (MIDA) (Yan et al., 2018), domain adaptation sparse representation classifier (DASRC) (Ni et al., 2021). For TCA and MEKT, the Gaussian kernel parameter and regularization parameters are set in the grid {2−6,..., 25}. For SRC, DPL, DASRC, and TDDPL, the dimension of subspace and the number of atoms in each class are set in the grid {20, 30,..., 100} and {10, 15,..., 30}, respectively. All the algorithms are implemented in MATLAB.

Experiments on the SEED Dataset

In this subsection, we compare the approaches on DE feature of the SEED dataset. Tables 2, 3 show the average accuracy of each approach for the o → o and m → o scenarios on the session 1, respectively. It can be seen that compared with other approaches, (1) the classification performances of the approaches using the transfer learning strategy outperform those of the approaches without the transfer learning strategy. It is demonstrated that transfer learning strategy can mitigate the impact of individual differences to a certain extent. Since the SRC and DPL classifiers simply combine SD and TD to build a model which may not be adapted to the test set data distribution, it is difficult to obtain ideal experimental results. It also indicates that transfer learning improves the approach's classification performance in across-subject EEG emotion analysis. (2) Regardless of the o → o or m → o scenarios, the TDDPL approach achieves the best or second best performance in Tables 2, 3. As seen the results in Table 2, the TDDPL approach improves by 1.70% compared with the best comparison approach. The TDDPL approach projects the EEG data of different subjects into the subspace to reduce the distribution difference between subjects. The shared analysis dictionary learned by the TDDPL produces discriminative low-rank coding, and the learned shared synthesis dictionary has good coding reconstruction ability. Therefore, the TDDPL approach performs well and is more stable in across-subject EEG emotion classification.

TABLE 2
www.frontiersin.org

Table 2. Average classification accuracies on session 1 of SEED dataset in the o → o scenario.

TABLE 3
www.frontiersin.org

Table 3. Average classification accuracies on session 1 of SEED dataset in the m → o scenario.

To verify the performance of our approach on other commonly used EEG features, we compared the average accuracy of the TDDPL approach with other approaches on session 2 and session 3 of the SEED dataset. The average classification accuracy is shown in Figures 1, 2. As shown in Figures 1, 2, the TDDPL approach has the best average classification accuracy on session 2 and session 3, and the experimental results are significantly better than other comparison approaches. The TDDPL approach improves by 10.36% compared with the traditional DPL classification approach, and also improves by 1.70% compared with the best comparison approach in the o → o scenario. The results reveal that the TDDPL approach has excellent across-subject adaptability, and it can accurately and effectively realize EEG emotion classification.

FIGURE 1
www.frontiersin.org

Figure 1. Average classification accuracies on session 2 of the SEED dataset.

FIGURE 2
www.frontiersin.org

Figure 2. Average classification accuracies on session 3 of the SEED dataset.

Experiments on the SEED IV Dataset

Next, we compare the approaches on the DE feature of the SEED IV dataset. Tables 4, 5 show the average classification accuracy of each approach for the o → o scenario and the m → o scenario on the session 1, respectively. As can be seen in Tables 4, 5, (1) Because the emotion classes of SEED IV dataset are more complex than the SEED dataset, the average accuracy of each approach is slightly lower than that of Tables 2, 3. (2) Similar to the results in Tables 2, 3, the transfer learning approaches (MIDA, MEKT, and DASRC) outperforms the non-transfer learning approaches (SRC and DPL), which indicates that simply mixing SD and TD into a training set is not suitable for EEG emotion recognition. (3) TDDPL achieved the best results on the vast majority of individuals. The label information of the SD data is transferred to TD through the shared dictionary pair D and P, thereby improving the performance of the task in TD. TDDPL also combines the advantages of a comprehensive dictionary approach and an analytical dictionary approach. Based on the learned analysis dictionary, projection matrix, and classifier parameters, the classifier learned in TD has good discriminative ability.

TABLE 4
www.frontiersin.org

Table 4. Average classification accuracies on session 1 of SEED IV dataset in the o → o scenario.

TABLE 5
www.frontiersin.org

Table 5. Average classification accuracies on session 1 of SEED IV dataset in the m → o scenario.

We compare the average accuracy of the TDDPL approach with other approaches in session 2 and session 3 of the SEED IV dataset, and the average classification accuracies are shown in Figures 3, 4. Similar to the results in Figures 1, 2, the classification performance of the TDDPL approach is the best, indicating that our proposed TDDPL approach can improve the performance of the learning task in the TD by using auxiliary data through dictionary pair learning. We also notice that the performance of all comparison approaches in the o → o scenario is slightly worse than that of m → o scenario of the SEED and SEED IV datasets. This is because in the o → o scenario, when single-subject data is used as the SD, if the correlation between the SD and the TD is low, the effect of transfer learning approach may not be good. In the m → o scenario, multiple SDs expand the capacity of the training set, which can also make the performance of transfer learning approach more stable.

FIGURE 3
www.frontiersin.org

Figure 3. Average classification accuracies on session 2 of the SEED IV dataset.

FIGURE 4
www.frontiersin.org

Figure 4. Average classification accuracies on session 3 of the SEED IV dataset.

Conclusion

Aiming at the problem of subject differences in the data distribution of EEG signals; simultaneously, due to insufficient information in the new domain in EEG emotion classification, this study studies how to mine knowledge from other related domains and transfer it to improve the performance of the task in TD. To this end, we propose a transfer discriminative dictionary pair learning approach based on subspace transfer learning. In the subspace, the distribution of TD is similar to that of the SD, and we learn a shared dictionary pair with discriminative ability. The label information of the SD is used to construct a classifier in the TD. However, EEG emotion classification based on transfer learning still faces many challenges. In the following research work, we will focus on the following aspects: (1) The TDDPL approach proposed in this paper is a batch learning mode with high time complexity and is not adopted used in online learning situations. How to improve the training efficiency of TDDPL and propose an online learning approach is an important topic to be studied urgently. (2) The TDDPL approach mainly studies the problem that the SD and TD data and label space are consistent, and it does not consider the situation that the TD label space is unavailable. In this case, how to construct an effective subspace, mine the similarities and differences between domains, and conduct effective knowledge transfer is a direction for future research. (3) This paper studies the homogeneous transfer learning problem, that is, the EEG feature space of the SD is the same as the TD. The problem becomes complicated when the feature spaces differ between domains. So far, the research based on heterogeneous transfer learning is not sufficient. How to use heterogeneous domain knowledge to improve the learning ability of the TD is an important content that needs to be studied in the future.

Model Parameter Analysis

The dimension of the subspace m and the number of sub-dictionary atoms rk per class are key parameters in the TDDPL approach. Figures 5, 6 show the average classification accuracy of the TDDPL approach with varying m and. Figure 5 shows the average classification accuracy of o → o and m → o scenarios on session 1 of the SEED dataset. Figure 6 shows the average classification accuracy of the two transfer learning scenarios on session 1 of the SEED IV dataset. It can be seen that the TDDPL approach is optimal when the subspace dimension is about 50 and the number of sub-dictionary atoms per class is about 20 in the SEED dataset. The TDDPL approach is optimal when the subspace dimension is 60 and the number of sub-dictionary atoms per class is about 25 in the SEED IV dataset. In addition, the dimension of the subspace is much smaller than that of the original space, indicating that domain-invariant knowledge can exist in low-dimensional subspaces. The low-dimensional subspace also reduces the training burden of the TDDPL approach.

FIGURE 5
www.frontiersin.org

Figure 5. The average classification accuracy of the TDDPL approach under varying m and rk on the SEED dataset. (A) o → o scenario, (B) m → o scenario.

FIGURE 6
www.frontiersin.org

Figure 6. The average classification accuracy of the TDDPL approach under varying m and rk on the SEED IV dataset. (A) o → o scenario, (B) m → o scenario.

Figure 7 shows the average classification accuracy of TDDPL under varying training samples in the TD in the o → o and m → o scenarios of the SEED and SEED IV datasets. As can be seen from Figure 7A, the classification performance of our approach is slightly improved with the increase of training samples in the TD, but its improvement speed is very slow. After reaching the highest accuracy, it is not very sensitive to the increase of training samples in TD. We can draw similar observations from Figure 7B. It shows that the TDDPL approach only needs a small number of training samples in the TD to establish the transfer learning model, and obtains the discriminative knowledge from SD.

FIGURE 7
www.frontiersin.org

Figure 7. The average classification accuracy of the TDDPL approach under varying Nt in the TD on (A) SEED dataset, (B) SEED IV dataset.

Data Availability Statement

Publicly available datasets were analyzed in this study. Public available SEED and SEED IV datasets used in this paper can be found in the following links: https://paperswithcode.com/dataset/seed-1, https://bcmi.sjtu.edu.cn/~seed/downloads.html.

Author Contributions

TN developed the theoretical framework and model in this study. YR and MD implemented the experiment. TN and YR wrote the manuscript. All authors read and approved the manuscript.

Funding

This work was supported in part by the Natural Science Foundation of Jiangsu Province under Grant Bk202101333.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Ameri, R., Pouyan, A., and Abolghasemi, V. (2016). Projective dictionary pair learning for EEG signal classification in brain computer interface applications. Neurocomputing. 218, 382–389, doi: 10.1016/j.neucom.2016.08.082

CrossRef Full Text | Google Scholar

Atkinson, J., and Campos, D. (2016). Improving BCI-based emotion recognition by combining EEG feature selection and kernel classifiers. Expert Syst. Appl. 47, 35–41, doi: 10.1016/j.eswa.2015.10.049

CrossRef Full Text | Google Scholar

Barthélemy, Q., Gouy-Pailler, C., Isaac, Y., Souloumiac, A., Larue, A., and Mars, J. I. (2013). Multivariate temporal dictionary learning for EEG. J. Neurosci. Methods. 215, 19–28, doi: 10.1016/j.jneumeth.2013.02.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Beyeler, M., Rounds, E. L., Carlson, K. D., Dutt, N., and Krichmar, J. L. (2019). Neural correlates of sparse coding and dimensionality reduction. PLoS Comput. Biol. 16, e10069008, doi: 10.1371/journal.pcbi.1006908

PubMed Abstract | CrossRef Full Text | Google Scholar

Chaplin, T. M. (2015). Gender and emotion expression: a developmental contextual perspective. Emotion Rev. 7, 14–21. doi: 10.1177/1754073914544408

PubMed Abstract | CrossRef Full Text | Google Scholar

Ding, Z., and Fu, Y. (2019). Deep transfer low-rank coding for cross-domain learning. IEEE Trans. Neural Netw. Learn. Syst. 30, 1768–1779. doi: 10.1109/TNNLS.2018.2874567

PubMed Abstract | CrossRef Full Text | Google Scholar

Du, H., Zhang, Y., Ma, L., and Zhang, F. (2021). Structured discriminant analysis dictionary learning for pattern classification. Knowl. Based Syst. 216, 106794. doi: 10.1016/j.knosys.2021.106794

CrossRef Full Text | Google Scholar

Gu, X., Fan, Y., and Zhou, J. (2021). Optimized projection and fisher discriminative dictionary learning for EEG emotion recognition. Front. Psychol. 12, 705528, doi: 10.3389/fpsyg.2021.705528

PubMed Abstract | CrossRef Full Text | Google Scholar

Jiang, Z., Lin, Z., and Davis, L. (2013). Label consistent K-SVD: learning a discriminative dictionary for recognition. IEEE PAMI. 35, 2651–2664, doi: 10.1109/TPAMI.2013.88

PubMed Abstract | CrossRef Full Text | Google Scholar

Lan, Z., Olga, S., Wang, L., Reinhold, S., and Muller-Putz, G. R. (2018). Domain adaptation techniques for EEG-based emotion recognition: a comparative study on two public datasets. IEEE Trans. Cogn. Develop. Syst. 11, 85–94, doi: 10.1109/TCDS.2018.2826840

CrossRef Full Text | Google Scholar

Li, J., Qiu, S., Shen, Y. Y., Liu, C. L., and He, H. (2020). Multisource transfer learning for cross-subject EEG emotion recognition. IEEE Trans Cybern. 50, 3281–3293, doi: 10.1109/TCYB.2019.2904052

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, P., Liu, H., Si, Y., Li, C., Li, F., Zhu, X., et al. (2019). EEG based emotion recognition by combining functional connectivity network and local activations. IEEE Trans Biomed Eng. 66, 2869–2881. doi: 10.1109/tbme.2019.2897651

PubMed Abstract | CrossRef Full Text | Google Scholar

Lin, Y. P., and Tzyy-Ping, J. (2017). Improving EEG-based emotion classification using conditional transfer learning. Front. Human Neurosci. 6, 00334, doi: 10.3389/fnhum.2017.00334

PubMed Abstract | CrossRef Full Text | Google Scholar

Mert, A., and Akan, A. (2018). Emotion recognition from EEG signals by using multivariate empirical mode decomposition. Pattern Anal. Appl. 21, 81–89. doi: 10.1007/s10044-016-0567-6

CrossRef Full Text | Google Scholar

Morioka, H., Kanemura, A., Hirayama, J., Shikauchi, M., Ogawa, T., Ikeda, S., et al. (2015). Learning a common dictionary for subject-transfer decoding with resting calibration. NeuroImage. 111, 167–178. doi: 10.1016/j.neuroimage.2015.02.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Ni, T., Ni, Y., Xue, J., and Wang, S. (2021). A domain adaptation sparse representation classifier for cross-domain electroencephalogram-based emotion classification. Front. Psychol. 12, 721266, doi: 10.3389/fpsyg.2021.721266

PubMed Abstract | CrossRef Full Text | Google Scholar

Pan, S. J., Tsang, I. W., Kwok, J. T., and Yang, Q. (2011). Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 22, 199–210, doi: 10.1109/TNN.2010.2091281

PubMed Abstract | CrossRef Full Text | Google Scholar

Ramakrishnan, A. G., and Panachakel, J. T. (2021). Decoding covert speech from EEG-a comprehensive review. Front. Neurosci. 4, 642251, doi: 10.3389/fnins.2021.642251

PubMed Abstract | CrossRef Full Text | Google Scholar

Sheykhivand, S., Rezaii, T. Y., Mousavi, Z., Delpak, A., and Farzamnia, A. (2020). Automatic identification of epileptic seizures from EEG signals using sparse representation-based classification. IEEE Access. 8, 138834–138845, doi: 10.1109/ACCESS.2020.3011877

CrossRef Full Text | Google Scholar

Shu, L., Xie, J., Yang, M., Li, Z., Li, Z., Liao, D., et al. (2018). A review of emotion recognition using physiological signals. Sensors. 18, 2074. doi: 10.3390/s18072074

PubMed Abstract | CrossRef Full Text | Google Scholar

Sorkhabi, M. M. (2014). Emotion detection from EEG signals with continuous wavelet analyzing. Psychophysiology. 38, 912–925, doi: 10.1111/1469-8986.3860912

PubMed Abstract | CrossRef Full Text | Google Scholar

Wan, Z., Yang, R., Huang, M., Zeng, N., and Liu, X. (2021). A review on transfer learning in EEG signal analysis. Neurocomput. 421, 1–14, doi: 10.1016/j.neucom.2020.09.017

CrossRef Full Text | Google Scholar

Wang, J., Guo, Y., Guo, J., Li, M., and Kong, X. (2017). Synthesis linear classifier based analysis dictionary learning for pattern classification. Neurocomputing. 238, 103–113, doi: 10.1016/j.neucom.2017.01.041

CrossRef Full Text | Google Scholar

Wright, J., Yang, A. Y., Ganesh, A., Sastry, S. S., and Ma, Y. (2009). Robust face recognition via sparse representation. IEEE PAMI. 31, 210–227, doi: 10.1109/TPAMI.2008.79

PubMed Abstract | CrossRef Full Text | Google Scholar

Yan, K., Kou, L., and Zhang, D. (2018). Learning domain-invariant subspace using domain features and independence maximization. IEEE Trans Cybern. 48, 288–299, doi: 10.1109/TCYB.2016.2633306

PubMed Abstract | CrossRef Full Text | Google Scholar

Zanini, P., Congedo, M., Jutten, C., Said, S., and Berthoumieu, Y. (2018). Transfer learning: a riemannian geometry framework with applications to brain-computer interfaces. IEEE Trans. Biomed. Engineer. 65, 1107–1116, doi: 10.1109/TBME.2017.2742541

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, Q., and Lee, M. (2009). Analysis of positive and negative emotions in natural scene using brain activity and GIST. Neurocomputing. 72, 1302–1306. doi: 10.1016/j.neucom.2008.11.007

CrossRef Full Text | Google Scholar

Zhang, W., and Wu, D. (2020). Manifold embedded knowledge transfer for brain-computer interfaces. IEEE Trans. Neural Syst. Rehabilitation Eng. 28, 1117–1127. doi: 10.1109/TNSRE.2020.2985996

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, Z., Jiang, W., Qin, J., Zhang, L., Li, F., Zhang, M., et al. (2018). Jointly learning structured analysis discriminative dictionary and analysis multiclass classifier. IEEE Trans. Neural Netw Learn Syst. 29, 3798–3814, doi: 10.1109/TNNLS.2017.2740224

PubMed Abstract | CrossRef Full Text | Google Scholar

Zheng, W., Liu, W., Lu, Y., and Cichocki, A. B. L. (2019). Emotionmeter: a multimodal framework for recognizing human emotions. IEEE Trans Cybern. 49, 1110–1122, doi: 10.1109/TCYB.2018.2797176

PubMed Abstract | CrossRef Full Text | Google Scholar

Zheng, W., and Lu, B. L. (2015). Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Autonomous Mental Dev. 7, 162–175. doi: 10.1109/TAMD.2015.2431497

CrossRef Full Text | Google Scholar

Zhu, J., Shen, Z., and Ni, T. (2022). Multi-frequent band collaborative EEG emotion classification method based on optimal projection and shared dictionary learning. Front. Aging Neurosci. 2, 848511, doi: 10.3389/fnagi.2022.848511

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhuang, N., Ying, Z., Tong, L., Chi, Z., Zhang, H., and Yan, B. (2017). Emotion recognition from EEG signals using multidimensional information in EMD domain. BioMed Res. Int. 2017, 8317357, doi: 10.1155/2017/8317357

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: electroencephalogram signals, transfer learning, dictionary pair learning, emotion classification, across-subject

Citation: Ruan Y, Du M and Ni T (2022) Transfer Discriminative Dictionary Pair Learning Approach for Across-Subject EEG Emotion Classification. Front. Psychol. 13:899983. doi: 10.3389/fpsyg.2022.899983

Received: 19 March 2022; Accepted: 14 April 2022;
Published: 10 May 2022.

Edited by:

Yuanpeng Zhang, Nantong University, China

Reviewed by:

Yi Li, Qingdao University, China
Jianwu Wan, Hohai University, China
Yufeng Yao, Changshu Institute of Technology, China

Copyright © 2022 Ruan, Du and Ni. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Tongguang Ni, bnRnJiN4MDAwNDA7Y2N6dS5lZHUuY24=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.