Skip to main content

ORIGINAL RESEARCH article

Front. Comput. Neurosci., 21 February 2022
This article is part of the Research Topic Advances in Machine Learning Methods Facilitating Collaborative Image-based Decision Making for Neuroscience View all 6 articles

Emotion Recognition Based on Dynamic Energy Features Using a Bi-LSTM Network

\r\nMeili Zhu*Meili Zhu*Qingqing WangQingqing WangJianglin LuoJianglin Luo
  • Modern Animation Technology Engineering Research Center of Jilin Higher Learning Institutions, Jilin Animation Institute, Changchun, China

Among electroencephalogram (EEG) signal emotion recognition methods based on deep learning, most methods have difficulty in using a high-quality model due to the low resolution and the small sample size of EEG images. To solve this problem, this study proposes a deep network model based on dynamic energy features. In this method, first, to reduce the noise superposition caused by feature analysis and extraction, the concept of an energy sequence is proposed. Second, to obtain the feature set reflecting the time persistence and multicomponent complexity of EEG signals, the construction method of the dynamic energy feature set is given. Finally, to make the network model suitable for small datasets, we used fully connected layers and bidirectional long short-term memory (Bi-LSTM) networks. To verify the effectiveness of the proposed method, we used leave one subject out (LOSO) and 10-fold cross validation (CV) strategies to carry out experiments on the SEED and DEAP datasets. The experimental results show that the accuracy of the proposed method can reach 89.42% (SEED) and 77.34% (DEAP).

Introduction

Emotion mainly refers to a kind of inner emotional response produced by the psychological needs of people. It is a series of subjective psychological experiences formed by integrating the feelings, perceptions, and behaviors of people (Li, 2017). In the daily life and work of people, emotion is everywhere. In medical care, different nursing measures can be provided according to the emotions of patients to improve the quality of care. When treating psychological and physical diseases, an emotional state can be used as an auxiliary diagnostic basis. In the human-computer interaction, by identifying emotional states, the interaction is more friendly and natural. Therefore, the qualitative recognition and classification of emotional states of people have a high research value.

Psychological and emotional states can be transmitted in a variety of signal forms, such as via facial expressions, voice intonation, text, body posture, and other directly observable emotional signals. An emotional state of a person can also be learned through physiological signals, such as electrical brain signals, skin temperature, and breathing signals. Compared with observable emotions, electroencephalogram (EEG) signals are obtained by directly reading brain activity. EEG signals are more real and difficult to disguise. Therefore, the results obtained by using EEG signals to classify emotions are more authentic and reliable (Ullsperger et al., 2014; Naseer and Hong, 2015; Chen et al., 2020). As a result, this article mainly discusses studying emotion recognition methods based on EEG signals.

In EEG emotion recognition research, the first type of method is based on traditional machine learning technology. This kind of method first expands the emotional feature extraction from the EEG signal and then expands the corresponding emotional pattern learning and classification of the feature data. Emotional feature extraction of EEG signals can be divided into four categories, including time-domain features, frequency-domain features, time-frequency features, and EEG signal non-linear dynamic features. For example, Atkinson et al. proposed an EEG emotion recognition method based on feature selection and a kernel classifier (Atkinson and Campos, 2016). In this study, the statistical characteristics of the EEG signal, frequency band power of different frequencies, Hjorth parameter, and fractal dimension are used. The statistical features include median, SD, and kurtosis coefficient. The frequency bands of EEG signals use θ (4–8 Hz), low α (8–10 Hz), α (8–12 Hz), β (12–30 Hz), and γ (30–45 Hz). Then, a feature selection method based on mutual information and a kernel classifier are used to optimize the performance of the emotion classification task. Mohammadi et al. decomposed the EEG signal using a wavelet transform, while the entropy feature and the energy feature were extracted from each sub-band signal. Then, the support vector machine (SVM) and k-nearest neighbor (KNN) pattern classifiers were used for training and testing (Mohammadi et al., 2017). The results showed that the accuracy of arousal was 86.75%, and the accuracy of valence was 84.05%. Fernández-Varela et al. evaluated the emotion recognition performance of six machine learning models in detail as follows: Fisher’s linear discriminant, SVM, artificial neural network, classification tree, KNN, and naive Bayes (Isaac et al., 2017). Li et al. (2018) studied the effectiveness and performance of 9 time-frequency domain features and 9 non-linear dynamic features through the combination of different feature selection and SVMs. Abhishek Tiwari et al. proposed motif series and graph-theoretical features, before analyzing three feature selection methods [ANOVA-based feature ranking and selection, minimum redundancy maximum relevance (mRMR) feature selection, and recursive feature elimination (RFE)] and three fusion strategies (feature fusion, score-level fusion, and output-associated fusion). They finally concluded that motif series features are better than spectral features and that feature-level fusion strategies can improve recognition accuracy. Score-level fusion can improve arousal prediction (Tiwari and Falk, 2019). Fu Yang et al. constructed a multidimensional high-dimensional feature set by extracting eight linear features, such as the Hjorth parameter, power spectral density, and SD, and two non-linear features, such as wavelet entropy and sample entropy. Then, the ST-SBSSVM method was proposed by combining the significance test, sequential backward selection, and the SVM (Yang et al., 2019). The accuracy of the ST-SBSSVM was 72% (DEAP) and 98% (SEED).

The second type of method is based on deep learning technology. Deep learning has powerful feature extraction and classification capabilities. The main deep learning models used in the field of EEG signal classification include convolutional neural networks (CNNs), recurrent neural networks (RNNs), and generative adversarial networks (GANs). For example, Wang and Shang (2013) used a deep belief network (DBN) to extract features from the original physiological data and constructed three classifiers to estimate the arousal, valence, and liking of emotions. Yang et al. (2018) and Banee et al. (2019) put 32 electrodes into a 9 × 9 matrix according to the position of the international 10–20 system. The former used a time sliding window to divide the continuous EEG signal and calculated the average value of the corresponding electrode in each window. The latter directly inputs the original signal value into the matrix. The final input of the CNN network is a two-dimensional matrix containing spatial location information. Li et al. (2017) used a sparse wavelet transform to transform the EEG signals of each channel into a two-dimensional spectrum and then input them into a CNN-RNN hybrid network. Hwang et al. (2020) generated an egg image with topological information based on differential entropy (DE) features and then used a deep network structure combining a CNN and an LSTM. Shawky et al. (2018) converted EEG signals into a three-dimensional continuous frame input depth network, which compared EEG signals with images while ignoring the spatial position information of electrodes in EEG signals. Ant topic et al. proposed two feature mapping methods, called topography (TOPO-FM) and holography (HOLO-FM). Then, the deep learning technique was employed for feature learning from TOPO-FM and HOLO-FM, and finally, SVM was used to classify emotions (Topic and Russo, 2021). In the above research, the transformation of the EEG feature form imitated the image learning mode of deep networks, including a CNN, an RNN, and a DBN as much as possible. In addition, there are studies on EEG signal analysis by using brain function networks, such as energy landscape (Krzemiński et al., 2020; Klepl et al., 2021), which provides a new idea for the study of EEG signals.

Electroencephalogram signals are non-stationary, non-linear, and complex physiological signals with multicomponent mechanisms (Golyandina et al., 2001). In the machine learning method, the time domain, frequency domain, and time-frequency domain characteristics ignore the fact that the EEG signal is a non-linear signal, and non-linear dynamic features need to extract sufficient structural information of EEG components. In addition, the traditional machine learning method extracts features from multiple channels of EEG signals and then connects them to a single feature vector without considering the temporal dynamics. In the method based on deep learning, it is difficult to train a high-quality model due to the small number of samples and the low resolution of EEG images. In addition, the feature analysis and extraction of EEG signals are easily affected by noise, which impacts the effect of the deep learning model. Therefore, this study discusses combining the features extracted by manual calculation with a neural network to solve two problems. The first problem is simply and effectively representing multichannel EEG signals before establishing the emotion recognition model. The second problem is constructing effective features to learn temporal dynamics. We regarded brain emotion processing as an overall coordinated physiological response. To save artificial features while retaining the original information as much as possible, we proposed emotion recognition based on dynamic energy features using a Bi-LSTM network, and the main contributions are as follows:

(1) The concept of energy sequence is proposed. Dynamic feature extraction based on the energy sequence can effectively avoid the noise superposition effect caused by the feature extraction of a single channel sequence.

(2) The feature set is constructed by the combination of mutual information and principal component analysis (PCA). The feature set can reflect the time persistence of the EEG signal and the complexity of the multicomponent structure. The network model adopts fully connected layers and long short-term memory (LSTM) networks, which are suitable for small datasets.

Materials and Methods

Experimental Data

The datasets used in this study are SEED and DEAP. SEED (Duan et al., 2013; Zheng and Lu, 2015) is a collection of EEG datasets provided by the BCMI laboratory. The dataset recorded 62 channels of EEG data at a sampling frequency of 1,000 Hz (200 Hz resampling) and collected 15 individuals (7 men and 8 women), as well as three kinds of emotional EEG data (positive, negative, and neutral). The size of the dataset is 3 × 15 × 15 × 62 × 4,800 (3 experiments, 15 videos/trials, 62 channels, and (200 Hz × 240 s) data). The DEAP (database for emotion analysis using physical signals) was collected by Koelstra et al. (2012) from Queen Mary University in London, the University of Twente in the Netherlands, the University of Geneva in Switzerland, and the Federal Institute of Technology in Switzerland. The dataset recorded 40 channels of physiological signals at a sampling frequency of 512 Hz (128 Hz resampling) and collected the data of four emotions (valence, arousal, dominion, and liking) of 32 people (16 men and 16 women). The size of the dataset was 32 × 40 × 40 × 8,064 (40 videos/trials, 40 channels, and 128 Hz × 63 s).

Data Preprocessing

Single-channel EEG signals can achieve strong results in attention state detection, sleep state detection, and identity recognition. Taran and Bajaj (2019) also obtained good results in emotion recognition through a single-channel signal. Inspired by this, this study combines the spatial distribution of the whole head electrode to preserve the integrity of cerebral cortex information. The data preprocessing is shown in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. Data preprocessing.

First, the spatial information of different channels is fused, and the concept of a dynamic energy sequence is proposed as follows:

ETs(t)=1Ni=1Ns(t,i)

where N represents the channels and s(t, i) is the value of time t in the i-th channel. The energy sequence projects the time sequence in multiple channels onto an equal-length energy sequence.

Assuming that the time signal of the i-th channel iss(t, i)and the noise is n(t, i), the noise and the signal space are independent of each other and distributed independently. Then, the time signal affected by the noise is as follows:

s(t,i)=s(t,i)+n(t,i)

The average value of the energy sequence is as follows:

E(ETs(t))=E(1Ni=1N[s(t,i)+n(t,i)])=E(1Ni=1Ns(t,i)+1Ni=1Nn(t,i))=s¯(t)+n¯(t)

where s¯(t)represents the average value of the sequence and n¯(t) represents the average value of the noise.

The variance of the energy series is as follows:

σETs(t)2=E([ETs(t)-E(ETs(t))]2)
=E({1Ni=1N[s(t,i)+n(t,i)]-[s¯(t)+n¯(t)]}2)
=E({[1Ni=1Ns(t,i)-s¯(t)]+[1Ni=1Nn(t,i)-n¯(t)]}2)
=σETs(t)2+1Nσn(t)2
=1Nσs(t)2+1Nσn(t)2

Equations (2) and (3) show that the mean value of the energy sequence after noise influence is the same as that of the single-channel sequence, and the variance is 1/N of the single-channel sequence. In the traditional EEG sequence recognition process, each channel is extracted separately, and then it is recognized or trained. During the recognition process, the noise will be superimposed, and the noise interference is much greater than that of single-channel sequence noise. Therefore, the method of projecting the sequence of multiple channels to the energy sequence can effectively smooth the noise, avoid the noise superimposition effect caused by extracting the features of the single-channel sequence separately, and improve the robustness.

During the procedure of EEG signal processing, even if preprocessing methods are used, such as artifact removal, downsampling, filtering, and component decomposition, the EEG signal noise still remains. Therefore, this study ignores the asymmetry of the brain emotion processing function, regards brain emotion processing as an overall coordinated physiological response, projects the time sequences of multiple channels to the energy time sequence, and saves the cost of artificial feature extraction while retaining the original information as much as possible. In our later experiments, considering the different activation degrees of different channels, the contribution to the template is also different. Therefore, we divided the brain lobe into 4 regions, which are the left anterior region, right anterior region, left posterior region, and right posterior region. According to the conclusion in the study (Li et al., 2018; Zheng et al., 2019), the area with high emotional activation response is set with a high coefficient. The specific region division and coefficient setting are shown in Table 1.

TABLE 1
www.frontiersin.org

Table 1. Regional division and coefficient setting.

Second, emotion is a continuous physiological and psychological process and state, and past emotional states may have an impact on the current emotions. To combine the time persistence of the emotional state and obtain the dynamic attributes of the emotional characteristics of EEG signals as much as possible, we introduced the time window variable tw on the time axis. The moving time window method is used to intercept the sequence fragments from the time series. We calculated the energy sequence corresponding to the same time window.

For SEED, we intercepted the data from the 2nd min to the 3rd min, and for DEAP, we intercepted the data from the 4th to the 63rd min. The relevant data settings are shown in Table 2.

TABLE 2
www.frontiersin.org

Table 2. Data parameter setting.

Dynamic Energy Feature Set

Based on the dynamic energy sequence, we extracted different types of features from each energy time sequence and then form the feature set according to the time sequence. The feature set formed by different types of features extracted from the dynamic energy sequence is a time-domain profile of multicomponent features, which can reflect the time-domain dynamic characteristics of EEG emotional signals and the complexity of the multicomponent structure. The pseudocode of the dynamic energy characteristic calculation is shown in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2. Pseudocode for computing dynamic energy feature set.

In this study, seven kinds of time-domain features, three kinds of frequency-domain features, and three kinds of dynamical features are extracted.

Time-Domain Features

The EEG signal is a discrete sequence obtained by sampling a continuous signal according to a certain sampling frequency through an EEG device. Suppose that s(n) represents the EEG signal value obtained by sampling a certain electrode, n = 1, 2, 3 … , N. N represents the total number of samples, μs represents the average value, and σs represents the SD.

(1) The average value is calculated as follows:

μs=1Nn=1N-1s(n)

(2) The SD is calculated as follows:

σs=1Nn=1N(s(n)-μs)2

(3) The first-order differential is calculated as follows:

δs=1N-1n=1N-1|s(n+1)-s(n)|

The normalized first-order differential (Yao, 2015) is calculated as follows:

δs¯=δsσs

(4) The second-order differential is calculated as follows:

γs=1N-2n=1N-2|s(n+2)-s(n)|

The normalized second-order differential is calculated as follows:

γs¯=γsσs

(5) The instability index (Kroupi et al., 2011) is calculated as follows:

s(n) is divided into M segments, and the mean value of each segment is calculated. The SD of these M mean values is defined as NSI.

(6) The energy is calculated as follows:

Es=n=1N|s(n)|2

(7) The fractal dimension (Ansari-Asl et al., 2007) is calculated as follows:

We can express s(n) and n = 1, 2, 3 … , N as {s(m),s(m+k),,s(m+N-mkk)}, where m = 1, 2, 3 … , k is the start time and k is the time interval, which can be seen as follows:

Hm(k)=N-1[N-mk]k2i=1[N-mk]|s(m+ik)-s(m+(i-1)k)|FDs=-logH(k)¯logk

where H(k)¯ denotes the mean of Hm(k).

Frequency Domain Features

(1) Hjorth parameters (Bo, 1970):

First, the EEG signal is transformed by a discrete Fourier transform (DFT). Then, the Hjorth parameters are calculated in theta (4–7 Hz), alpha (8–13 Hz), beta (14–30 Hz), gamma (31–50 Hz), and the full band.

The DFT is calculated as follows:

S(k)=DFT[s(n)]=n=0N-1s(n)WNnk=n=0N-1s(n)e-j(2πN)nk

The activity is as follows:

Ac=1Nk=0N-1(S(k)-S(k)¯)2

The mobility is as follows:

Mo=var(S(k))var(S(k))

The complexity is as follows:

Co=Mo(S(k))Mo(S(k))

(2) Power spectrum and power spectral density:

In this study, the autoregressive model (AR mode) is used to estimate the power spectrum of EEG signals, and the process is as follows:

s(n)is expressed by the difference equation as s(n)=-k=1paks(n-k)+ω(n),

where ω(n)represents the white noise sequence with zero mean and σω2 variance. krepresents the order of the AR model.

The transfer function of the AR model is as follows:

H(z)=11+k=1pakz-k

The power spectrum is as follows:

Ps(ω)=σω2|A(ejω)|2=σω2|1+k=1pake-jωk|2

Dynamical Features

(1) Approximate entropy (Pincus, 1991):

Constructing the vector Xi(m)=(xi,xi+1,xi+2xi+m-1), 1 ≤ i ≤ N-m + 1, N represents the length of the time series, m represents the embedded dimension, then counts the number of vectors Cm(i, r) similar to vector Xi(m), and r is the similarity tolerance. The method is as follows:

Cm(i,j,r)={1,d(Xi(m),Xj(m))<r0,otherwise
Cm(i,r)=j=1N-mCm(i,j,r)

Now the formula is defined as follows:

Pim(r)=1N-m+1Cm(i,r),
Φm(r)=1N-m+1i=1N-m+1ln(Pim(i))

We can repeat the above steps to calculate Φm + 1(r); then, the approximate entropy is as follows:

ApEn=Φm(r)-Φm+1(r)

(2) Wavelet entropy (Sharma et al., 2015):

In this study, the DB4 wavelet transform is used to decompose the EEG signal in six levels, and then the Shannon entropy is extracted from the approximate component and the detail component.

(3) Differential entropy (Zheng and Lu, 2015):

DE=-+12πσi2e-(x-μ)22σi2log(12πσi2e-(x-μ)22σi2)dx=12log(2πeσi2)

Feature Dimension Reduction

Feature dimension reduction includes feature extraction and feature selection, which is very important for recognition and classification. The effective dimensionality reduction method can not only improve the speed of the algorithm but also filter out irrelevant attributes and reduce the interference of irrelevant information with classification to improve the accuracy of the classification and prevent overfitting.

The basic idea of feature dimensionality reduction is to score each feature vector according to some evaluation function, rank the feature vectors according to the score value from high to low, and take some feature vectors with a high score value as a feature set. In this study, we considered the effectiveness of a single feature label and the information redundancy between multiple features. We adopted the method of combining mutual information (Hossain et al., 2014) and PCA (Ian and Jorge, 2016) to reduce dimensionality. The process is as follows:

Suppose that the feature space of the sample dataset X is Rm×n, each row of data Xi is (xi1, xi2, xi3, ⋯ xin), and xij represents the j-th feature value of the i-th sample.

Step 1: Calculate the proportion of the value of the i-th sample on the j-th feature to the overall value of the j-th feature as follows:

p(xi)=xiji=1mxij

Step 2: Calculate the information entropy of the i-th sample as follows:

H(xi)=-i=1np(xi)logp(xi)

Step 3: Calculate the mutual information matrix as follows:

I(X,Y)=-i=1nj=1mp(xi,yj)logp(xi,yj)p(xi)p(yj)
Ixy=[I11In1In1Inn]

Step 4: The matrix decomposition is as follows:

Ixy=BΛB

where Λ is the diagonal matrix composed of the eigenvalues (μ1, μ2, ⋯ μn) of ∑Ixy and B is a matrix composed of eigenvectors (β1, β2, ⋯ , βn) corresponding to the eigenvalues.

Step 5: Calculate the contribution rate σk and the cumulative contribution rate δk as follows:

σk=μkk=1nμk,δk=k=1nσk

Step 6: Select new features.

The eigenvectors (β1, β2, ⋯ , βl) corresponding to the first k eigenvalues with 85–95% contribution rates are selected as the principal component decision matrix. The new feature matrix is as follows:

Z=BlX

The Network Structure

The network structure used in this study is shown in Figure 3. The architecture consists of a Bi-LSTM network, a dropout layer, an LSTM network, and a full connection layer output.

FIGURE 3
www.frontiersin.org

Figure 3. Schematic diagram of network structure.

Bidirectional Long Short-Term Memory

Bidirectional LSTM (Bi-LSTM) is based on LSTM (Hochreiter and Schmidhuber, 1997), which combines the information of the input sequence in both the forward and backward directions. In this study, the vector outputs by the two LSTM layers are connected and sent to the next layer. The Bi-LSTM is selected for the following reasons. First, mental emotion is a continuous physiological and psychological process and state. EEG signals are time-dependent. Therefore, emotion recognition of EEG signals may require the information of the whole input sequence, and the Bi-LSTM can better capture the context information in the sequence. Second, manual feature extraction is when the signal extracted from the EEG time series signals uses a sliding time window, which still belongs to a time series and can be modeled by the Bi-LSTM. Finally, for EEG signal recognition, the number of samples is small. Even when using data enhancement, it is difficult to train a good depth network model, and the Bi-LSTM is suitable for small datasets.

Dropout

To prevent overfitting, the dropout layer (Hinton et al., 2012) is linked after the Bi-LSTM layer. In each iteration, the neuron is randomly output to zero according to the specified proportion.

Long Short-Term Memory

To further extract features, the output data of the dropout layer is sent to the LSTM.

Parameter settings of the model are shown in Table 3.

TABLE 3
www.frontiersin.org

Table 3. Model parameter setting.

Results

To verify the feasibility and effectiveness of the proposed method, we performed four groups of experiments: comparison with traditional methods, feature type selection, individual emotion recognition, and comparison with other methods.

(1) Comparison with traditional methods:

The experiment is carried out by using the feature extraction method proposed in this study. The contrast method is the traditional method of extracting features from a single-channel signal. The methods are as follows: extracting features from signals in each channel, then using PCA for dimension reduction, and finally using SVM for recognition. To input the features into the SVM model, the feature vectors are composed of the mean value of the corresponding eigenvalues in the time window. In this way, the time-domain dynamics are considered to a certain extent. The relevant parameters of SVM are as follows: the type is C-SVC, the kernel function is RBF, and the setting method of penalty coefficient is Scan the parameter space 2[–3:8] in step 1 to search for the best value.

The experimental dataset is SEED, and positive and negative emotions are identified. The results are shown in Figure 4.

FIGURE 4
www.frontiersin.org

Figure 4. Comparison of recognition rates between proposed and traditional methods.

In this study, we first obtained the energy sequence of each channel and then extracted features from the energy sequence. Therefore, the feature dimension of the energy sequence is 1/62 of that of the single channel. As seen in Figure 4, the recognition rate of the proposed method is 72.5% and that of the traditional method is 72.2%, which shows that the proposed dynamic energy sequence can retain the characteristics of each channel of the EEG signal and effectively reduce the feature dimension.

(2) Feature type selection:

In this study, the data of 6 subjects in the seed dataset are randomly selected. According to the feature extraction method in the “Dynamic energy feature set” section, 49 types of features are extracted. Then, the MIPCA method is used to analyze the feature set F270×49. The cumulative contribution rate of the principal components is shown in Figure 5. In Figure 5, the x-axis represents the principal component serial number, and the y-axis represents the cumulative contribution rate. The y-axis red mark indicates that the cumulative contribution rate has reached 94.65, and the x-axis red mark indicates that the number of principal components when the cumulative contribution rate is 94.65. In our experiment, the principal components whose cumulative contribution rate is before 95% are selected to determine the final feature set. Considering that different subjects may have different suitable features, we conducted three experiments at random and conducted statistical analysis on the results. The final features are shown in Table 4.

FIGURE 5
www.frontiersin.org

Figure 5. Cumulative contribution rate of principal components.

TABLE 4
www.frontiersin.org

Table 4. Top features contributing to the prediction of emotion using the MIPCA method for SEED.

It can be seen from Table 4 that among the selected features, there are 8 time-domain features with the largest number. This shows that the time-domain features have a strong resolution. In the frequency domain, the features in beta, gamma, and all bands are better than those in other bands. This is consistent with the conclusion in the study by Li et al. (2018). For the DEAP dataset, in addition to the features in Table 4, we added the wavelet entropy (gamma) and the power spectrum sum (all).

(3) Emotion recognition:

In this study, we adopted two verification strategies, namely, leave one subject out (LOSO) and 10-fold cross validation (CV). For SEED, we classified two-dimensional emotions and three-dimensional emotions, and for DEAP, we identified the accuracy of arousal and valence. The results of SEED are shown in Figure 6. Figure 6A shows the recognition accuracy of two-dimensional emotion under the LOSO strategy. The average value of accuracy is 83.87%. Figure 6B shows the recognition accuracy of three-dimensional emotion under the LOSO strategy. The average value of accuracy is 69.76%. Figure 6C shows the recognition accuracy of two-dimensional emotion under the 10-fold CV strategy. Figure 6D shows the recognition accuracy of three-dimensional emotion under the 10-fold CV strategy.

FIGURE 6
www.frontiersin.org

Figure 6. The results of the SEED datasets: (A) leave one subject out (LOSO) for the 2-classification. (B) LOSO for the 3-classification classes. (C) 10-fold CV for the 2-classification. (D) 10-fold CV for the 3-classification.

The DEAP results are shown in Figure 7. Figure 7A shows the accuracy of arousal and valence under the LOSO strategy. The average value of arousal was 63.35% and the valence was 61.39%. Figure 7B shows that the accuracy of arousal and valence is under the 10-fold CV strategy. The average value of arousal is 77.3%, and the valence is 75.7%. The results of Figures 6, 7 verify the effectiveness of our method.

FIGURE 7
www.frontiersin.org

Figure 7. The results of the DEAP dataset: (A) the result of LOSO and (B) the result of 10-fold CV.

(4) Comparison with other methods:

We compared the results obtained by the proposed method with other methods that adopt the same dataset and adopt the same verification strategy. The comparison results are shown in Table 5.

TABLE 5
www.frontiersin.org

Table 5. Comparison results.

Discussion

In this study, we proposed the concept of an energy sequence to reduce the influence of noise caused by feature extraction and then construct a dynamic energy feature set including time-domain features, frequency domain features, and non-linear features. The network model adopts fully connected layers and LSTM networks, which are suitable for small datasets. In Figure 1, for the dynamic energy sequence feature set + SVM, the recognition rate is 72.5%. For the traditional feature set + SVM, the recognition rate is 72.2%, which shows that the dynamic energy sequence can preserve the features of the EEG signal and effectively reduce the feature dimension. To further extract effective features, this study comprehensively considers the effectiveness of a single feature for labeling and the information redundancy between multiple features. Through the feature type selection experiment, 17 kinds of feature vectors were selected from the SEED dataset, and 19 kinds of feature vectors were selected from the DEAP dataset. Finally, to verify the feasibility of our method and to facilitate the comparison with advanced methods, this study selects two verification strategies, namely, LOSO and 10-fold CV. For the SEED dataset, with 2-classification with LOSO, in the study by Li et al. (2018), the accuracy is 83.33%; in the study by Yang et al. (2019), it is 89%; and in our method, it is 83.87%. For the 3-classification problem, in the study by Zheng et al. (2019), the accuracy is 60.93%, and in our method, it is 69.76%. For the 2-classification with a 10-fold CV, in the study by Topic and Russo (2021), the accuracy is 88.45%, and in our method, it is 89.42%. For the DEAP dataset, with 2-classification with LOSO, in the study by Li et al. (2018), the average accuracy is 59.06%; for our method V, the average accuracy is 61.39%, and for A, it is 63.35%. With a 10-fold CV, the result of our method (V: 75.78%, A: 77.34%) is better than the result (V: 71.09%, A: 72.58%) in the study by Siddharth et al. (2019) and close to the result (V: 76.61%, A: 77.72%) in the study by Topic and Russo (2021). Therefore, our method is effective. Emotion is a physiological and psychological activity state induced by internal or external stimuli. Although different scalp regions are affected by emotion, when emotion changes, EEG signals will change harmoniously as a whole. Therefore, this study analyses the synchronization of signals in different regions as a whole, which is consistent with the phenomenon that emotional changes will lead to the overall change of EEG signals. The limitation of this study is that taking all channel signals as a whole cannot verify the relationship between different regions and different emotions. In the future study, we will continue to study the noise problem in EEG signal analysis and processing, as well as the relationship between different regions and emotions.

Conclusion

In this study, the concept of an energy sequence is proposed, and it is verified through theory and experiment that an emotional recognition method based on an energy sequence can reduce the influence of noise and improve accuracy. In addition, due to the small size of the EEG dataset, this study uses the combination of traditional features and a Bi-LSTM deep network. Compared with other traditional methods and CNN deep neural models, this method can achieve higher average accuracy.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://bcmi.sjtu.edu.cn/home/seed/downloads.html and http://www.eecs.qmul.ac.uk/mmv/datasets/deap/readme.html.

Author Contributions

MZ developed the method based on dynamic energy features, performed all the data analysis, and wrote the manuscript. QW advised data analysis and edited the manuscript. JL supervised and led the planning and implementation of research activities and revised the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This study was supported by the Jilin Science and Technology Development Plan Project (Grant No. 20210201129GX) and the Young Academic Backbone Program Jilin Animation Institute.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

The authors thank the editor and the reviewers for their useful feedback that improved this manuscript.

References

Ansari-Asl, K., Chanel, G., and Pun, T. (2007). A channel selection method for EEG classification in emotion assessment based on synchronization likelihood. Kos island: European Signal Processing Conference.

Google Scholar

Atkinson, J., and Campos, D. (2016). “Improving BCI-based emotion recognition by combining EEG feature selection and kernel classifiers”. Expert Syst. With Applic. 47, 35–41. doi: 10.1016/j.eswa.2015.10.049

CrossRef Full Text | Google Scholar

Banee, B. D., Pradeep, K., Debakanta, K., Saswat, K. R., Korra, S. B., and Ramesh, K. M. (2019). “A spatio-temporal model for EEG-based person identification”. Multimed. Tools Applic. 78, 28157–28177.

Google Scholar

Bo, H. (1970). “EEG analysis based on time domain properties”. Electroencephalogr. Clin. Neurophysiol. 29, 306–310.

Google Scholar

Chen, W., Chen, L. L., Song, Z. Z., Lou, X. G., and Li, D. D. (2020). “EEG-based emotion recognition using simple recurrent units network and ensemble learning”. Biomed. Signal Proc. Control 58:101756. doi: 10.1186/s12868-016-0283-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Duan, R. N., Zhu, J. Y., and Lu, B. L. (2013). “Differential entropy feature for EEG-based emotion classification,” in International IEEE/EMBS Conference on Neural Engineering (New Jersey, NJ: IEEE).

Google Scholar

Golyandina, N., Nekrutkin, V. V., and Zhigljavsky, A. A. (2001). Analysis of Time Series Structure: SSA and related techniques. Boca Raton, FL: CRC.

Google Scholar

Hinton, G. E., Srivastava, N., Krizhevsky, A., Sutskever, I., and Salakhutdinov, R. (2012). “Improving neural networks by preventing co-adaptation of feature detectors”. Comp. Sci. 3, 12–223. doi: 10.1109/TNNLS.2017.2750679

PubMed Abstract | CrossRef Full Text | Google Scholar

Hochreiter, S., and Schmidhuber, J. (1997). “Long short-term memory”. Neural Comput. 9, 1735–1780.

Google Scholar

Hossain, M. A., Jia, X. P., and Pickering, M. R. (2014). “Subspace Detection Using a Mutual Information Measure for Hyperspectral Image Classification”. IEEE Geosci. Rem. Sens. Lett. 11, 424–428. doi: 10.1109/lgrs.2013.2264471

CrossRef Full Text | Google Scholar

Hwang, S., Hong, K., Son, G. Y., and Byun, H. (2020). “Learning CNN features from DE features for EEG-based emotion recognition”. Pattern Anal. Applicat. 23, 1323–1335.

Google Scholar

Ian, T. J., and Jorge, C. (2016). “Principal component analysis: a review and recent developments”. Philosop. Transac. R. Soc. A 374, 20150202–20150202. doi: 10.1098/rsta.2015.0202

PubMed Abstract | CrossRef Full Text | Google Scholar

Isaac, F. V., Elena, H. P., Diego, A. E., and Vicente, M. B. (2017). “Combining machine learning models for the automatic detection of EEG arousals”. Neurocomputing 268, 100–108. doi: 10.3389/fnhum.2021.621493

PubMed Abstract | CrossRef Full Text | Google Scholar

Klepl, D., He, F., Wu, M., Marco, M. D., Blackburn, D., and Sarrigiannis, P. G. (2021). Characterising Alzheimer’s Disease with EEG-based Energy Landscape Analysis. IEEE J. Biomed. Health Informat. 2021, 1–1. doi: 10.1109/JBHI.2021.3105397

PubMed Abstract | CrossRef Full Text | Google Scholar

Koelstra, S., Mühl, C., Soleymani, M., Lee, J. S., Yazdani, A., Ebrahimi, T., et al. (2012). “DEAP: A Database for Emotion Analysis;Using Physiological Signals”. IEEE Transact. Affect. Comput. 3, 18–31.

Google Scholar

Kroupi, E., Yazdani, A., and Ebrahimi, T. (2011). “EEG correlates of different emotional states elicited during watching music videos,” in Affective Computing and Intelligent Interaction (Memphis, TN: ACII).

Google Scholar

Krzemiński, D., Masuda, N., Hamandi, K., Singh, K. D., Routley, B. C., and Zhang, J. X. (2020). Energy landscape of resting magnetoencephalography reveals fronto-parietal network impairments in epilepsy. Network Neurosci. 4, 1–45. doi: 10.1162/netn_a_00125

CrossRef Full Text | Google Scholar

Li, H. S. (2017). Psychology of consciousness. Beijing: China legal publishing house.

Google Scholar

Li, X., Song, D. W., Zhang, P., Hou, Y. X., and Hu, B. (2017). “Deep fusion of multi-channel neurophysiological signal for emotion recognition and monitoring”. Int. J. Data Mining Bioinform. 18, 1–27.

Google Scholar

Li, X., Song, D. W., Zhang, P., Zhang, Y. Z., Hou, Y. X., and Hu, B. (2018). Exploring EEG Features in Cross-Subject Emotion Recognition. Front. Neurosci. 12, 162–162. doi: 10.3389/fnins.2018.00162

PubMed Abstract | CrossRef Full Text | Google Scholar

Mohammadi, Z., Frounchi, J., and Amiri, M. (2017). “Wavelet-based emotion recognition system using EEG signal”. Neural Comput. Applicat. 28, 1985–1990. doi: 10.1109/EMBC46164.2021.9630188

PubMed Abstract | CrossRef Full Text | Google Scholar

Naseer, N., and Hong, K. S. (2015). “fNIRS-based brain-computer interfaces: a review”. Front. Hum. Neurosci. 9, 3–3. doi: 10.3389/fnhum.2015.00003

PubMed Abstract | CrossRef Full Text | Google Scholar

Pincus, S. M. (1991). “Approximate entropy as a measure of system complexity”. Proc. Natl. Acad. Sci. U S A. 88, 2297–2301.

Google Scholar

Sharma, R., Pachori, R. B., and Acharya, U. (2015). “An Integrated Index for the Identification of Focal Electroencephalogram Signals Using Discrete Wavelet Transform and Entropy Measures”. Entropy 17, 5218–5240.

Google Scholar

Shawky, E., El-Khoribi, R., and Shoman, M. A. I. (2018). EEG-Based Emotion Recognition using 3D Convolutional Neural Networks. Int. J. Adv. Comp. Sci. Applicat. 9:329.

Google Scholar

Siddharth, S., Jung, T. P., and Sejnowski, T. J. (2019). Utilizing Deep Learning Towards Multi-modal Bio-sensing and Vision-based Affective Computing. IEEE Transact. Affect. Comput. 2019, 1–1.

Google Scholar

Taran, S., and Bajaj, V. (2019). “Emotion recognition from single-channel EEG signals using a two-stage correlation and instantaneous frequency-based filtering method”. Comp. Methods Prog. Biomed. 173, 157–165. doi: 10.1016/j.cmpb.2019.03.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Tiwari, A., and Falk, T. H. (2019). Fusion of Motif- and Spectrum-Related Features for Improved EEG-Based Emotion Recognition. Computat. Intellig. Neurosci. 2019:3076324. doi: 10.1155/2019/3076324

PubMed Abstract | CrossRef Full Text | Google Scholar

Topic, A., and Russo, M. (2021). Emotion recognition based on EEG feature maps through deep learning network. Engine. Sci. Technol. Int. J. 24:12.

Google Scholar

Ullsperger, M., Fischer, A. G., Nigbur, R., and Endrass, T. (2014). “Neural mechanisms and temporal dynamics of performance monitoring”. Trends Cognit. Sci. 18, 259–267. doi: 10.1016/j.tics.2014.02.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, D., and Shang, Y. (2013). “Modeling Physiological Data with Deep Belief Networks”. Int. J. Informat. Educ. Technol. 3, 505–511. doi: 10.7763/IJIET.2013.V3.326

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, F., Zhao, X. C., Jiang, W. G., Gao, P. F., and Liu, G. Y. (2019). “Multi-method Fusion of Cross-Subject Emotion Recognition Based on High-Dimensional EEG Features”. Front. Computat. Neurosci. 13:53. doi: 10.3389/fncom.2019.00053

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, Y. L., Wu, Q. F., Qiu, M., Wang, Y. D., and Chen, X. W. (2018). Emotion Recognition from Multi-Channel EEG through Parallel Convolutional Recurrent Neural Network. Int. Joint Confer. Neural Netw. 2018:8489331.

Google Scholar

Yao, Z. P. (2015). Calculus. Huazhong: Huazhong Normal University press.

Google Scholar

Zheng, W. L., and Lu, B. L. (2015). “Investigating Critical Frequency Bands and Channels for EEG-Based Emotion Recognition with Deep Neural Networks”. IEEE Transact. Autonom. Mental Dev. 7, 162–175.

Google Scholar

Zheng, W. L., Zhu, J. Y., and Lu, B. L. (2019). “Identifying Stable Patterns over Time for Emotion Recognition from EEG”. IEEE Transact. Affect. Comp. 10, 417–429.

Google Scholar

Keywords: EEG, emotion recognition, dynamic energy feature, Bi-LSTM, energy sequence

Citation: Zhu M, Wang Q and Luo J (2022) Emotion Recognition Based on Dynamic Energy Features Using a Bi-LSTM Network. Front. Comput. Neurosci. 15:741086. doi: 10.3389/fncom.2021.741086

Received: 14 July 2021; Accepted: 31 December 2021;
Published: 21 February 2022.

Edited by:

Si Wu, Peking University, China

Reviewed by:

Fei He, Coventry University, United Kingdom
Abhishek Tiwari, Université du Québec, Canada

Copyright © 2022 Zhu, Wang and Luo. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Meili Zhu, zhumeilizhu@163.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.