- 1College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China
- 2School of Engineering and Applied Science, University of Pennsylvania, Pennsylvania, PA, United States
- 3Department of Auromation, Tiangong University, Tianjin, China
Previous studies have successfully applied a lightweight recurrent neural network (RNN) called Echo State Network (ESN) for EEG-based emotion recognition. These studies use intrinsic plasticity (IP) and synaptic plasticity (SP) to tune the hidden reservoir layer of ESN, yet they require extra training procedures and are often computationally complex. Recent neuroscientific research reveals that the brain is modular, consisting of internally dense and externally sparse subnetworks. Furthermore, it has been proved that this modular topology facilitates information processing efficiency in both biological and artificial neural networks (ANNs). Motivated by these findings, we propose Modular Echo State Network (M-ESN), where the hidden layer of ESN is directly initialized to a more efficient modular structure. In this paper, we first describe our novel implementation method, which enables us to find the optimal module numbers, local and global connectivity. Then, the M-ESN is benchmarked on the DEAP dataset. Lastly, we explain why network modularity improves model performance. We demonstrate that modular organization leads to a more diverse distribution of node degrees, which increases network heterogeneity and subsequently improves classification accuracy. On the emotion arousal, valence, and stress/calm classification tasks, our M-ESN outperforms regular ESN by 5.44, 5.90, and 5.42%, respectively, while this difference when comparing with adaptation rules tuned ESNs are 0.77, 5.49, and 0.95%. Notably, our results are obtained using M-ESN with a much smaller reservoir size and simpler training process.
1 Introduction
Emotions play an essential role in the human decision-making process (Picard, 1997). Automated recognition of emotions has broad applications in improving human-computer interactions (Atkinson and Campos, 2016), facilitating the diagnosis and treatment of affective disorders (Cai et al., 2020), and helping content providers enhance user experiences. Traditionally, emotional recognition relies on non-physiological cues such as facial expressions, speech, and behavior. However, subjects can inhibit these reactions, thus causing inaccurate classification results (Mauss and Robinson, 2009). Emotion recognition based on physiological activities such as EEG signals can circumvent this limitation.
Nonetheless, EEG is a complex signal that often requires advanced computational models to decode. Recently, a great number of machine learning and deep learning techniques have been applied to interpret EEG signals for emotion recognition (Li et al., 2022). Zheng and Lu (2015) utilized deep belief network and support vector machine to classify positive, negative, and neutral emotions. RNN frameworks, such as Long Short-Term Memory (Alhagry et al., 2017) and Spatial–Temporal RNN (Zhang et al., 2018), are also widely employed as the RNN structures are natural fits for sequential data like EEG signals. Convolutional neural networks (CNNs) (Lin et al., 2017; Chen et al., 2019) and hybrid architectures (Li et al., 2016) also obtained remarkable results. However, several challenges remain: (1) CNN-based architectures, despite their exemplary performance in detecting spatial information, their feed-forward design makes it less efficient in leveraging complex long-distance temporal patterns. (2) On the other hand, RNN-based models trade off their time and space complexity to mitigate the vanishing gradient problems. Therefore, it is challenging to implement them on wearable devices where computational resources are limited.
A special category of RNN called Echo State Network (ESN) (Jaeger, 2001a,b) offers a promising solution to the above-mentioned shortcomings. ESN contains a reservoir layer that simulates brain synaptic connections using a randomly connected network (see Figure 1A). The random connections non-linearly project the input into the internal states of the reservoir, making ESN particularly powerful on many time-series prediction and classification tasks. However, the reservoir layer is generally untrained due to the limitation of backpropagation on circular connections; only the last layer (i.e., the readout layer) is trained on the input data. This configuration allows ESN to exploit long-distance dependencies while maintaining exceptional computational simplicity, yet it also makes ESN performance particularly vulnerable to random initialization. In previous studies using ESN to decode emotional information in EEG signals, neural adaptation rules such as intrinsic plasticity (IP) (Schrauwen et al., 2008) and synapses plasticity (SP) (Oja, 1982; Castellani et al., 1999) were used to alleviate this deficiency. IP rule maximizes the information entropy in the reservoir by adjusting the activation function of the reservoir, while SP rules update the connections weights to make reservoir dynamics more stable.
Existing attempts using ESN for EEG-based emotion recognition are 2-fold. The first kind adopts ESN as EEG feature selectors (Koprinkova-Hristova et al., 2015; Bozhkov et al., 2016). In this setting, an IP-tuned ESN takes EEG signals to generate a lower-dimensional representation, while a clustering algorithm was applied to the n-leading lower-dimensional features to classify emotions. Bozhkov et al. (2017) showed that IP-tuned ESNs outperform regular ESNs because they can yield more separable features in the hyper-plane. On the other hand, the second type adopts ESN directly as classifiers (Fourati et al., 2017), in which ESNs predict emotion labels in an end-to-end manner. Several adaptation rules such as IP, anti-Oja, and BCM rule were employed on ESN (Fourati et al., 2020b) for better network robustness. An ESN with leaky integrators and multiple reservoir layers called Leaky DeepESN-IP (Fourati et al., 2020a) was also investigated to mitigate the impacts of inter-subject variability.
Nonetheless, we argue that these ESN methods can be further improved. ESN resembles the biological brain using a random network. Since biological brains are far from being entirely random (Bullmore and Sporns, 2009; Bullmore and Sporns, 2012; Tripathy et al., 2013), this over-simplification may hamper ESN performance. Moreover, even though adaptation rules are readily available, their algorithm and training procedures are complex. In an effort to resolve the above limitations, there are several pioneering works that attempt to initialize ESN to a predefined topology. Deng and Zhang (2007) proposed scale-free highly clustered ESN, an ESN with neurons distributed in a scale-free structure. Similarly, Li et al. (2016) generated a multi-clustered structure by first placing pioneering nodes in a three-dimensional space, then adding new nodes and connecting them to their closest neighbor based on their Euclidean distance. Rodriguez et al. (2021) initialized ESN to a two-module structure for optimal information diffusion. They demonstrated that for ESN using step-activation functions, there exists an optimal level of modularity that will maximize the model performance on the memory capacity task and the recall task.
The main advantage of ESN model is its simple structure, and there is no need to train the weights inside the reservoir. However, the general ESN network mainly adopts a random structure, which is not a structure with optimal performance. In order to further improve the performance of ESN, the weights inside the reservoir have been trained in some studies, including synaptic plasticity as well as intrinsic plasticity methods. Although these methods can improve the performance of ESN, they also increase the complexity of training. Modularity is a prevalent structure in brain neural networks, which is the result of the evolution and optimization of brain neural networks; therefore, this study directly draws on the modular structure of brain networks, which promotes the performance of ESN, while the design is relatively simple.
Motivated by several recent findings indicating brain modularity improves cognitive abilities (Crossley et al., 2014; Bertolero et al., 2018), in this proposed work, we continue to explore the possibility of adopting modular structure in ESN. Differing from the previous study (Rodriguez et al., 2021), which uses a single parameter to optimize the modular structure, we propose a novel triple-parameter optimization method for modular ESN. Our method allows flexible control over the module number, intra-module connectivity, and inter-module connectivity. We present a comprehensive experimental analysis of how each of these parameters impacts network performance. The proposed Modular Echo State Network (M-ESN) demonstrates significant performance improvement on the DEAP benchmark for EEG emotion classification without any neural adaptation procedure. In summary, our major contributions are as follows:
• We propose a bio-inspired lightweight framework for EEG classification, namely M-ESN. We present a novel triple-parameter method for defining and optimizing the modular topology.
• On the DEAP benchmark, our proposed M-ESN with 600 neurons outperformed the neuron-adaptation-rule-tuned ESN with 1,500 neurons by 0.77, 5.49, and 0.95% on the arousal, valence, and stress/calm classification tasks.
• We provide explanations for the modularity enhancing ESN performance from several perspectives. We found that the increasing modularity will induce changes in network topology. We show that these topological changes increase network heterogeneity and thereby improve the information processing efficiency.
2 Related works
2.1 Modularity in biological neural network
Modular structure, also referred to as community structure, describes networks in which neurons are more interconnected within their own cluster and less connected to neurons in other clusters (see Figure 1B). Modular structure is found in the neuronal system of many species, ranging from nematode C. elegans to mammals (Kim and Kaiser, 2014), suggesting it may confer genetic advantages.
This structure is also confirmed by well-establishing neuroscientific evidence (Sporns et al., 2005; Meunier et al., 2010; Sporns and Betzel, 2016). Structural evidence shows neurons within the same anatomical region in human brains are densely connected by synapses, while long-distancing white matter tracts sparsely connect these segregated regions to enable interregional information transfer (Avena-Koenigsberger et al., 2017). Moreover, the functional modeling of the human brain, which characterizes neuronal dynamics, also appears to be modular (Zhou et al., 2006). Finally, various anatomical regions typically have diverse gene expressions (Krienen et al., 2015), indicating brain is evolved in a modular manner to support function specialization (Bassett et al., 2010; Taylor et al., 2017; Figure 1A).
2.2 Modularity promotes network efficiency
Emerging neuroscientific findings suggest modular network promotes network efficiency. Crossley et al. (2014) investigated the modularity in brain functional coactivation networks, in which they concluded that brain modules promote cognitive specialization. Bertolero et al. (2018) offered indirect evidence that brain modularity enhances cognitive abilities. They investigated the role of hub structure, which is ubiquitous in modular networks (Bullmore and Sporns, 2009), on cognitive ability. They indicated that individuals with more diversely connected hubs perform universally better on several cognitive tasks. Lastly, several studies have reported that modular structure promotes the reuse of recurring network patterns (Kashtan and Alon, 2005) and network adaptability (Clune et al., 2013).
To summarize, existing literature mainly attributes modular topology to (1) leveraging the brain wiring costs and the efficacy of the information propagation (Betzel et al., 2017); (2) providing functional versatility and adaptability (Bassett et al., 2011); and (3) enabling structural flexibility that allows the neural system remain unaffected by local modifications (Kim and Kaiser, 2014). Motivated by these findings, we adopt the modular structure in ESN to improve the model performance.
2.3 EEG-based emotion recognition using hybrid CNN and LSTM classification
Electroencephalography (EEG)-based emotion classification is an important research area in emotion recognition, with a significant challenge being the individual differences and temporal variability in EEG recordings. Shen et al. (2020) proposed a novel four-dimensional convolutional recurrent neural network method to address this issue, effectively integrating frequency, spatial, and temporal information to overcome the interference caused by individual and temporal variability, thereby significantly improving the accuracy of emotion recognition. However, this method may face issues of high model training complexity and computational costs. Zhang et al. (2020) explored the potential of emotion recognition using multiple deep learning architectures, and validated the effectiveness of the CNN-LSTM hybrid model in processing EEG signals. This model successfully leveraged the advantages of CNN in feature extraction and the ability of LSTM in handling long-term dependencies in time-series data, but it may also lead to overfitting, especially with limited data. Additionally, Zamani and Wulansari (2021) and Samavat et al. (2022) presented models based on 1D-CNN, RNN, and a multi-input hybrid model based on CNN and Bi-LSTM, focusing on the analysis of specific frequency and temporal features. They enhanced the performance of emotion recognition by capturing temporal information more comprehensively through bi-directional LSTM, despite the high complexity in model structure design and the need for a large amount of annotated data for training. As for the hybrid emotion model proposed by Patlar Akbulut (2022), the utilization of transfer learning with large-scale sensor signals further improved the accuracy of emotion classification, highlighting the advantages of multimodal data and transfer learning in overcoming subject differences and data scarcity issues. Cimtay and Ekmekcioglu (2020) focused on the utilization of pre-trained CNN models, exploring the potential of cross-subject and cross-dataset emotion recognition, thus avoiding the resource consumption of training models from scratch. However, pre-trained models may face challenges in transfer efficiency and fine-tuning precision.
3 Methods
3.1 Dataset
We used the famous DEAP benchmark (Koelstra et al., 2011) on our model to provide fair comparison results with other existing methods. The DEAP dataset is collected when 16 male and 16 female participants watching a collection of 40 music videos. While the subjects are watching video clips, 32 channels of EEG signals and eight channels of peripheral signals are collected at 512 Hz and down-sampled to 128 Hz. Only the first 32 channels, i.e., the EEG channels, are used in our study. Each video clip lasts 63 s. The first 3 s of pre-trail baseline are removed for all trails. Therefore, each trail has a size of 60 × 128 × 32 (seconds × sampling rate × channels). The data labels are obtained by prompting participants to complete the self-assessment manikins (Bradley and Lang, 1998) to rate their level of emotional valence, arousal, dominance, and liking after each video clip. A total of 32 × 40 (participants × video clips) trails and corresponding labels are included in the dataset.
Establishing evidence has shown that EEG activity is rhythmic (Zheng and Lu, 2015). Such rhythmic patterns are most demonstrable in five frequency sub-bands, which are delta (1–4 Hz), theta (4–8 Hz), alpha (8–12 Hz), beta (12–30 Hz), and gamma (30–45 Hz). Since the DEAP dataset applied a 4–45 Hz band-pass filter after collecting the signals, the delta band is not included in our study.
3.2 Data preprocessing
A great body of literature has employed a feature extraction procedure (Subasi, 2007; Liu and Sourina, 2013; Zheng et al., 2017; Fourati et al., 2020b) for EEG signals before classification to improve EEG classification performance. In accordance with previous research (Fourati et al., 2020b), we used the power features of EEG signals for classification, which compute the energy of sub-bands over the energy of the entire frequency band. Our data preprocessing method consists of two procedures:
• Data cropping: first, to generate a sufficiently large dataset, we apply a non-overlapping sliding window for all trails to crop the original EEG samples into smaller data segments. For different classification tasks, we use different lengths of the sliding window. The selection of sliding window size will be explained in detail in section 4.
• Extract frequency bands features: then, the Fast Fourier Transform (FFT) was employed to transform data segments into frequency band features. L2-Normalization is applied to each data segment to reduce the inter-subject variability of EEG signals and prevent overfitting. After data segments were transformed into frequency sub-bands, the energy of each sub-bands is computed using Eq. 1. While and are the higher and lower bond of the current frequency band, is the normalized frequency band signal. The extracted feature (Eq. 2) is obtained by dividing sub-band energy by the total band energy. For each channel, we extract four features from four frequency bands. In this manner, each data segment with 32 channels is transformed into a feature vector with data points.
3.3 Memory capacity
Memory capacity (MC) refers to the network’s ability to store and retrieve past information. It serves as a measure of the echo state network’s capability to store and reconstruct past information. According to the work of Jaeger (2001a,b), memory capacity (Eq. 3) is defined as the maximum possible retrospective of an independently and identically distributed input sequence. Specifically, memory capacity is measured by assessing the correlation between the network output and previous input. For an echo state network with N units and identity activation function, the upper bound of its memory capacity is N.
The formula for memory capacity is as follows:
Here, represents memory capacity, indicates the covariance of two time series, denotes variance, is the input shown k steps before the current input, and is its reconstruction at the network output, where is the weight vector of the k-th output unit.
3.4 M-ESN
3.4.1 ESN backbone
We use ESN with leaky integrator neurons as the model backbone (Jaeger, 2001a,b). In contrast to our proposed M-ESN, it is referred to as basic ESN in the rest of this paper. A basic ESN has three layers: input layer, reservoir, and output layer (Figure 2). The connection weights between the input layer and reservoir, within the reservoir, and between reservoir and readout layer, are denoted by , , and , respectively (Figure 2). , , are randomly generated, while is scaled such that its spectral radius (i.e., the largest absolute eigenvalue of ) . This operation is suggested by Lukosevicius and Jaeger (2009) to guard the Echo State Property (ESP) so that information from prior states will be asymptotically washed out. The internal states of the reservoir using Eq. 4 are updated as follows:
Where and denotes the reservoir states at the time and . denotes the activation function of the reservoir layer, which usually is a hyperbolic tangent function, and is the leak rate that controls the update speed of the reservoir. and its corresponding label at time are stacked into and . The readout weights are calculated through ridge regression using Eq. 5, where is an identity matrix and is the regularization term introduced to prevent overfitting.
Finally, the predicted classes are obtained using Eq. 6 as follows:
3.4.2 Modular structure
Our M-ESN improves basic ESN (Figure 2) by replacing its random reservoir layer with a modular network (Figure 3C). Existing literature introduced a single parameter method (Nematzadeh et al., 2014) to modularize an ESN with two communities (Rodriguez et al., 2021). Under this setting, edges are generated according to a pre-defined global connection density . Then, fractions of edges bridge nodes within the same community and fractions of them connect nodes across communities (see Figure 1B). Since the number of modules remains fixed and the inter-community connectivity is dependent on the intra-community connectivity , their proposed ESN is optimal in a one-dimensional parameter space. In other words, the ESN is optimized with respect to intra-community connectivity .However, as the modular structure is collectively determined by the number of modules, intra-and inter-community connectivity, it may be insufficient to use only one parameter. Thus, we propose in this study to optimize M-ESN with a more comprehensive approach. We use to govern the intra-community density and to govern global connectivity. is the probability of arbitrary two nodes being connected within the same community, while governs the likelihood of a link exists between two nodes in different communities. We set less than or equal to as inter-community connections are often sparser than intra-community connections. When equals , the reservoir is equivalent to a random network. The modularity controls the module counts within reservoir. These three parameters together allow more flexible control over modularity and global connectivity, such that our M-ESN is optimized in three-dimensional parameter space ( , , and ). A complete generation process of M-ESN is described in Figures 3A–C.
Figure 3. (A) N internal neurons are first generated and randomly assigned to M clusters. (B) Then, intra-community connections are added according to . (C) Cross community connections are added according to and the reservoir layer is connected to the input layer and the output layer. (D) The flow chart of a complete experiment process. In Phase I, the optimal parameters are searched in the given parameter space. In Phase II, the optimized M-ESN is trained and tested.
3.4.3 ESN parameters
We use the M-ESN with 600 neurons to perform the experiments. In the previous step, an adjacency matrix based on , , and will be generated. Then, connections were added according to that adjacency matrix. We use bidirectional connections such that each connection has a corresponding inverse connection. The weights of these inverse connection pairs are randomly assigned according to a normal distribution within range (0, 1), with approximately 50% of these pairs are in opposite sign. We use a activation function for the reservoir layer. The input weights are uniformly distributed, with values scaled to (−0.1, 0.1), such that the neural dynamics will be more distinguishable by the activation function.
Finally, implementing a modular structure requires the network topology and connection weights to remain unchanged during the course of the training. Consequently, adaptation rules are not included in this study.
4 Experiment
We test M-ESN on three tasks, that is, to discriminate emotional valence, arousal, and stress/calm. The DEAP dataset describes emotional status in the following four domains: valence, arousal, dominance, and liking. Each of the four domains is represented by a numerical rating from 1 to 9. In accordance with prior research (Fourati et al., 2020b), we classify numerical values as high arousal/valence (HA/HV) and those as low arousal/valence (LA/LV). The calm and stress label is defined collectively by numeric valence and arousal levels, where a signal is classified as stress if and , and as calm if and .As discussed in section 3.2, each 60-s trail is sliced into data segments using sliding windows. For the arousal and valence classification task, the sliding window is set to 6 s. However, only 2,790 data segments have a stress/calm labels when using the 6 s sliding window. Thus, a smaller sliding window (t = 2 s) is used for stress/calm discrimination task to generate a larger dataset (n = 8,370). Our method contains two steps. In the first step, we fix the spectral radius and leakage rate of M-ESN to search for optimal combinations of modules count, local and global connection density on the valence discrimination task. In the second stage, the optimized modular structure is set fixed. We tune the ESN using the leakage rate for different classification tasks. A complete procedure is demonstrated in Figure 3D.
4.1 M-ESN optimization
It is well-established that ESN performance is determined by the reservoir topology (Zhang et al., 2011). For M-ESN, the reservoir topology is governed by , , and . Hence, in the first phase of the experiment, we fix the spectral radius of M-ESN to 0.85 and the leakage rate to 0.25, as they work best on basic ESN. Then, we search for the parameters that optimize M-ESN in the three-dimensional discrete parameter space ( , , ). The ESN is assessed on the valence discrimination task. The pseudo-code of this process is described in Algorithm 1.
We perform 5-fold cross-validation to reduce the bias introduced by the random train-test split. This 5-fold cross-validation process is conducted by: (1) splitting the dataset into five non-overlapping partitions; (2) leaving one partition out as the test set without repetition and using the remaining four partitions to train the model; and (3) repeating the previous two steps five times and averaging the accuracies. Given that ESN performance is constantly haunted by the random initialization of the reservoir layer, we repeat 5-fold validation three times and average the accuracies. For each value, we plot the model performance with respect to M values and values in a 2D contour diagram as depicted in Figure 4 to test all parameter combinations iteratively. We plot the 2D diagram using M as y-axis and as x-axis. The highest classification accuracy at each value is displayed in Table 1. The model is optimized when =0.05, =0.02, and =6.
Figure 4. 2D accuracy diagram with respect to and at various value. The classification accuracies are at its highest value when = 0.05, = 0.02, and = 6.
The structure and weights of modular reservoir of ESN remain unchanged during training, on the one hand, because this modularized reservoir is optimally designed with optimal classification accuracy, on the other hand, because the reservoir is an RNN structure, its weights are not trained, avoiding the problems of gradient vanishing and gradient explosion, reducing the training complexity of the ESN, and making the ESN simpler to implement, which is also an inherent advantage of ESN.
4.2 Classification results
Once the optimal parameters are found, we fix the three modular parameters and fine-tune the M-ESN on different leaking rates for three classification tasks. The optimal leaking rate for valence, arousal, and stress/calm classification are 0.25, 0.3, and 0.25, respectively. Then, we train the M-ESN using the optimal parameter combinations to predict test set labels. We plot the confusion matrix with the predicted classes and their actual labels to calculate the precision, sensitivity, specificity, accuracy, and F1-score (Figure 5). As demonstrated in Table 2, after the modular structure is introduced, we observe an increment in almost all metrics.
Table 2. Comparing ESN and M-ESN on three emotion recognition tasks with optimal parameters in terms of precision, sensitivity, specificity, accuracy, and F1-score.
In Table 3, we compared our results with ESN-IP, ESN-anti-Oja, and ESN-BCM presented by Fourati et al. (2020b). In their work, they provided results trained with raw signals and extracted features. Both kinds of input are trained with three different training schemes. To ensure a fair comparison, we only selected their results trained with frequency band features using the offline mode, which are consistent with our approach. Notably, our results are obtained using M-ESN with the size of 600, while their ESNs have 1,500 internal neurons. In all three tasks, our M-ESN outperforms the plasticity-rules-tuned ESNs with a considerable improvement. This result suggests that the modular structure may provide structural advantages for network performance.
Our results provide several insights. First, as shown in Table 1, the highest prediction accuracy at each value decreases monotonically as increases. This pattern aligns with previous studies that suggest reservoir sparsity has a significant impact on ESN performance (Zhang et al., 2011). Since the maximum value of equals , a lower will have a lower maximum , resulting in a sparser reservoir.
Second, a protruding pattern in our result (Figure 4) is that there appears to be a huge “jump” in its classification accuracy when the number of modules increases from 1 to 2 (Figure 4). As , , spectral radius, and the leaking rate remain unchanged, we conclude that the introduction of modular structure is responsible for this improvement. Moreover, the best performing M-ESNs have better scores than the best performing basic ESN on almost all evaluation metrics (Table 2).
Furthermore, as depicted in Figure 4, classification accuracy decreases as intercommunity connectivity increases. When , the modular structure disappears as inter-community connectivity equals intra-community connectivity. This further support that modular structure contributes to the improved classification accuracy.
4.3 Memory capacity
The classification performance of ESNs heavily depends on the encoding capability of the reservoir. In order to provide an explanation for the modular optimization design results shown in Figure 4, we further searched for the optimal values of the module parameters based on the reservoir’s memory capacity as the optimization criterion. As shown in Figure 6, it can be observed that the reservoir achieves the maximum memory capacity when = 0.05, = 0.04, and = 5. This optimization result is in close agreement with the results obtained using classification accuracy as the optimization criterion in Figure 4. The discrepancies between the two results can be attributed to certain random factors. Furthermore, when using classification accuracy as the optimization criterion, the weights of the output layer also influence the classification accuracy.
Figure 6 demonstrates that the values of , , and have significant impacts on the memory capacity of ESNs. Firstly, overall, a smaller value of the intra-module connection probability leads to a larger memory capacity, indicating that a sparser intra-module connectivity promotes the enhancement of memory capacity. Secondly, with an increasing number of modules ( ), the memory capacity initially increases and then decreases, implying that too many or too few modules are not conducive to memory capacity, and there exists an optimal number of modules. A larger number of modules may result in longer information propagation paths, thereby increasing the signal propagation delay during memory processes and reducing the memory capacity. Additionally, an increasing inter-module connection probability also demonstrates an initial increase and then decrease in memory capacity. This phenomenon may be attributed to the fact that when is small, its increase benefits information transfer between different modules, while excessively large values of weaken the modular characteristics, potentially decreasing the network’s memory capacity. In conclusion, the modularization parameters affect the memory capacity in the same way as the classification accuracy.
4.4 Network heterogeneity
However, it is still unclear why modular architecture improves network performance. To explain the working mechanism of modular structure in M-ESN, we examine the microscopic changes induced by modular topology, i.e., how modular structure affects neuronal degrees and connection strengths. We found that the modular structure alters the number of in-degree connections for individual neurons. Figure 7 compares the distribution of in-degree connections between a basic ESN and an M-ESN. It appears to be a significant change before and after introducing modular structure. The in-degree connection of most neurons in basic ESN falls between (15, 45), while most neurons in M-ESN have in-degree connection closely distributed around zero.
Figure 7. (A) In-degree distribution of M-ESN. (B) In-degree distribution of M-ESN with zoomed-in y-axis. For both ESN and M-ESN: = 0.05, N = 600. For M-ESN: = 6.
In Table 4, quantitative results reflect more detailed insights. Although both the variance and the mean of neuron degrees for M-ESN are smaller than basic ESN, the Standard Deviation (SD) over the mean for the M-ESN is significantly larger. Dividing SD by mean is referred to as Coefficient of Variation (CV), which is commonly employed to measure the network heterogeneity in previous literature (Ju et al., 2013; Litwin-Kumar et al., 2017). Comparing to the CV of 5.48 for basic ESN, M-ESN has a CV of 304.17, suggesting M-ESN becomes much more heterogeneous after adopting the community structure.
This increased network heterogeneity provides several benefits. First, in the field of neuroscience, it is well-established that neurons in brain are far from being homogeneous (Bullmore and Sporns, 2012; Tripathy et al., 2013). A more heterogeneous network can better simulate biological neural networks. Second, neurons in homogenous networks will yield similar features as they have similar connection weights and degrees. A more heterogeneous network, however, can provide richer features so that they can be more discernable in the hyperspace (Zeldenrust et al., 2021). Similarly, as M-ESN is more heterogeneous, its internal states will be more discernable by ridge regression in the readout layer. This finding is consistent with numerous studies in neuroscience and network science that suggest heterogeneous networks enhance cognitive abilities in brains and model performance in ANNs (Bullmore and Sporns, 2012; Volo and Destexhe., 2021; Zeldenrust et al., 2021).
In addition, a mean in-degree connection close to zero (0.0032) suggests that the majority of neurons within M-ESN have only 1 or 2 incoming connections. A larger CV and variance, however, imply that there are a few neurons that have more than average connection numbers. This network structure with many sparsely connected neurons and a few densely connected cores forms a hub structure. While previous neuroscientific finding (Bertolero et al., 2018) suggests the hub structure facilitates cognitive abilities in human brains, our research extends their studies and proves that hub structure also aids in the efficiency of M-ESN.
Lastly, our finding offers complimentary insights into previous literature (Rodriguez et al., 2021). They indicate increasing modularity will lead to decreasing network performance on ESN using non-step-like activation functions. In our result, however, when modular structure is introduced in our activated M-ESN, we observed a substantial increase in accuracy. We believe this divergence may be task-relevant. Rodriguez et al., 2021 researched a modularized ESN on memory capacity tasks, in which task the model performance is highly dependent on the level of information diffusion within the reservoir. When using step-like activation functions, community structures serve as containers that reduce noise and promote signal. However, since information diffusion will always occur when using non-step-like activation functions, increasing modularity cannot improve the memory capacity in their case. In contrast, we utilized M-ESN for classification tasks, which are highly sensitive to small temporal changes. Non-step-like activation functions broadcast such small changes to the entire reservoir. Consequently, in our experiments, community structures may serve to provide more diverse representations of the original signals, making them more discernable in the hyperspace.
5 Discussion and conclusion
A recent study (Kasabov, 2014) proposed a new SNN architecture, called NeuCube, based on a 3D evolving SNN learning from Spatio-and spectro-temporal brain data (STBD) and creates connections between clusters of neurons that manifest chains of neuronal activity. In the present study, the design of modular reservoir structure is also inspired the spatial structure information of the brain network, and we will consider more deeply how to introduce the spatial information into the design of the reservoir in our future research.
FORCE training (Depasquale et al., 2018) is a common method for training ESN, which can train the weights within the reservoir network, as well as the weights of the input and output layers. In this study, drawing on the fact that the connection structure of brain neural networks is modular, the reservoir adopts a modular structure, and its connection structure and weights are determined during initialization, and remain unchanged during training, so modular reservoir retains the advantages of simple training of reservoir computation, and also has a good performance of information processing. FORCE training can be considered in future research to train the weights of the reservoir and used in EEG classification tasks.
In this study, inspired by brain science, we obtained a remarkable improvement in the classification accuracy by adopting the modular structure in ESN for EEG emotion classification. Our work demonstrated the feasibility and superiority of replicating biologically observed structures on RNN to improve model performance. The main reason why the modular ESN performs better than the regular ESN is that its structure is optimized. The reservoir of the regular ESN adopts a randomly connected structure, and the modular ESN adopts the modular structural features of the brain neural network, which improves the performance of the ESN in the EEG classification task. We offered explanations of modularity enhance network performance and reported that neurons become more heterogeneous as the network becomes more modular. The performance enhancement of the modular ESN lies in the stronger structural heterogeneity, and larger coding capacity of the modular reservoir.
The advantage of ESN is that the model size is relatively small and has good performance. The increase of the reservoir size does not necessarily improve the performance of the ESN, too large a reservoir size will saturate the performance of the ESN, and even sometimes lead to overfitting of the ESN. The reservoir of ESN adopts a modularized structure, and the weights inside the reservoir are not trained, which greatly reduces the complexity of ESN training, and the design is relatively simple. Compared to the random reservoir of regular ESN, the modularized reservoir has higher structural heterogeneity and larger coding capacity, thus enhancing improvement of the ESN performance. Future studies could potentially adopt different intra-community connectivity in different communities or seek to reproduce more complex brain structures.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Author contributions
LY: Writing – original draft, Writing – review & editing. ZW: Writing – original draft, Writing – review & editing. GW: Writing – review & editing. LL: Writing – review & editing. JW: Writing – review & editing. ML: Writing – original draft.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported by the National Natural Science Foundation of China (Grant No: 61876132), Shen-zhen Science and Technology Major Project (Grant No: KJZD20230923114615032), the major project of Science and Technology Research Program of Chongqing Education Commission (Grant No: KJZD-M202302001), and the educational reform research project of Shenzhen Technology University (Grant No: 20222011).
Acknowledgments
This work was supported by the National Natural Science Foundation of China (Grant No: 61876132), Shenzhen Science and Technology Major Project (Grant No: KJZD20230923114615032), the major project of science and technology research program of Chongqing Education Commission (Grant No: KJZD-M202302001), the educational reform research project of Shenzhen Technology University (Grant No: 20222011) the Special Funds for the Cultivation of Guangdong College Students’ Scientific and Technological Innovation (Grant No: pdjh2023b0469), and the Shenzhen University of Technology Self-made Experimental Instruments and Equipment Project (Grant No: JSZZ202301006).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Alhagry, S., Fahmy, A. A., and El-Khoribi, R. A. (2017). Emotion recognition based on EEG using LSTM recurrent neural network. Int. J. Adv. Comput. Sci. Appl. 8, 355–358. doi: 10.14569/IJACSA.2017.081046
Atkinson, J., and Campos, D. (2016). Improving BCI-based emotion recognition by combining EEG feature selection and kernel classifiers. Expert Syst. Appl. 47, 35–41. doi: 10.1016/j.eswa.2015.10.049
Avena-Koenigsberger, A., Misic, B., and Sporns, O. (2017). Communication dynamics in complex brain networks. Nat. Rev. Neurosci. 19, 17–33. doi: 10.1038/nrn.2017.149
Bassett, D. S., Greenfield, D. L., Meyer-Lindenberg, A., Weinberger, D. R., Moore, S. W., and Bullmore, E. T. (2010). Efficient physical embedding of topologically complex information processing networks in brains and computer circuits. PLoS Comput. Biol. 6:e1000748. doi: 10.1371/journal.pcbi.1000748
Bassett, D. S., Wymbs, N. F., Porter, M. A., Mucha, P. J., Carlson, J. M., and Grafton, S. T. (2011). Dynamic reconfiguration of human brain networks during learning. Proc. Natl. Acad. Sci. 108, 7641–7646. doi: 10.1073/pnas.1018985108
Bertolero, M. A., Thomas Yeo, B. T., Bassett, D. S., and D’Esposito, M. (2018). A mechanistic model of connector hubs, modularity and cognition. Nat. Hum. Behav. 2, 765–777. doi: 10.1038/s41562-018-0420-6
Betzel, R. F., Medagalia, J. D., Papadopoulos, L., Baum, G. L., Gur, R., Gur, R., et al. (2017). The modular organization of human anatomical brain networks: accounting for the cost of wiring. Netw. Neurosci. 1, 42–68. doi: 10.1162/NETN_a_00002
Bozhkov, L., Koprinkova-Hristova, P., and Georgieva, P. (2016). Learning to decode human emotions with Echo state networks. Neural Netw. 78, 112–119. doi: 10.1016/j.neunet.2015.07.005
Bozhkov, L., Koprinkova-Hristova, P., and Georgieva, P. (2017). Reservoir computing for emotion valence discrimination from EEG signals. Neurocomputing 231, 28–40. doi: 10.1016/j.neucom.2016.03.108
Bradley, M. M., and Lang, P. J. (1998). Measuring emotion: the self-assessment manikin and the semantic differential. J. Behav. Ther. Exp. Psychiatry 25, 49–59. doi: 10.1016/0005-7916(94)90063-9
Bullmore, E., and Sporns, O. (2009). Complex brain networks: graph theoretical analysis of structural and functional systems. Nat. Rev. Neurosci. 10, 186–198. doi: 10.1038/nrn2575
Bullmore, E., and Sporns, O. (2012). The economy of brain network organization. Nat. Rev. Neurosci. 13, 336–349. doi: 10.1038/nrn3214
Cai, H., Qu, Z., Li, Z., Zhang, Y., Hu, X., and Hu, B. (2020). Feature-level fusion approaches based on multimodal EEG data for depression recognition. Inform. Fusion 59, 127–138. doi: 10.1016/j.inffus.2020.01.008
Castellani, G. C., Intrator, N., Shouval, H. Z., and Cooper, L. N. (1999). Solutions of the BCM learning rule in a network of lateral interacting nonlinear neurons. Netw. Comput. Neural Syst. 10, 111–121. doi: 10.1088/0954-898X_10_2_001
Chen, J. X., Zhang, P. W., Mao, Z. J., Huang, Y. F., Jiang, D. M., and Zhang, Y. N. (2019). Accurate EEG-based emotion recognition on combined features using deep convolutional neural networks. IEEE Spect. Sect. New Trends Brain Sign. Process. Analys. 7, 44317–44328. doi: 10.1109/ACCESS.2019.2908285
Cimtay, Y., and Ekmekcioglu, E. (2020). Investigating the use of pretrained convolutional neural network on cross-subject and cross-dataset EEG emotion recognition. Sensors 20:2034. doi: 10.3390/s20072034
Clune, J., Mouret, J. B., and Lipson, H. (2013). The evolutionary origins of modularity. Proc. R. Soc. B 280:20122863. doi: 10.1098/rspb.2012.2863
Crossley, N. A., Mechelli, A., Scott, J., Carletti, F., Fox, P. T., McGuire, P., et al. (2014). The hubs of the human connectome are generally implicated in the anatomy of brain disorders. Brain 137, 2382–2395.
Deng, Z., and Zhang, Y. (2007). Collective behavior of a small-world recurrent neural system with scale-free distribution. IEEE Trans. Neural Netw. 18, 1364–1375.
Depasquale, B., Cueva, C. J., Rajan, K., Escola, G. S., and Abbott, L. F. (2018). Full-FORCE: a target-based method for training recurrent networks. PLoS One 13:e0191527. doi: 10.1371/journal.pone.0191527
Fourati, R., Ammar, B., Aouiti, C., Sanchez-Medina, J., and Alimi, A. M. (2017). “Optimized Echo state network with intrinsic plasticity for EEG-based emotion recognition” in International Conference on Neural Information Processing. Springer, 718–727.
Fourati, R., Ammar, B., Jin, Y., and Alimi, A.M. (2020a). “EEG feature learning with intrinsic plasticity based deep Echo state network” in International Joint Conference on Neural Networks. IEEE.
Fourati, R., Ammar, B., Sanchez-Medina, J., and Alimi, A. M. (2020b). Unsupervised learning in reservoir computing for EEG-based emotion recognition. IEEE Trans. Affect. Comput. 13, 972–984. doi: 10.1109/TAFFC.2020.2982143
Jaeger, H. (2001a). The “echo state” approach to analysing and training recurrent neural networks-with an erratum note. German National Research Center for Information Technology GMD Technical Report 148.
Ju, H., Xu, J. X., Chong, E., and VanDongen, A. M. (2013). Effects of synaptic connectivity on liquid state machine performance. Neural Netw. 38, 39–51.
Kasabov, N. K. (2014). NeuCube: a spiking neural network architecture for mapping, learning and understanding of spatio-temporal brain data. Neural Netw. 52, 62–76. doi: 10.1016/j.neunet.2014.01.006
Kashtan, N., and Alon, U. (2005). Spontaneous evolution of modularity and network motifs. Proc. Natl. Acad. Sci. 102, 13773–13778. doi: 10.1073/pnas.0503610102
Kim, J. S., and Kaiser, M. (2014). From Caenorhabditis elegans to the human connectome: a specific modular organization increases metabolic, functional and developmental efficiency. Philos. Trans. R. Soc. B 369:20130529. doi: 10.1098/rstb.2013.0529
Koelstra, S., Muhl, C., Soleymani, M., Lee, J. S., Yazdani, A., Ebrahimi, T., et al. (2011). DEAP: a database for emotion analysis using physiological signals. IEEE Trans. Affect. Comput. 3, 18–31. doi: 10.1109/T-AFFC.2011.15
Koprinkova-Hristova, P., Bozhkov, L., and Georgieva, P. (2015). Echo State Networks for Feature Selection in Affective Computing. Lecture Notes in Computer Science 9086:131–141. doi: 10.1007/978-3-319-18944-4_11
Krienen, F. M., Yeo, B. T. T., Ge, T., and Sherwood, C. C. (2015). Transcriptional profiles of supragranular-enriched genes associate with corticocortical network architecture in the human brain. Proc. Natl. Acad. Sci. 113, E469–E478. doi: 10.1073/pnas.1510903113
Li, X., Song, D., Zhang, P., Yu, G., Hou, Y., and Hu, B. (2016). “Emotion recognition from multi-channel EEG data through convolutional recurrent neural network” in IEEE International Conference on Bioinformatics and Biomedicine. IEEE.
Li, X., Zhang, Y., Tiwari, P., Song, D., Hu, B., Yang, M., et al. (2022). EEG based emotion recognition: a tutorial and review. ACM Comput. Surv. 55, 1–57. doi: 10.1145/3524499
Lin, W., Li, C., and Sun, S. (2017). “Deep convolutional neural network for emotion recognition using EEG and peripheral physiological signal” in International Conference on Image and Graphics. Springer.
Litwin-Kumar, A., Harris, K. D., Axel, R., Sompolinsky, H., and Abbott, L. F. (2017). Optimal degrees of synaptic connectivity. Neuron 93, 1153–1164.e7. doi: 10.1016/j.neuron.2017.01.030
Liu, Y., and Sourina, O. (2013). Real-time fractal-based valence level recognition from EEG. Trans. Comput. Sci. XVIII 7848, 101–120. doi: 10.1007/978-3-642-38803-3_6
Lukosevicius, M., and Jaeger, H. (2009). Reservoir computing approaches to recurrent neural network training. Comput. Sci. Rev. 3, 127–149. doi: 10.1016/j.cosrev.2009.03.005
Mauss, I. B., and Robinson, M. D. (2009). Measures of emotion: a review. Cognit. Emot. 23, 209–237. doi: 10.1080/02699930802204677
Meunier, D., Lambiotte, R., and Bullmore, E. T. (2010). Modular and hierarchically modular organization of brain networks. Front. Neurosci. 4:200. doi: 10.3389/fnins.2010.00200
Nematzadeh, A., Ferrara, E., Flammini, A., and Ahn, Y. (2014). Optimal network modularity for information diffusion. Phys. Rev. Lett. 113:088701. doi: 10.1103/physrevlett.113.088701
Oja, E. (1982). Simplified neuron model as a principal component analyzer. J. Math. Biol. 15, 267–273. doi: 10.1007/BF00275687
Patlar Akbulut, F. (2022). Hybrid deep convolutional model-based emotion recognition using multiple physiological signals. Comput. Methods Biomech. Biomed. Eng. 25, 1678–1690. doi: 10.1080/10255842.2022.2032682
Rodriguez, N., Izquierdo, E., and Ahn, Y. (2021). Optimal modularity and memory capacity of neural reservoirs. Netw. Neurosci. 3, 551–566. doi: 10.1162/netn_a_00082
Samavat, A., Khalili, E., Ayati, B., and Ayati, M. (2022). Deep learning model with adaptive regularization for EEG-based emotion recognition using temporal and frequency features. IEEE Access 10, 24520–24527. doi: 10.1109/ACCESS.2022.3155647
Schrauwen, B., Wardermann, M., Verstraeten, D., Steil, J. J., and Stroobandt, D. (2008). Improving reservoirs using intrinsic plasticity. Neurocomputing 71, 1159–1171. doi: 10.1016/j.neucom.2007.12.020
Shen, F., Dai, G., Lin, G., Zhang, J., Kong, W., and Zeng, H. (2020). EEG-based emotion recognition using 4D convolutional recurrent neural network. Cogn. Neurodyn. 14, 815–828. doi: 10.1007/s11571-020-09634-1
Sporns, O., and Betzel, R. F. (2016). Modular brain networks. Annu. Rev. Psychol. 67, 613–640. doi: 10.1146/annurev-psych-122414-033634
Sporns, O., Tononi, G., and Kotter, R. (2005). The human connectome: a structural description of the human brain. PLoS Comput. Biol. 1:e42. doi: 10.1371/journal.pcbi.0010042
Subasi, A. (2007). EEG signal classification using wavelet feature extraction and a mixture of expert model. Expert Syst. Appl. 32, 1084–1093. doi: 10.1016/j.eswa.2006.02.005
Taylor, P. N., Wang, Y., and Kaiser, M. (2017). Within brain area tractography suggests local modularity using high resolution connectomics. Sci. Rep. 7:39859. doi: 10.1038/srep39859
Tripathy, S. J., Padmanabhan, K., Gerkin, R. C., and Urban, N. N. (2013). Intermediate intrinsic diversity enhances neural population coding. Proc. Natl. Acad. Sci. 110, 8248–8253. doi: 10.1073/pnas.1221214110
Volo, M. D., and Destexhe, A. (2021). Optimal responsiveness and information flow in networks of heterogeneous neurons. Sci. Rep. 11:17611. doi: 10.1038/s41598-021-96745-2
Zamani, F., and Wulansari, R. (2021). Emotion classification using 1D-CNN and RNN based on deap dataset. Nat. Lang. Process, 363–378. doi: 10.5121/csit.2021.112328
Zeldenrust, F., Gutkin, B., and Deneve, S. (2021). Efficient and robust coding in heterogeneous recurrent networks. PLoS Comput. Biol. 17:e1008673. doi: 10.1371/journal.pcbi.1008673
Zhang, Y., Chen, J., Tan, J. H., Chen, Y., Chen, Y., Li, D., et al. (2020). An investigation of deep learning models for EEG-based emotion recognition. Front. Neurosci. 14:622759. doi: 10.3389/fnins.2020.622759
Zhang, B., Miller, D. J., and Wang, Y. (2011). Nonlinear system modeling with random matrices: Echo state networks revisited. IEEE Trans. Neural Netw. Learn. Syst. 23, 175–182. doi: 10.1109/TNNLS.2011.2178562
Zhang, T., Zheng, W., Cui, Z., Zong, Y., and Li, Y. (2018). Spatial-temporal recurrent neural network for emotion recognition. IEEE Trans. Cybernet. 49, 839–847. doi: 10.1109/TCYB.2017.2788081
Zheng, W.-L., and Lu, B.-L. (2015). Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural network. IEEE Trans. Autonom. Mental Dev. 7, 162–175. doi: 10.1109/TAMD.2015.2431497
Zheng, W.-L., Zhu, J.-Y., and Lu, B.-L. (2017). Identifying stable patterns over time for emotion recognition from EEG. IEEE Trans. Affect. Comput. 10, 417–429. doi: 10.1109/TAFFC.2017.2712143
Keywords: modular echo state network, emotion recognition, EEG, memory capacity, heterogeneity
Citation: Yang L, Wang Z, Wang G, Liang L, Liu M and Wang J (2024) Brain-inspired modular echo state network for EEG-based emotion recognition. Front. Neurosci. 18:1305284. doi: 10.3389/fnins.2024.1305284
Edited by:
Benjamin Thompson, University of Waterloo, CanadaReviewed by:
Di Zhou, Kanazawa University, JapanMan Fai Leung, Anglia Ruskin University, United Kingdom
Copyright © 2024 Yang, Wang, Wang, Liang, Liu and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Junsong Wang, d2pzb25nMjAwNEAxMjYuY29t
†Completed this research while he visited Shenzhen Technology University during the summer 2022