- 1Signal Processing and Recognition Group, Department of Electrical and Electronic Engineering, Universidad Nacional de Colombia, Manizales, Colombia
- 2Diseño Electrónico y Técnicas de Tratamiento de Señal, Universidad Politecnica de Cartagena, Cartagena, Spain
Lately, research on computational models of emotion had been getting much attention due to their potential for understanding the mechanisms of emotions and their promising broad range of applications that potentially bridge the gap between human and machine interactions. We propose a new method for emotion classification that relies on features extracted from those active brain areas that are most likely related to emotions. To this end, we carry out the selection of spatially compact regions of interest that are computed using the brain neural activity reconstructed from Electroencephalography data. Throughout this study, we consider three representative feature extraction methods widely applied to emotion detection tasks, including Power spectral density, Wavelet, and Hjorth parameters. Further feature selection is carried out using principal component analysis. For validation purpose, these features are used to feed a support vector machine classifier that is trained under the leave-one-out cross-validation strategy. Obtained results on real affective data show that incorporation of the proposed training method in combination with the enhanced spatial resolution provided by the source estimation allows improving the performed accuracy of discrimination in most of the considered emotions, namely: dominance, valence, and liking.
1. Introduction
The design of reliable systems and devices, which can recognize, interpret, and process human emotions, still poses a challenging task. As long as the latent factors that generate emotions are unobservable, the main problem is how to rely on visible manifestations of emotions to reproduce and verify them. There are several objective measures, proposed for estimating unseen psychological activity, that are often related to expressive movements (facial expressions, gestures, movements of limbs, vocal emission, etc.), where facial expression analysis is one of the most frequently addressed themes (Marrero-Fernandez et al., 2014). Although these models can be more manageable to implement, human beings considerably vary in the manner which they manifest motivational states through movement patterns (Wentzel and Wigfield, 2009). Moreover, humans convey emotional information both intentionally and non-intentionally, producing consciously regulating or naturally suppressing emotional expressions. Therefore, derived motivational movement patterns may be too short for the goal of emotion recognition (Virvou et al., 2015).
As an option to expressive movements is the use of physiological components generated by the autonomic nervous system, such as electrocardiogram (Agrafioti et al., 2012), skin conductance (Muthu Meena et al., 2015), respiration and blood pressure (Dan-Glauser and Gross, 2015) to improve the accuracy and robustness of emotion recognition systems. In comparison with methods based on expressive movements, the responses of peripheral physiological signals tend to provide more detailed and complex information as an indicator for estimating emotional states (Wang et al., 2014). However, neuroscience has provided strong evidence that several cortical and subcortical regions (e.g., insula, prefrontal regions, thalamus, amygdala, hippocampus, basal ganglia) are implicated in emotional perception and regulation (Kober et al., 2008). Consequently, a growing number of studies have been enabled recently to engage and thus measure emotional functions using neuroimaging paradigms mostly through brain signals captured from the central nervous system such as electroencephalogram (EEG) and functional magnetic resonance imaging (fMRI). Nonetheless, the first technique is more desired in emotion recognition systems due to its non-invasiveness, cost effectiveness, and simple acquisition (Robinson and Robinson, 2009). Another aspect of consideration is the brain activity nature of emotional functions, which are partially localized in both space (some cortical and subcortical regions) and frequency (mostly upon the neural oscillation bands). Although EEG suffers from its poor spatial resolution and raised susceptibility to noise, this biosignal is a plausible approach in emotion recognition due to its provided high temporal resolution to investigate the brain dynamics (Liu et al., 2014).
Most of the research studies devoted to feature extraction put emphasis on the spectral power of theta, alpha, beta, and gamma bands (Jirayucharoensak et al., 2014). Due to the nonstationary behavior of EEG signals, the features, which aim at discrimination of time-series, are also extracted, including time-varying (Alvarez-Meza et al., 2015), time-frequency (Sirca et al., 2015), and nonlinear analysis methods (Liu and Sourina, 2013; Ahammed, 2015). Nevertheless, some issues remain open in characterizing EEG data for emotional recognition. Moreover, interpretation of extracted features from EEG sensor level is not straightforward due to a couple of basic reasons: the low spatial resolution provided by EEG recordings and the influence of field spread that severely corrupt data, i.e., neural activity travels from cortical to scalp surface through several head tissues (Schoffelen and Gross, 2009). To overcome this concern, some methods are focused on characterizing the connectivity networks of brain sources reconstructed from scalp EEG signals (Brookes et al., 2014), where usually, a realistic model of the brain cortex serves as the source space. However, some methodological aspects remain still open. Particularly, mapping solutions with a simplistic spatial structure are assumed to be favored regarding neurophysiological plausibility. Nonetheless, the simplest approaches are the minimum norm estimate that minimizes the overall power of the sources and Low-Resolution Tomography (LORETA) that explicitly enforces spatial smoothness of the sources, making either approach to producing poor spatial accuracy. Furthermore, it has also been argued that only a small fraction of the brain should be consistently activated in event-related experimental designs (Castaño-Candamil et al., 2015).
Consequently, the used mapping algorithm must provide an accurate location of those brain areas that are most related to the studied task (or Regions-of-interest – ROIs). Therefore, the ROIs identification can be performed relying on some prior knowledge about their association with the underlying research task. In other words, the evidence provided previously by other functional imaging studies is accepted as discussed in Schoffelen and Gross (2009). Additionally, some ROI selection methods have been proposed mainly based on anatomical constraints (Nordhjem et al., 2016), fostering brain areas with similar temporal responses (Edelman et al., 2016), or as part of a feature extraction procedure under a supervised classification framework, i.e., regarding the classification accuracy as proposed in Illán et al. (2011). Nevertheless, the widely allowed strategy is the determination of those reconstructed brain areas with a potent activity during the experimental task (Robinson and Robinson, 2009), even though determination of the best strategy of ROI selection for each task under consideration remains an open issue. Finally, once the ROIs are suitably selected, the feature extraction procedure is carried out over their time series, improving the low spatial resolution of raw EEG data, and avoiding the field spread effect (Toppi et al., 2016).
Although many features may be extracted from both the raw EEG data or the ROI time courses, their main restriction is how to extract discriminative features as much as possible since several features may not contain relevant information introducing redundancy (Verleysen and François, 2005). Consequently, extracted feature set may decrease classification accuracy. Therefore, an emerging challenge for emotion discrimination is how to tackle the problem of large training feature spaces. To this end, relevance analysis algorithms are used, by instance, the one developed in Alvarez-Meza et al. (2015) that is built on the eigenvalues and eigenvectors of the feature covariance matrix. In this case, the relevance measure is assumed to point out to the best input attributes that exhibit higher overall correlations to the estimated principal components. For validation experiments of effective emotion discrimination, SVMs are a powerful machine learning approach that has been successfully applied to solving problems with complex multi-class features, overcoming some of the drawbacks relating to over-fitting (López et al., 2011).
Under the aforementioned considerations, we present a new method for discrimination of emotions that is founded upon the extraction of relevant ROI time series from EEG data, yielding an improved accurate location of the brain areas related to emotion states. In order to map accurately the scalp recordings to the source space, we solve the ill-posed EEG inverse problem by employing the Multiple Sparse Priors Algorithm (MSP) as to encourage sparse and focal solutions (Friston et al., 2008). Further, the data-driven selection of the regions of interest is accomplished along with their respective time courses estimation, holding two main aspects: (i) An adequate selection of the brain areas should allow characterizing the brain activity directly related to the specifically studied task, and (ii) Quality of the estimated time courses biases the goodness of estimated features. Obtained results on real affective data show that incorporation of the proposed training method in combination with enhanced spatial resolution allows improving the performed accuracy of discrimination in most of the considered emotions, namely: Arousal, dominance, valence, and liking.
The rest of the paper is organized as follows: Section 2 describes materials and methods, the experimental setup is explained in detail in Section 3, this is followed by the discussion and concluding remarks in Section 4.
2. Materials and Methods
2.1. Brain Source Activity with Spatially Localized Solutions
In order to assess the needed connectivity analysis between distant brain regions, all measured EEG signals must be evaluated upon the source space, i.e., the scalp neural field recordings should be projected into brain volume that we carry out within the inverse problem framework. To this end, the measured EEG data are expressed through a multivariate linear model that incorporates a distributed source representation with fixed positions and orientations as below:
where X ∈ ℝC×T is the EEG data comprising C channels and T time samples, J ∈ ℝD×T is the amplitude of D current dipoles distributed through the cortical surface with fixed orientation perpendicular to it, and M ∈ ℝC×D (termed lead field matrix) is the gain matrix that holds all available relationships between sources and EEG data influenced by a zero mean Gaussian noise ϵ ∈ ℝC×T with covariance .
Since we assume J to be a zero mean Gaussian process that has prior covariance cov{J} = QJ, with , then, the brain activity estimation, is carried out by solving the following optimization problem (Grech et al., 2008):
where λ∈ℝ+ is the regularization parameter and notation stands for the Mahalanobis distance. Consequently, the first term relates to the quadratic error function while the second one introduces additional information for solving a given ill-posed task, usually, in a form of constraints that are imposed upon the source activity.
The inverse-problem optimization above-stated yields the following estimate:
This estimation demands some prior knowledge about the modeled sensor noise covariance Qϵ and source covariance QJ. To supply this information about the former matrix, we set Qϵ = exp(λϵ)IC, where is the appropriately sized identity matrix that is scaled by a hyper-parameter, modulating the sensor noise variance (Phillips et al., 2002). In turn, the prior covariance matrix can be built as a sum over an assumed set that holds P spatial areas {Qp} so that each one relates to a single potentially activated cortex area, whose contribution to the mapped neural activity is expressed by the respective adding hyper-parameter, (termed weight), as follows:
The above prior covariance matrix gives rise to the possibility for implementing sparse-based approaches, in particular, the Multiple Sparse Priors (MSP).
In practice, the joint optimization of real-valued hyper-parameter sets {λϵ, λp} can be accomplished by employing any standard variational scheme such as Expectation Maximization along with the greedy search algorithm (Belardinelli et al., 2012). Besides, the covariance components are also constructed so that each one regards a single locally smooth, focal patch of the cortex. That is, , where is the set of non-overlapping patches that must cover the entire cortical surface. As suggested in Friston et al. (2008), the brain patch set are the columns of a Green's function matrix QG = exp(σGM), with , where is the graph Laplacian that comprises inter-dipole connectivity information about all neighboring dipoles. Here, the constant σ∈ℝ+ rules the smoothness of current distribution or spatial extent of the activated areas. As a rule, the graph Laplacian GM is calculated using an available head model through the adjacency matrix that incorporates all needed cortical information.
Moreover, a simpler approach, termed LORETA, can be directly formed by using the Green's function as the source prior covariance matrix, that it, by adjusting (Harrison et al., 2007)
As a result, considering as the time course of the d−th dipole, where d = 1, …, D, the brain activity energy ē∈ℝD×1 is reconstructed as follows:
where notations ◦ and 𝔼{·} stand for the Hadamard product and expectation operator, respectively.
2.2. Selection of Spatially Compact Regions of Interest
After computing the energy of brain activity, we focus on developing a data-driven approach to select the spatial ROI location accurately as to obtain the label vector (with ι ∈ ℕD×1) that better encodes the membership of each dipole to the set holding NROI ROIs, being ιd the assigned label to the d–th dipole. For this purpose, we rely on the focal, powerful brain activity as an effective indicator of each region of interest, which we assume solely compact and without any space partition. Therefore, we take advantage of the already calculated brain activity reconstruction power and the neighboring information available in the Green's function QG to the select each ROI set element.
Concretely, all estimated brain activity energy is thoroughly scanned over the brain surface, i.e., {ēd : ∀d ∈ d}, with the aim to detect the set of most powerful dipoles and their respective neighborhoods to avoid biasing the ROI estimates. This procedure is as follows: Firstly, we remove all power peaks that are below 10% of the amplitude activity with the maximum power. Consequently, those dipoles will not be labeled as members of any ROI. Then, we assume the first ROI as the neighboring spatial region that is enclosed within a prior fixed radius, ρ∈ℝ+, and centered at the dipole location that measures the most powerful neural activity. Likewise, the second ROI is further selected that we center at the dipole with next highest power, leaving out those dipoles that have been already labeled. This selection procedure is repeated until the whole set of most powerful dipoles is labeled, assuring that each one belongs to a single ROI. Note that the amount NROI is fixed automatically.
At this step, one label has been assigned to each dipole exceeding the prefixed threshold. Nevertheless, the ROI time courses estimation need to be addressed, as their quality determines the ROI-based feature selection and hence the extraction benefits. To this end, we average the estimated brain activity over the time courses, jd, that belong to the same ROI. So, the average over time courses is computed as:
where δ() is the delta function. Thus, the matrix is built where each r−th row holds the corresponding ROI time series yr, ∀r∈NROI.
2.3. ROI-Based Feature Extraction
Throughout this study, we consider three representative feature extraction methods that have been widely applied to emotion detection tasks (Alvarez-Meza et al., 2015):
2.3.1. Power Spectral Density Parameters
For the available ROI time series y, we estimate the Power Spectral Density (PSD) noted as where NB∈ℕ is the number of frequency bins that is fixed according to the spectral band of interest, where the most discriminative information for MI is concentrated. Provided the EEG sample frequency , the PSD vector s = {sf:f = 1, …, NB}, with sf∈ℝ and NB = ⌊Fs ∕ 2⌋, is estimated by means of the nonparametric Welch's method that calculates the widely known Fast Fourier Transform algorithm of a set of M∈ℕ overlapping segments, which are split from the preprocessed EEG data vector. Due to the non-stationary nature of EEG data, the piecewise stationary analysis is carried out over the set of extracted overlapping segments that are further windowed by a smooth-time weighting window α∈ℝL that lasts L∈ℕ (L<T), yielding a set of the windowed segments {vm∈ℝL:m = 1, …, M}, where (i = 1, …, L) is the i-th element of vm. Consequently, the following modified periodogram vector , , is computed based on the Discrete Fourier Transform as follows:
Therefore, we extract each PSD element as sf = uf ∕ Mν, where the value .
2.3.2. Wavelet-Based Parameters
Continuous Wavelet Transform (CWT) is a inner-product-based transformation quantifies similarity between a given equally sampled time series at time spacing δt∈ℝ and a previously fixed base function γ(η), termed mother wavelet ruled by a dimensionless parameter vector η∈ℝ. Namely, each time element of the CWT vector ςg∈ℂT is extracted from the preprocessed EEG time-series z at scale g∈ℝ by accomplishing their convolution with the scaled and shifted mother wavelet in the form:
where (*) denotes the complex conjugate.
To effectively address the trade-off between time and frequency resolution in the non-stationary signal analysis, Discrete Wavelet Transform (DWT) had been developed that provides multi-resolution and non-redundant representation by decomposing the considered time-series into some sub-bands at different scales, yielding more precise time-frequency information about z. Aiming to extract all suitable time-frequency information from DWT, the following detail vector bj∈ℂ at level j is defined
where aj, k = ∑t∈Tzthj, k(t), with aj, k∈ℂ, hj, k(t)∈ℂ is the impulse response of a given wavelet filter. Then, provided a wavelet ψ(·), the DWT-based decomposition of z is computed as zt = ∑j, k∈ℤaj, k ψ j, k(t).
2.3.3. Hjorth Parameters
For each windowed segment vm, a set of the following three vector parameters describe EEG signals on the time domain:
3. Experimental Set-Up
Figure 1 shows the main scheme of the proposed emotion discrimination using EEG source features. Owing to highlight the latent patterns in Y, we propose a time-series discrimination methodology appraising the following stages: (i) EEG source estimation, (ii) Selection of Regions of interest (ROI), and (iii) feature extraction and classification.
Figure 1. Scheme of the proposed methodology for discrimination of emotions using extracted features from EEG source reconstruction. The blocks marked in dashed lines are the subject of present study.
3.1. Testing EEG Database and Preprocessing
The pre-processed EEG and subjective data used in the present study were obtained from the publicly available database for emotion analysis using physiological signals (DEAP) (Koelstra et al., 2012). Thirty-two healthy participants (50% females and 50% males in average aging 26.9 years) were recruited and consented to participate in the study. Thirty-two-channel EEG data were recorded using a Biosemi Active Two system. Data were acquired at a sampling rate 512 Hz, placing the electrodes on surface scalp according to the International 10-20 system. Pre-processing included the following steps: common referencing, down sampling to 128 Hz, high-pass filtering from 4 Hz, and eye blink artifact removal via independent component analysis.
All participants were presented with forty, one-minute long music videos with varying emotional content. Before every video, there was a baseline period of five seconds so that each participant was asked to fixate at across in the middle of screen. Following the presentation of each video, the participants were provided enough time to rate the music videos on a discrete 9-point scale for valence, arousal, dominance, and liking. Valence, arousal, and dominance dimensions were scored using the self-assessment manikins (SAM) to gouge user emotional states (Siegert et al., 2011). For liking (i.e., how much did you like the video?), thumbs up and thumbs down icons were used. Familiarity was rated after the end of the experiment on a 5-point integer scale (from “never heard it before” to “listen to it regularly”).
3.2. Brain Activity Mapping
As regards the inverse solution used for validation purpose, two approaches are performed: the baseline LORETA (see Equation 4) and MSP (see Equation 3) that are implemented using the SPM software. For the MSP approach, the employed spatial dictionary comprises 512 basis as to cover the whole cortical surface. Also, we fix the value for spatial coherence prior as σ = 0.6, which propagates spatial dependences over three or four mesh vertices that are, on average, about 6 mm apart to get a trade-off between spatial accuracy and local coherence.
Figure 2 shows an example of the reconstructed brain activity from EEG data using either tested approach of mapping for the subject N 27. For the sake of a better visual perception, the top row displays three examples of topographic figures that are obtained for the trials labeled as #38, #39, and #40. Aiming to get the physiological interpretation of estimated sources, the middle topographic also includes the most commonly referred brain activity areas related to emotions. Next, Figures 2A,D display the corresponding reconstructed neural activity using LORETA and MPS, respectively. The former mapping technique yields more blurred estimation producing wide zones (sometimes excessively) of cortical activation. In contrast, the MSP mapping results in more compact regions of estimated activation.
Figure 2. Examples of estimated neural brain activity for N 27, trials labeled as #38, #39, and #40. The activated regions are highlighted. (A) LORETA, (B) LORETA ROI set, (C) α band calculated by LORETA-based ROI set, (D) MSP, (E) MSP ROI set, (F) α band calculated by MSP-based ROI set.
Upon the reconstructed neural activity, the ROI set is computed as shown in Figures 2B,E that displays the activated cortical locations over the topographic plots. So, both mappings discover some powerful common sources like somatosensory cortex, primary motor cortex, superior temporal gyrus, temporal lobe in the concrete case. However, the disparity between reconstructed neural activity brings two sets containing different activated areas. Thus, MSP finds considerable activity in the occipital lobe and orbitofrontal cortex, while LORETA detects in the association cortex. Therefore, the use of either mapping approach not necessarily leads to a unique ROI set.
Overall, the selected ROI set allows determining the location of those electrodes influencing the most, from which the features are to be extracted.
3.3. Feature Extraction from Selected ROI Sets
For the sake of analysis, the feature set is extracted from three scenarios of input signal, namely: from the measured EEG channels, X, the mapped activity J, and the estimated ROI set Y. So, the following parameters are fixed for each studied feature extraction method:
(i) PSD-based and feature subsets: For each band of interest (α, β, γ, and θ), the segment length value L, needed during calculation of the PSD and Hjort parameters, is adjusted as L>Fr ∕ Fs, where Fr = 4 Hz is the smallest considered frequency. Accordingly, the calculated PSD features are the two first statistical moments of s computed for all bands.
(ii) WT-based feature subset: A suitable wavelet function must be used to optimize the classifier performance. We select the Morlet wavelet for the CWT analysis because its wave shape and EEG signals are alike and it allows extracting features better localized in the frequency domain. Thus, we extract the short-time instantaneous CWT amplitudes using a set of the Morlet wavelets centered as follows: 2 Hz (to extract the δ band), 10 Hz (α band), 20 Hz (β band) and 50 Hz (γ band).
Figures 2C,F shows the trajectories of α band computed for either case of ROI set (subject N 27). Time series highlighted in red is the average as one of the features extracted. Table 1 shows the features extracted from EEG signals.
3.4. Classifier Training and Validation
In this stage, the statistical measures of the short-time parameters are computed, as described in Section 2, to extract the input feature space matrix Ξ = {ξn:n = 1, …, Ntr} with . Hence, the row vector holds NF = C×F concatenated features for the n-th trial, being NF∈ℕ the number of provided features of the tr-th trial for a given channel. Here, C = 32, F = 45, and Ntr = 40.
Before accomplishing the classification stage, a stochastic relevance analysis of the extracted feature set Ξ is performed by means of its eigenvalues and eigenvectors. Consequently, the input features are ranked following the yielded relevance vector, as described in detail in Alvarez-Meza et al. (2015). The main assumption behind this relevance analysis is that the largest values of the ranking vector should point out to the better input attributes since they exhibit higher overall correlations to the estimated principal components. The variance explained is adjusted to 90% for mapping the most relevant components.
Additionally, the set of relevant features feeds a soft-margin support vector machine (SVM) based classifier that is trained in the following leave-one-out cross-validation methodology: (i) D-1 samples in the database are used to learn the most relevant features and to train the classifier, while the remaining sample is used for testing, (ii) repeat the preceding stage until all samples have been used as testing sample. This procedure is implemented to solve the four different binary classification problems: low/high arousal, low/high dominance, low/high valence and low/high liking. Note that the reason for choosing a leave-one-sample-out cross-validation scheme is that the number of samples per subject is not enough for generating significant training and testing sets, making this strategy the most used in emotion discrimination for the DEAP database.
With the aim to configure the low and high classes, the subjective scores are thresholded at the mid-point of a 9-pointscale, i.e., at 5. Since this procedure results in unbalanced classes for each subjective rating, the F1-score, along with classification accuracies, is employed to describe reliably the results of classifier performance while tackling the class imbalance, as suggested in Gupta et al. (2016).
Figure 3 shows the accuracy performed by the SVM classifier for each one of the considered emotions. In terms of feature extraction, there are three experiment setups of emotion discrimination: Features extracted directly from the EEG data, avoiding any mapping activity (noted as EEG), features from ROI sets computed for LORETA (LOR-ROI) and MSP (MSP-ROI) reconstruction techniques. For a better illustration, the cardinal of subjects is rearranged following the obtained accuracy by the proposed extraction of selected ROI time series (i.e., the MSP-ROI experiment). All subjects are ranked by accuracy score in decreasing order.
Figure 3. Performed accuracy by the considered experiment setups for all considered emotions. The cardinal of subjects is rearranged following the obtained accuracy by MSP-ROI, though the original database cardinal is shown. (A) Liking, (B) Dominance, (C) Valence, (D) Arousal.
Figure 3A shows the performed F1 values for liking discrimination that has been widely reported as the emotion having the most complex random structure. The use of features extracted from spatially compact ROI set results in a higher accuracy achieved by either tested approach of brain activity mapping. For most of the subjects, the MSP-ROI approach overcomes LOR-ROI and EEG settings. Although the same situation holds for the Dominance state, the advantage of MSP-ROI performance somewhat reduces (see Figure 3B). Further, in the case of Valence discrimination, MSP-ROI also performs the best, but the LOR-ROI behaves very close to the plain EEG (see Figure 3C). Lastly, neither ROI-based strategy seems to outperform the plain EEG for Arousal emotion (see Figure 3D).
To compare all considered approaches of neural analysis, we quantify the statistical difference regarding the F1 score and classifier accuracy ac, validating whether the MSP-ROI performance over subjects is higher than the one obtained by other approaches. To this end, a paired sample t-test is carried out in which the null hypothesis states that there are not significant differences between MSP-ROI and each compared approach in terms of the performed population mean values. Otherwise, the alternative hypothesis states that our population mean are greater that the other our population-means calculated over all subjects are greater that other values.
Table 2 summarizes the achieved performance by each examined training setup, regarding the maximum and average values for all subjects.Superscripted stars indicate whether the population mean of MSP-ROI is greater than the population mean of EEG (taking into account accuracy and F1-score separately ** = p < 0.01, * = p < 0.05). Subscripted diamonds indicate whether the population mean of MSP-ROI is greater than the population mean of LOR-ROI (taking into account accuracy and F1-score separately ◇◇ = p < 0.01, ◇ = p < 0.05). From the obtained results for either performance measure (F1 score and accuracy), it can be noted that the use of MSP mapping allows increasing the system accuracy so that it even outperforms LORETA in either case of setting regardless the discriminating emotion. Therefore, the incorporation of features, extracted from spatially compact Regions of Interest, enables to enhance further the performed accuracy of emotion discrimination.
4. Discussion and Concluding Remarks
We propose a method for supporting discrimination of emotions that employs features extracted from spatially compact regions of interest. To this end, we aim to compute the characterizing ROI set upon the reconstructed brain neural activity, using EEG recordings.
However, during training and validation of the proposed discrimination method, the following factors are to be considered:
– The first aspect to reflect is the influence of intrinsic uncertainty provided by the acquired EEG data. As seen in Figure 4A that shows the entropy as measure of uncertainty, the utmost disparity between subjects is as much as ten times. In fact, some of the outlier recordings (v.g. #27, #8, and #23) are quite noisy. However, there is another source of great incertitude that is related to the labeling procedure used during construction of databases for affect representation and recognition. Thus, Figure 4B displays the ranked values of entropy for the DEAP label set. As already discussed in Verma and Tiwary (2016), the large value of standard deviation for each emotion points out on their labeling difficulty.
– The following aspect concerns the technique that we apply for imaging EEG activity. The first tested method of neural mapping is LORETA that has the characteristic that localization is preserved with a certain amount of dispersion, i.e., it has a relatively low spatial resolution (Pascual-Marqui et al., 1994). Consequently, the reconstructed brain activity is more blurred, producing wide zones (sometimes excessively) of cortical activation. In contrast, the use of sparse-based approaches (like MSP) notably improves identification of the source signals from noisy electroencephalographic measurements (Becker et al., 2015). As a result, MSP allows performing brain source imaging so that we obtain more compact regions of activation. Furthermore, due to the low spatial resolution, Loreta algorithm frequently infers false regions of activation, where powerful common sources should not be present. Therefore, although LORETA has been widely used in the last years to localize electrical generators of scalp EEG data, its confidence of estimated areas of activation may be not enough.
– Generally speaking, a challenging issue relating to the emotion detection is how to identify ROI sets precisely at very short temporal scales; this dilemma remains common for all cognitive tasks (Hassan et al., 2015). Similar studies have selected the ROIS set by identifying particular gyral landmarks on subject specific cortex models (Mattia et al., 2006). However, we rather employ a data-driven approach that focuses on accurately encoding the membership of each dipole to the estimated areas with salient cortical activation (identified as the Region-of-Interest set) so that every single ROI is assumed solely compact and without any space partition. Thus, each selected ROI set allows considering the time series of neural activity, contributing the most to the emotion states. As a result, the introduced ROI sets enhance the performed detection accuracy.
– We carry out the binary classification as the only one reported in the literature for emotion recognition. However, the use of hard thresholding algorithms for binarizing a label set leads to losing most of the emotional richness. Moreover, the scores near the midpoints and extreme values may have different implications. Therefore, other strategies of labeling should be considered to better capture the richness of emotion dimensions like the use of regression.
– The next factor of value is the classifier setup. We use SVM classifier with radial basis kernel that is assumed more robust to class-imbalance of training emotion data (Daimi and Saha, 2014). For this purpose, we consider three representative feature extraction methods that have been widely applied to emotion detection tasks. However, due to the high dimension of the input feature space, feature selection is carried out trough PCA before feeding the classifier. As a result, we obtain a reduced set of features that represent the most the properties of input training space.
– Lastly, Table 3 summarizes the comparison of emotion identification systems that had been recently reported in the literature, regarding two commonly used performance measures: F1 and accuracy. As seen, although the proposed approach obtains competitive scores of accuracy for the same dataset used here, other methods are performing higher values employing similar features and classifiers. Nonetheless, the discussed MSP-ROI method reaches the biggest F1 values for all considered emotion classes. This matter deserves particular attention because of the imbalanced class of validated database, making the F1-score more reliable to quantify the classifier performance in this study (Gupta et al., 2016). Note that the obtained results agree with those studies in BCI areas like in Edelman et al. (2015), where the use of spatially compact regions of interest have also been applied, allowing to better identify the network activity of these tasks in source space.
Figure 4. Uncertainty estimated for the DEAP database. (A) Measued EEG data, (B) Labeling procedure.
Therefore, we conclude that incorporation of the proposed training method in combination with enhanced spatial resolution allows improving the performed accuracy of discrimination in most of the considered emotions, namely: dominance, valence, and liking.
As future work, the authors intend to refine the recognition method by investigating a finer and more complex class of emotions. Besides, the use of more elaborate methods for extraction of temporal information from the ROI set is to be strongly considered such as the connectivity analysis. Another, aspect of improving is the feature selection that should take into consideration the nonstationary behavior of EEG data.
Author Contributions
JP: His research interests are feature extraction/selection for training pattern recognition systems, bio-engineering, neuroscience and machine learning. JM: His research interests include machine learning and signal processing methods applied to image and video data analysis as well as bioengineering tasks. GC: His teaching and research interests include information and signal theory, digital signal processing and bioengineering.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The handling Editor declared a shared affiliation, though no other collaboration, with one of the authors JIP-B and states that the process nevertheless met the standards of a fair and objective review.
Acknowledgments
This research was supported by COLCIENCIAS project Evaluacin asistida de potenciales evocados cognitivos como marcador del transtorno por dficit de atencin e hiperactividad (TDAH), Programa Nacional de Formacion de Investigadores Generacion del Bicentenario (2011) and Programa nacional de becas de doctorado convocatoria 647 de 2014.
References
Agrafioti, F., Hatzinakos, D., and Anderson, A. (2012). Ecg pattern analysis for emotion detection. IEEE Trans. Affect. Comput. 3, 102–115. doi: 10.1109/T-AFFC.2011.28
Ahammed, K. (2015). Identification of human emotions via univariate and multivarite multiscale entropy. Am. J. Eng. Appl. Sci. 8, 410–416. doi: 10.3844/ajeassp.2015.410.416
Alvarez-Meza, A., Velasquez-Martinez, L., and Castellanos-Dominguez, G. (2015). Time-series discrimination using feature relevance analysis in motor imagery classification. Neurocomputing 151(Pt 1):122–129. doi: 10.1016/j.neucom.2014.07.077
Becker, H., Albera, L., Comon, P., Gribonval, R., Wendling, F., and Merlet, I. (2015). Brain-source imaging: from sparse to tensor models. IEEE Signal Process Mag. 32, 100–112. doi: 10.1109/MSP.2015.2413711
Belardinelli, P., Ortiz, E., Barnes, G., Noppeney, U., and Preissl, H. (2012). Source reconstruction accuracy of MEG and EEG bayesian inversion approaches. PLoS ONE 7:e51985. doi: 10.1371/journal.pone.0051985
Brookes, M. J., O'Neill, G. C., Hall, E. L., Woolrich, M. W., Baker, A., Corner, S. P., et al. (2014). Measuring temporal, spectral and spatial changes in electrophysiological brain network connectivity. NeuroImage 91, 282–299. doi: 10.1016/j.neuroimage.2013.12.066
Castaño-Candamil, S., Höhne, J., Martínez-Vargas, J.-D., An, X.-W., Castellanos-Domínguez, G., and Haufe, S. (2015). Solving the EEG inverse problem based on space-time-frequency structured sparsity constraints. NeuroImage 118, 598–612. doi: 10.1016/j.neuroimage.2015.05.052
Daimi, S. N., and Saha, G. (2014). Classification of emotions induced by music videos and correlation with participants' rating. Exp. Syst. Appl. 41, 6057–6065. doi: 10.1016/j.eswa.2014.03.050
Dan-Glauser, E. S., and Gross, J. J. (2015). The temporal dynamics of emotional acceptance: experience, expression, and physiology. Biol. Psychol. 108, 1–12. doi: 10.1016/j.biopsycho.2015.03.005
Edelman, B., Baxter, B., and He, B. (2015). “Decoding and mapping of right hand motor imagery tasks using EEG source imaging,” in 2015 7th International IEEE/EMBS Conference on Neural Engineering (NER) (Montpellier: IEEE), 194–197. doi: 10.1109/NER.2015.7146593
Edelman, B., Baxter, B., and He, B. (2016). EEG source imaging enhances the decoding of complex right-hand motor imagery tasks. IEEE Trans. Biomed. Eng. 63, 4–14. doi: 10.1109/TBME.2015.2467312
Friston, K., Harrison, L., Daunizeau, J., Kiebel, S., Phillips, C., Trujillo-Barreto, N., et al. (2008). Multiple sparse priors for the M/EEG inverse problem. NeuroImage 39, 1104–1120. doi: 10.1016/j.neuroimage.2007.09.048
Grech, R., Cassar, T., Muscat, J., Camilleri, K., Fabri, S., Zervakis, M., et al. (2008). Review on solving the inverse problem in EEG source analysis. J. NeuroEng. Rehabil. 5, 792–800. doi: 10.1186/1743-0003-5-25
Gupta, R., Laghari, K. R., and Falk, T. H. (2016). Relevance vector classifier decision fusion and EEG graph-theoretic features for automatic affective state characterization. Neurocomputing 174(Pt B):875–884. doi: 10.1016/j.neucom.2015.09.085
Harrison, L., Penny, W., Ashburner, J., Trujillo-Barreto, N., and Friston, K. (2007). Diffusion-based spatial priors for imaging. NeuroImage 38, 677–695. doi: 10.1016/j.neuroimage.2007.07.032
Hassan, M., Dufor, O., Benquet, P., Berrou, C., and Wendling, F. (2015). “Identification of brain networks with high time/space resolution using dense EEG,” in 2015 7th International IEEE/EMBS Conference on Neural Engineering (NER) (Montpellier: IEEE), 1060–1063. doi: 10.1109/NER.2015.7146810
Jirayucharoensak, S., Pan-Ngum, S., and Israsena, P. (2014). EEG-based emotion recognition using deep learning network with principal component based covariate shift adaptation. Sci. World J. 2014:627892. doi: 10.1155/2014/627892
Kober, H., Barrett, L. F., Joseph, J., Bliss-Moreau, E., Lindquist, K., and Wager, T. D. (2008). Functional grouping and cortical-subcortical interactions in emotion: a meta-analysis of neuroimaging studies. NeuroImage 42, 998–1031. doi: 10.1016/j.neuroimage.2008.03.059
Koelstra, S., Muhl, C., Soleymani, M., Lee, J.-S., Yazdani, A., Ebrahimi, T., et al. (2012). Deap: a database for emotion analysis; using physiological signals. IEEE Trans. Affec. Comput. 3, 18–31. doi: 10.1109/T-AFFC.2011.15
Illán, I. A., Górriz, J. M., López, M. M., Ramírez, J., Salas-Gonzalez, D., Segovia, F., et al. (2011). Computer aided diagnosis of Alzheimer's disease using component based SVM. Appl. Soft Comput. 11, 2376–2382. doi: 10.1016/j.asoc.2010.08.019
Liu, Y., and Sourina, O. (2013). “EEG databases for emotion recognition,” in 2013 International Conference on Cyberworlds (CW) (Yokohama), 302–309. doi: 10.1109/CW.2013.52
Liu, Y.-H., Cheng, W.-T., Hsiao, Y.-T., Wu, C.-T., and Jeng, M.-D. (2014). “Eeg-based emotion recognition based on kernel fisher's discriminant analysis and spectral powers,” in 2014 IEEE International Conference on Systems, Man and Cybernetics (SMC) (San Diego, CA), 2221–2225. doi: 10.1109/SMC.2014.6974254
López, M., Ramírez, J., Górriz, J. M., Alvarez, I., Salas-Gonzalez, D., Segovia, F., et al. (2011). Principal component analysis-based techniques and supervised classification schemes for the early detection of alzheimer's disease. Neurocomputing 74, 1260–1271. doi: 10.1016/j.neucom.2010.06.025
Marrero-Fernandez, P., Montoya-Padron, A., i Capo, A. J., and Rubio, J. M. B. (2014). Evaluating the research in automatic emotion recognition. IETE Tech. Rev. 31, 220–232. doi: 10.1080/02564602.2014.906863
Mattia, D., Cincotti, F., Mattiocco, M., Scivoletto, G., Marciani, M. G., and Babiloni, F. (2006). Motor-related cortical dynamics to intact movements in tetraplegics as revealed by high-resolution EEG. Hum. Brain Mapp. 27, 510–519. doi: 10.1002/hbm.20195
Muthu Meena, S., Vimala, K., and Kalaivani, V. (2015). Emotional stress recognition using multi-modal bio-signals. Biomet. Bioinform. 7, 17–22.
Nordhjem, B., Curcic-Blake, B. J. T., Meppelink, A. M., Renken, R. J., De Jong, B. M., Leenders, K. L., et al. (2016). Lateral and medial ventral occipitotemporal regions interact during the recognition of images revealed from noise. Front. Hum. Neurosci. 9:678. doi: 10.3389/fnhum.2015.00678
Pascual-Marqui, R., Michel, C., and Lehmann, D. (1994). Low resolution electromagnetic tomography: a new method for localizing electrical activity in the brain. Int. J. Psychophysiol. 18, 49–65. doi: 10.1016/0167-8760(84)90014-X
Phillips, C., Rugg, M. D., and Friston, K. J. (2002). Systematic regularization of linear inverse solutions of the EEG source localization problem. NeuroImage 17, 287–301. doi: 10.1006/nimg.2002.1175
Robinson, I. B., and Robinson, M. D. (2009). Measures of emotion: a review. Cogn. Emot. 23, 209–237. doi: 10.1080/02699930802204677
Schoffelen, J.-M., and Gross, J. (2009). Source connectivity analysis with MEG and EEG. Hum. Brain Mapp. 30, 1857–1865. doi: 10.1002/hbm.20745
Siegert, I., Bock, R., Vlasenko, B., Philippou-Hubner, D., and Wendemuth, A. (2011). “Appropriate emotional labelling of non-acted speech using basic emotions, geneva emotion wheel and self assessment manikins,” in 2011 IEEE International Conference on Multimedia and Expo (ICME) (Barcelona), 1–6. doi: 10.1109/ICME.2011.6011929
Sirca, F., Onorati, F., Mainardi, L., and Russo, V. (2015). Time-varying spectral analysis of single-channel EEG: application in affective protocol. J. Med. Biol. Eng. 35, 367–374. doi: 10.1007/s40846-015-0044-5
Soleymani, M., Pantic, M., and Pun, T. (2012). Multimodal emotion recognition in response to videos. IEEE Trans. Affect. Comput. 3, 211–223. doi: 10.1109/T-AFFC.2011.37
Toppi, J., Astolfi, L., Poudel, G. R., Innes, C. R., Babiloni, F., and Jones, R. D. (2016). Time-varying effective connectivity of the cortical neuroelectric activity associated with behavioural microsleeps. NeuroImage 124, 421–432. doi: 10.1016/j.neuroimage.2015.08.059
Verleysen, M., and François, D. (2005). “The curse of dimensionality in data mining and time series prediction,” in Computational Intelligence and Bioinspired Systems, eds J. Cabestany, A. Prieto, and F. Sandoval (Berlin; Heidelberg: Springer), 758–770. doi: 10.1007/11494669_93
Verma, G. K., and Tiwary, U. S. (2016). Affect representation and recognition in 3d continuous valence–arousal–dominance space. Multi. Tools Appl. 1–25. doi: 10.1007/s11042-015-3119-y
Virvou, M., Tsihrintzis, G. A., Alepis, E., Stathopoulou, I.-O., and Kabassi, K. (2015). Intelligent Interactive Multimedia Systems and Services in Practice, Chapter On the Use of Multi-attribute Decision Making for Combining Audio-Lingual and Visual-Facial Modalities in Emotion Recognition Cham: Springer International Publishing.
Wang, X.-W., Nie, D., and Lu, B.-L. (2014). Emotional state classification from EEG data using machine learning approach. Neurocomputing 129, 94–106. doi: 10.1016/j.neucom.2013.06.046
Keywords: emotion classification, electroencephalograph, brain mapping, LORETA, MSP, region of interest
Citation: Padilla-Buritica JI, Martinez-Vargas JD and Castellanos-Dominguez G (2016) Emotion Discrimination Using Spatially Compact Regions of Interest Extracted from Imaging EEG Activity. Front. Comput. Neurosci. 10:55. doi: 10.3389/fncom.2016.00055
Received: 01 February 2016; Accepted: 24 May 2016;
Published: 20 July 2016.
Edited by:
Jose Manuel Ferrandez, Universidad Politecnica de Cartagena, SpainCopyright © 2016 Padilla-Buritica, Martinez-Vargas and Castellanos-Dominguez. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Jorge I. Padilla-Buritica, amlwYWRpbGxhQHVuYWwuZWR1LmNv