- Computer Science Department, Universidad Carlos III de Madrid, Madrid, Spain
Nowadays, the recognition of emotions in people with sensory disabilities still represents a challenge due to the difficulty of generalizing and modeling the set of brain signals. In recent years, the technology that has been used to study a person’s behavior and emotions based on brain signals is the brain–computer interface (BCI). Although previous works have already proposed the classification of emotions in people with sensory disabilities using machine learning techniques, a model of recognition of emotions in people with visual disabilities has not yet been evaluated. Consequently, in this work, the authors present a twofold framework focused on people with visual disabilities. Firstly, auditory stimuli have been used, and a component of acquisition and extraction of brain signals has been defined. Secondly, analysis techniques for the modeling of emotions have been developed, and machine learning models for the classification of emotions have been defined. Based on the results, the algorithm with the best performance in the validation is random forest (RF), with an accuracy of 85 and 88% in the classification for negative and positive emotions, respectively. According to the results, the framework is able to classify positive and negative emotions, but the experimentation performed also shows that the framework performance depends on the number of features in the dataset and the quality of the Electroencephalogram (EEG) signals is a determining factor.
Introduction
The recognition of human emotions was proposed long ago as a way for the development of current computing, with the aim of designing machines that recognize emotions to improve the interaction between humans and computer systems (Picard, 2003). Besides, it represents a challenge since this could mean that computers respond in real time and in a personalized way to the affective or emotional states of a person (Kumar et al., 2016).
Emotions in a person play an important role in non-verbal communication and are essential for understanding human behavior (Liu et al., 2011). Moreover, some research related to analyzing emotional behavior and automatic recognition of emotions using machine learning techniques have generated high expectations. First, emotions have been studied from behavioral signals, named emotional signals (Ekman and Friesen, 1969) and from the analysis of body posture and movement (Ekman et al., 1991). On the other hand, various approaches and ways have been tested for the classification of emotions in people under different circumstances, like music (Vamvakousis and Ramirez, 2015), autism (El Kaliouby et al., 2006), the recognition of emotions using electrodermal activity sensors (Al Machot et al., 2019), or e-Healthcare applications (Ali et al., 2016).
Although the human emotional experience has a vital role in our lives, scientific knowledge about human emotions is still minimal (Soleymani et al., 2012). Consequently, emotions classification is a problem considered challenging since emotional states do not have precisely defined limits, and often, the perception between users differs. Therefore, research on the recognition and emotions classification is of importance in real-life applications (Anagnostopoulos et al., 2012; Zhang et al., 2020).
Through the recognition of speech and the processing of facial gestures, it has been possible to classify a person’s emotions, and commonly these approaches have given good results. However, it has been detected that people can manipulate these methods. Therefore, to achieve objectivity in the technique, the source of the emotion must not be easily manipulated (Ackermann et al., 2016). Consequently, a new reliable and objective approach is required to avoid these cases.
In recent years, through the brain–computer interface (BCI), the behavior and emotions of a person have been studied (Nijboer et al., 2008; Chaudhary et al., 2016; Pattnaik and Sarraf, 2018). These studies indicate that BCI technology offers an additional benefit as it is a method that cannot easily be manipulated by the person. Therefore, it is possible to obtain valid and accurate results by analyzing brain signals obtained using a BCI (Patil and Behele, 2018).
People with any sensory disability often do not have access to current technology. For this reason, it is necessary to develop new ways of communication and interaction between the human and the computer to give support to these people considering their disability and the degree it affects them. Therefore, the effective adaptation of a brain–machine interface to the recognition of the user’s emotional state can be beneficial to society (Picard, 2010).
Commonly, people with total or partial visual disability have difficulties completing their daily tasks (Leung et al., 2020). This is associated with dependence when carrying out daily activities, even in some cases with decreased physical activity (Rubin et al., 2001). Frequently, people with visual disabilities need to use support tools that allow them to interact with the environment around them, so they must alter their behavior according to their needs. Therefore, changing how people with visual disabilities communicate, intervene, and express themselves in their environment through the recognition of their emotions would improve their quality of life. Besides, this would positively impact their daily lives since it would put them on an equal footing in current technology access and use.
In various studies related to the classification of emotions through biological signals such as brain signals, music has been used as a source to induce human emotions. Besides, music is considered capable of evoking a series of emotions and affect people’s mood (Koelsch, 2010). However, music’s influence on emotions is often unknown due to individual preference and appreciation for music (Naser and Saha, 2021). The action of listening to music and psychological processes such as perception, attention, learning, and memory are involved. Therefore, music has been considered a useful tool to help study the human brain’s functions (Koelsch, 2012). Additionally, music can provoke strong emotional responses in listeners (Nineuil et al., 2020). Moreover, it has been shown that music is used for understanding human emotions and their underlying brain mechanisms (Banerjee et al., 2016). For these reasons, music is considered adequate to induce and study various human emotions, including positive and negative (Peretz et al., 1998). Taking into account various psychological aspects and the effects of music on emotions, music has been studied in the regulation of moods in people (Van Der Zwaag et al., 2013), the effects of music on memory (Irish et al., 2006), recognition of brain patterns while listening to music (Sakharov et al., 2005), etc.
Previously, this paper’s authors analyzed the research that proposes the classification of emotions in people with visual disabilities (López-Hernández et al., 2019). The results showed that new approaches that specifically consider people with visual disabilities and the study of their emotions are still required. Based on these results, the design of a system that classifies the affective states of people with visual disabilities was proposed by identifying a person’s emotional responses when they are auditory stimulated.
For the reasons mentioned above, this research’s main motivation is to provide an integrated framework for acquiring brain signals through a BCI, characterizing brain activity models, and defining machine learning models for the automatic classification of emotions, focused on people with visual disabilities.
This study expects to obtain new evidence on the application of BCIs, affective computing, and machine learning, oriented toward the development of communication and interaction alternatives between systems and people with visual disabilities.
Likewise, the challenges associated with this research are the analysis and evaluation of emotional behavior as well as the perception of the responses to an auditory stimulus of people with visual disabilities.
Related Work
Next, a review of related works on applying a BCI for the recognition and classification of people’s emotions using machine learning algorithms is presented.
An EEG signal-based system for automatic recognition of emotions was proposed to examine different methods of extracting EEG features, channel locations, and frequency bands (Ackermann et al., 2016). Machine learning algorithms such as support vector machines (SVMs), random forests (RFs), and decision trees (DTs) were evaluated with pre-processed data for the analysis of emotions, based on physiological data provided during the training and testing tasks. In their results and experimental findings, the authors report that the RF algorithm behaves better in recognition of emotions from signals coming from the EEG. Likewise, they mention that although it is possible to recognize human emotions from other sources, the most reliable way is through EEG signals due to this approach’s objectivity.
Another study proposes the identification of four emotions through the analysis of EEG and the exploration of machine learning algorithms such as Multiclass SVM for the emotion classification task (Patil and Behele, 2018). The results indicate that the model obtains a 91.96% precision in the classification of emotions. Likewise, it is mentioned that an EEG is a more reliable data source for the study of emotions since the subject cannot alter the data.
An approach to the acquisition and processing of the EEG signals obtained using the Emotiv Epoc+ device and the evaluation of a neural network model for the classification of emotional states of people without disabilities reports results of 85.94, 79.69, and 78.13% for valence, excitement, and dominance, respectively (Sánchez-Reolid et al., 2018).
A model for identifying human emotions using EEG signals and Multi-Feature Input Deep Forest Model has been used as an alternative to classifying five emotions, neutral, angry, sad, happy, and pleasant (Fang et al., 2021). In this study, EEG signals from a public dataset for emotion analysis (DEAP) are used. Data processing involves dividing the EGG signals into several frequency bands, processing the power spectral density, the differential entropy of each frequency band, and the original signal as features of the model. Results show that the MFDF model achieves 19.53% more precision with the compared algorithms (RF, SVM, and KNN).
The detection of emotions from EEG signals is also studied by Ramirez and Vamvakousis (2012). The paper describes an automatic approach to emotion detection based on brain activity using the Emotiv Epoc+ headset. In this study, a group of men and women were stimulated auditorily with 12 sounds from the IADS database. During the extraction of features, alpha waves (8–12 Hz) and beta waves (12–30 Hz) were considered, using a bandpass filter and the Fourier analysis of frequency. Linear Declining Analysis (LDA) and SVM algorithms were evaluated for the two-class classification task. Finally, the results indicated that the best classification results for excitation and valence were 83.35 and 86.33%.
A new normalization method of features named stratified normalization is studied to classify emotions from EEG signals (Fdez et al., 2021). In this research, the SEED dataset is used, and the data on the effects of three independent variables (labeling method, normalization method, and feature extraction method) are recorded. This method proposes an alternative for the normalization of features to improve the precision of the recognition of emotions between people. The results indicate 91.6% in the classification of two categories (positive and negative) and 79.6% in the classification of three categories (positive, negative, and neutral).
Other research shows the analysis and evaluation of machine learning, SVM, and K-nearest neighbors (KNN) methods to classify a person’s emotions while observing a visual stimulus (Mehmood and Lee, 2015). In this research, five people (without disabilities) participated in the experiment, and the EEG data were recorded through the Emotiv Epoc+ headset. The processing of the EEG signals was through the EEGLAB toolbox applying the Independent Component Analysis (ICA) technique. The best result of the application of the automatic learning methods for the classification of emotions was 61% accuracy for KNN, as opposed to SVM, which obtained 38.9% accuracy.
K-nearest neighbors algorithm and its functioning to classify emotions are described by Kaundanya et al. (2015). This proposal is a method for EEG signal acquisition tasks, pre-processing, feature extraction, and emotion classification. Several subjects were stimulated for the emotions of sadness–happiness, and the data acquisition was performed with the ADInstruments device. The recorded EEG signals were processed by applying a bandpass filter (3–35 Hz) to remove the signals’ noise. The results indicated that the KNN algorithm is viable for the classification task.
Decision tree classifiers for EEG signals have been used in different research works. A fast and accurate DT structure-based classification method is used for classifying EEG data with computer cursor up/down/right/left movement images (Aydemir and Kayikcioglu, 2014). The detect epileptic seizure in EEG signals uses a hybrid system based on DT classifier and fast Fourier transform (FFT) (Polat and Güneş, 2007). DT and a BCI have also been used to assist patients who are nearly or entirely “locked-in,” i.e., cognitively intact but unable to move or communicate (Kennedy and Adams, 2003).
In summary, several approaches have been proposed to classify emotions, from voice recognition or facial expressions to BCI, to extract brain signals (EEG). Although these methods have been tested in different settings and their results are correct, the literature mentions that the most reliable method is the use of brain signals (EEG) due to its objectivity in reading the data. Additionally, the classification of emotions for the development of new systems that respond to the emotional states of a person has been analyzed in different research works. In addition, machine learning models, previously labeled datasets and different scenarios have been explored, and the results demonstrate the viability of the proposals. However, the authors consider that new scenarios must be evaluated, considering the classification of affective states in people with disabilities.
The remainder of this article is organized as follows: In section “Related Work,” the related works are discussed, and section “Materials and Methods” presents the methods and materials used in this research. In section “Proposed Framework,” the proposed framework is described, highlighting its principal features. Section “Results” shows the results obtained from the experimentation after the implementation of the proposed model. Possible causes of the results are discussed in section “Discussion.” Finally, section “Conclusion” presents the conclusions and future works related to the implementation of systems capable of recognizing and responding in real time to the emotions of a person with disabilities.
Materials and Methods
This section describes the tools and methods used during the experimental phase of this research.
Headset Emotiv Epoc+
Emotiv Epoc+ is a high-resolution portable EEG device, which is used to record brain signals. This device has 14 electrodes for reading brain signals and two CML/DRL reference electrodes. It is designed to operate quickly during the tasks of acquisition and processing of brain signals (Emotiv Epoc, 2019). The configuration of the device Emotiv Epoc+, for the acquisition of the EEG signals, is supported by the sensors: AF3, F7, F3, FC5, T7, P7, O1, O2, P8, T8, FC6, F4, F8, and AF4. Figure 1 shows the electrode locations for Emotiv Epoc+.
Participants
The present study was developed in collaboration with the ONCE social group from Madrid, Spain. This group involves the National Organization of the Spanish Blind (ONCE) and other related entities. ONCE supports people with needs derived from blindness or a severe visual impairment by providing of specialized Social Services.
This study involved seven participants, five men and two women, from whom the ONCE social group invited four people to participate in the study. The age of the participants is between 40 and 55 years old. Previously, all participants reported having normal hearing, and before the experiment, they gave their consent in a confidentiality document to process personal data and participate in the study. Likewise, they were informed of the procedure and of their right to suspend the study. Considering the participants, the experiment was carried out following the principles of the Declaration of Helsinki.
Stimuli
For the experimentation of this study, two classical music audios with different musical styles were selected, the first being joy–happiness and the second being fear–suspense. From these, 40 stimuli (audios) with 5 s each have been generated and selected. The purpose of using stimuli of different musical styles is to induce different affective states (emotions) in the participants.
Proposed Framework
The related work has exposed different research works for the identification of a person’s emotional responses when they are auditory stimulated. As it has been stated, despite the number of works in this area, future research is needed for developing high-performance BCI systems to allow people with needs to perform activities of daily living (Yuan and He, 2014).
Furthermore, due to the advancement of computational tools, the task of recognition and classification of human emotions based on machine learning models has generated interest (Asghar et al., 2019). For this reason, in this research, different machine learning models are evaluated looking for the one with the best performance in this problem.
This manuscript presents a new framework focused on people with visual disabilities, taking into account those findings. The framework is composed of different components and stages:
(A) Data acquisition: EEG signals data acquisition by a BCI interface and brain activity models characterization.
(B) Pre-processing: Analysis of brain activity models and EEG data signals for feeding the training and test process of the machine learning models.
(C) Machine Learning: Definition and evaluation of different machine learning models.
(D) Classification: Automatic classification of basic emotions (positive or negative).
The different components and stages of the framework are shown in Figure 2 and detailed in the following subsections.
Data Acquisition
The first step of the proposed framework is data acquisition and collection. During the presentation of a stimulus, the Emotiv Epoc+ device is used to read and record real-time people’s brain activity. These sensors are placed on the participants as shown in Figure 1, following the guidelines of the international 10–20 standard for electrode positioning (Sharbrough et al., 1991).
Experiments
Before starting the experimentation stage, the participant is informed of the data recording procedure and the process to evaluate each stimulus. Subsequently, the participant performs a test with the Emotiv Epoc+ headband to ensure the correct reading of the data; in addition, volume tests with the audio device were carried out to validate that the participant is comfortable.
During each test, the task of inducing different affective states or emotions in the participants was presented using auditory stimulation. Each participant listens to 40 previously selected auditory stimuli, divided into four groups of 10 stimuli, where each stimulus is presented for 5 s. Between each stimulus, the participant has 3 s to rate the stimulus heard and 5 s of silence to evoke a neutral emotional state in the participant. The experimentation process is divided into four stages that are described below.
Experiment 1
In this test, 10 trials’ data are recorded presenting 10 stimuli (in that order): five of Joy–Happiness and five of Fear–Suspense. Likewise, the participant will rate each stimulus as positive (pleasant) or negative (unpleasant) according to their musical preferences.
Experiment 2
During this test, 10 trials are carried out from presenting 10 stimuli (in random order): five of joy–happiness and five of fear–suspense. In this test, the participant rates each stimulus according to their musical preferences, positive (pleasant), or negative (unpleasant), respectively.
Experiment 2
For this test, data from 10 trials are saved by an orderly presentation of five joy–happiness stimuli and five fear–suspense stimuli. Each participant rates each stimulus according to their musical preferences (positive or negative).
Experiment 4
In this last test, a random order is considered for recording the data of 10 trials using five stimuli of joy–happiness and five of fear–suspense, respectively. Each stimulus is rated positive or negative, depending on each participant.
Finally, the data obtained from the signal’s EEG are digitized into a file for each user. The file contains all the information related to the experiment: type of stimulus, the time interval the stimulus was presented, the wave magnitude for each electrode, and the participant’s evaluation for each stimulus.
Pre-processing
This component of the framework processes the source data of the EEG signals of each participant obtained in the data acquisition stage. A component of the framework extracts the signal from each sensor, applies FFT, and filters the signal using a filter band pass between 0.5 and 30 Hz. The result of this task is the conversion of the data into the signal frequencies, delta (0.5–4 Hz), theta (4–8 Hz), alpha (8–15 Hz), beta (15–30 Hz), and an average wave magnitude of each electrode, reducing the amount of data generated and therefore improving its understanding.
Machine Learning
The machine learning component of the framework includes eight machine learning classifiers: RF, logistic regression (LR), multilayer perceptron (MLP), KNN, Linear Discriminant Analysis (LDA), Naive Bayes (NB), DT, and neural networks (NNs) where different experiments are configured, altering the main parameter for each algorithm.
The KNN method is a popular classification method in data mining and statistics because of its simple implementation and significant classification performance. KNN classifier is a type of instance-based learning or non-generalizing learning: it does not attempt to construct a general internal model but simply stores instances of the training data. Classification is computed from a simple majority vote of the nearest neighbors of each point: a query point is assigned to the data class, which has the most representatives within the nearest neighbors of the point. However, it is impractical for traditional KNN methods to assign a fixed k value (even though set by experts) to all test samples (Zhang et al., 2017, 2018). Considering this, for the KNN classifier, the number of neighbors (n_neighbors) parameter is modified in the experimentation (moving from 1 to 20).
Random forest classifier is an ensemble classifier that produces multiple DTs, using a randomly selected subset of training samples and variables. Over the last two decades, the use of the RF classifier has received increasing attention due to the excellent classification results obtained and the speed of processing (Belgiu and Drǎgu, 2016). RF algorithm is a meta estimator that fits a number of DT classifiers on various subsamples of the dataset and uses averaging to improve the predictive accuracy and control over-fitting. The subsample size is always the same as the original input sample size, but the samples are drawn with replacement in this case. The number of trees (NT) in the RF algorithm for supervised learning has to be set by the user. It is unclear whether the NT parameter should be set to the largest computationally manageable value or whether a smaller NT parameter may be enough or, in some cases, even better (Cutler et al., 2012; Probst and Boulesteix, 2018). RF is an algorithm that has been shown to have excellent performance for classification tasks. It uses a set of trees (n_estimators), which are based on the technique of sampling the data (Vaid et al., 2015). Taking this into account, for the RF classifier, the NT parameter (n_estimators) is modified in the experimentation (moving from 1 to 25).
Decision trees are a non-parametric supervised learning method used for classification and regression. The goal is to create a model that predicts the value of a target variable by learning simple decision rules inferred from the data features. Many systems have been developed for constructing DTs from collections of examples (Quinlan, 1987). The study for the number of features to consider when looking for the best split has been under research for many years (Kotsiantis, 2013; Fratello and Tagliaferri, 2018). Taking this into account, for the DT classifier, the number of features (max_features) for the best split is modified in the experimentation (moving from 1 to 14).
Linear Declining Analysis algorithm is a classifier that works with a linear decision limit. However, it is also considered a technique for feature extraction and dimensionality reduction. LDA projects the data in a vector space with a covariance matrix and an average vector of lower dimensions. Finally, the samples are classified according to the closest average vector (Torkkola, 2001; Ye et al., 2005). Furthermore, LDA has been used to reduce the number of dimensions in datasets, while trying to retain as much information as possible.
Logistic regression is a statistical method that examines the relationship between a dependent variable (target) and a set of independent variables (input), which is applied in regression problems, binary classification, and multiclassification. The LR algorithm technique is based on finding a prediction function and a loss function and identifying the parameters that minimize the loss function. In LR classification problems, you first create a cost function and then apply an iterative optimization process to identify the optimal model parameters (Tsangaratos and Ilia, 2016; Fan et al., 2020).
Naive Bayes classifier is a probabilistic algorithm known for being simple and efficient in classification tasks. From a set of training data, it estimates a joint probability between the features (X) and the targets (Y) (Tsangaratos and Ilia, 2016). NB learns the parameters separately for each attribute, which simplifies learning, even in large datasets (Mccallum and Nigam, 1998).
Multilayer perceptron is a type of NN frequently used in pattern recognition problems. Due to its ease of implementation and adaptability to small datasets (Subasi, 2007). The MLP implementation consists of three sequential layers: input (s), hidden (s), and output (s). The hidden layer processes and serves as an intermediary between the input and the output layer (Orhan et al., 2011).
A NN is a model that has been studied in supervised learning approaches. These models are composed of a large number of interconnected neurons, on which parallel calculations are performed to process data and obtain certain knowledge. The learning of a NN is based on rules that simulate biological learning mechanisms. For classification tasks, NN models are important for their ability to adapt and fit the data (Subasi, 2007; Naraei et al., 2017).
Taking this into account, Table 1 shows the configuration of each algorithm and the parameters that have been modified during experimentation.
Classification
At this stage and taking into account the main insights obtained from the related work, machine learning classifiers are trained, evaluated, and validated through cross-validation and different precision metrics. Figure 3 describes the workflow adopted to validate the results of the classification models using cross-validation and to obtain a comparison of the best results in the model’s evaluation process.
All the data recorded from the trials that have been collected during the experimentation stage have been included in two different approaches, looking for the machine learning model with the best fit for the recognition and classification of emotions. The evaluations of users include negative samples (stimuli that the user qualified as unpleasant), neutral samples (periods when the user did not hear any stimuli), and positive samples (stimuli rated as pleasant by the user). Approach A includes negative (unpleasant) and neutral cases. Instead, approach B contains neutral and positive cases. The purpose of generating these approaches is to identify and recognize the difference between cases with different emotional state using machine learning models.
The percentage of samples for training and testing, the dimensionality of the feature vector, and the number of classes to evaluate the machine learning models’ performance are presented in Table 2. The features are extracted from each sensor’s EEG signal spectrum as a potential feature to feed the machine learning model.
The number of features of each dataset depends on the type of waves included. Approaches A1 and B1 use 56 features for the binary classification of “unpleasant – neutral” and “neutral – pleasant,” respectively, extracted from 14 sensors (56 = 14 ∗ 4 {delta, theta, alpha, and beta}). Instead, approaches A2 and B2 employ 28 features for binary classification of “unpleasant – neutral” and “neutral – pleasant,” respectively, obtained from 14 sensors (28 = 14 ∗ 2 {alpha and beta}). In A2 and B2 approaches, it has been chosen to evaluate the alpha and beta frequencies since it is known from the literature that these frequencies reflect active mental states. Therefore, it is proposed to evaluate the importance of this type of frequencies in emotion classification tasks.
During classification tasks, first, the precision of each machine learning algorithm is validated using cross-validation (10 folds). Subsequently, each algorithm is evaluated by training a new model and with a set of samples reserved for its validation. The percentages of samples assigned in each step of the process have been described in Table 2. The results of the accuracy, precision, recall, and F1 score metrics will indicate the ability of the model to generalize new cases. Additionally, it is essential to validate that the model does not present a classification bias toward one of the problem classes.
Results
In this section, the results of the proposed framework are presented. First, the percentage of samples evaluated as negative or positive per participant is reported. Later, machine learning algorithms for classifying affective states in people with visual disabilities are compared. Finally, each algorithm’s performance is analyzed, and precision analysis of the proposed models is presented.
Pre-processing: Dataset Analysis
First, an analysis of each experiment’s responses reveals the percentage of samples that have been evaluated negatively or positively by all participants. Table 3 reports these results, ID Participant, Stimuli evaluated, % of negative samples, and % of positive samples associated with the evaluation by all participants.
Once the EEG signals have been extracted and the average wave values processed from the 14 electrodes of the Emotiv Epoc+ headband, data analysis for inspecting, cleaning, and transforming data to highlight useful information has been performed. For visualizing this process, two boxplot diagrams have been defined. The usefulness of the boxplot diagram is that it offers, by simple visual inspection, a rough idea of the central tendency (through the median), dispersion (through the interquartile) of the symmetry of the distribution (through the symmetry of the graph), and possible outliers in each classifier. The rectangular part of the plot extends from the lower quartile to the upper quartile, covering the center half of each sample. The center lines within each box show the location of the sample medians. The whiskers extend from the box to the minimum and maximum values in each sample, except for any outside or far outside points, which will be plotted separately (Gonzalez-Carrasco et al., 2014; Molnar, 2019).
The detection of these outliers is crucial for understanding possible causes and implications of their presence (Cousineau and Chartier, 2010; Leys et al., 2013). Moreover, the importance of outliers has been studied in different domains and problems (Felt et al., 2017; Iwata et al., 2018; Peiffer and Armytage, 2019).
Figure 4 depicts the variability of the distribution of brain signal values in positive and negative emotions for the 14 electrodes of the BCI. Based on the visual analysis of the dataset, there are some outliers for the electrodes in the distribution of the values (median around 4200 μV). In the same way, the behavior of the signals is quite similar for positive and negative emotions. For this reason, it is essential to mention that the existence of outliers could determine the reaction and behavior of the brain to a given stimulus. Therefore, for the experimentation of this research, these outliers have been taken into account.
Machine Learning
As mentioned above, the goal of the framework is to predict the basic emotions, positive or negative, of the participants. For this reason, at the machine learning stage, several machine learning algorithms and two approaches were evaluated to obtain the best performance. This section shows the process of evaluation and validation performed to determine the contribution of the research.
Classification
Approach A
Firstly, the models’ performance for approaches A1 and A2 with 56 and 28 features, respectively, is presented in Figure 5. The configuration to evaluate the performance contemplates cross-validation with 10 folds for all models. Subsequently, the comparison of the performance of each model for approaches A1 and A2 is presented in Table 4. The first part describes each model’s results with the precision achieved during the cross-validation, the average, standard deviation, and the precision minimum and maximum. Next, the evaluation of the models is detailed; this includes the accuracy result for training and test tasks and the metrics obtained in precision, recall, and F1 score.
In the results for approach A1 with 56 features, it is observed that RF is the model that best adapts to the problem of recognition of Negative and Neutral emotions. RF achieves a mean accuracy of 87%. Moreover, the performance of the KNN, NB, and DT models is very similar, reaching between 74 and 77%. Otherwise, the remaining LR, MLP, LDA, and NN models’ average performance is less than 70% mean accuracy. The evaluation of approach A1 with 56 features shows that the best model is RF with an average precision of 87% and a minimum accuracy of 80%, and a maximum of 94%. KNN behavior indicates an average of 75%, besides a minimum and maximum accuracy of 65 and 85%, respectively. On the other hand, NB averages 74% accuracy, with 67% as the minimum level and 85% as the maximum level. Meanwhile, the DT model achieves an average accuracy of 77%, with a minimum performance of 63% and a maximum of 85%. The average performance of the LR, MLP, LDA, and NN models is less than 70%, with minimal accuracy ranging from 44% up to a maximum accuracy of 76%.
Finally, the results of the validation of each model are shown. These indicate that RF achieves the best result for the classification of new cases; it obtains 83% accuracy. Additionally, the F1 score metric reports 73% for negative cases and 87% for neutral cases.
The performance of the models of approach A2 with 28 features (alpha and beta frequencies) is shown in the right part of Figure 5. The results again indicate that RF with 85% mean accuracy is the model with better levels. On the other hand, KNN, LDA, NB, and DT reach a mean accuracy between 70 and 80%. These models’ minimum values are between 64 and 69% and maximum values are from 81 to 85%. Finally, for LR, MLP, and NN models, the results indicate the lower performance with values less than 70% mean accuracy.
Likewise, in Table 4, the evaluation results of approach A2 with the alpha and beta frequencies are detailed. The results show that RF achieves the best mean result with 85% accuracy, with a minimum of 74% and a maximum of 96%. The results of the KNN model indicate a mean accuracy of 76% and a minimum and maximum of 67 and 84%, respectively. Instead, NB achieves a mean of 77%, a minimum of 69%, and a maximum of 85% accuracy. On the other hand, DT reports a mean of 75% accuracy and 69 and 85% as minimum and maximum, respectively. The LDA model achieves a mean of 72%, with a minimum of 64% and a maximum of 85%. The lowest performance models are LR, MLP, and NN; they show minimums of 46% to 57%, mean between 59 and 66%, and maximums of 69 up to 76% accuracy.
Finally, the models’ validation data with approach A2 with 28 features are presented in the lower part of Table 4. The data show that RF obtains the best result with 85% accuracy in the classification of new cases. Furthermore, this result is validated with the F1 Score metric, which shows 80% for negative classes and 88% for neutral classes.
Based on the results, it is observed that RF is the model that best adapts to the classification of new cases, considering the two scenarios proposed for approaches A1 and A2. Besides, the data show that RF is capable of classifying similarly for cases that are negative and neutral.
Approach B
Initially, Figure 6 details the different models’ behavior evaluated for approaches B1 and B2 with 56 and 28 features, respectively. Next, the statistics of the behavior of the models on approach B are presented in Table 5.
Firstly, in the results of approach B1 with 56 features, it is observed that RF is the best model, achieving a mean precision of 86%. The models that reach similar values of 70% are KNN (74%), NB (70%), and DT (77%). In contrast, LR, MLP, LDA, and NN show a mean performance below 67% accuracy. The data RF shows a mean performance of 86% accuracy, and a minimum and maximum of 80 and 93%, respectively. On the other hand, KNN results indicate a maximum accuracy of 84%, a mean of 74%, and a minimum of 67%. NB model shows a mean accuracy of 70%, and minimum and maximum values of 62 and 79%. DT results indicate a mean accuracy of 77%, as well as a minimum and maximum of 66 and 82%, respectively. Finally, LR, MLP, LDA, and NN models show a mean accuracy of 66%, minimums from 45 to 56%, and maximums from 59 to 79%.
On the other hand, the models’ behavior for approach B2 with 28 features is represented in the right part of Figure 6. First, it is observed that RF obtains a mean accuracy of 87%. On the other hand, the KNN, NB, and DT models achieve a mean accuracy of more than 70%. Lastly, the LR, MLP, LDA, and NN models report results below 68%.
The performance statistics of approach B2 with 28 features show RF as the best model, which obtains a mean accuracy of 87%; maximum and minimum correspond to 93 and 83%, respectively. The KNN model achieves a mean accuracy of 71%, a minimum value of 65%, and a maximum of 79%. On the other hand, although NB and DT obtain a superior performance of 74% mean accuracy, it is below the RF model. The data minimum and maximum of NB are 64 and 83% accuracy. Instead, DT achieves a minimum value of 70% and a maximum value of 83%. The lowest results belong to LR, MLP, LDA, and NN models; they obtain a mean accuracy of less than 68%. The minimum percentages of these models range from 46 to 62% and the maximums range from 62 to 72%.
Finally, Table 5 presents the results of the validation of each model. The data indicate that RF is the model with the best result; it obtains 88% in the generalization of new cases, and F1 Score validation metric indicates 89% for neutral classes and 86% for positive classes.
According to the results for approaches B1 and B2, RF obtains the best performance in classifying neutral and positive cases. Besides, the data indicate that RF shows balanced performance in the classification of the proposed cases.
In order to demonstrate the best performance of the evaluated models, the average classification precision achieved in the cross-validation task and the result in the test task of each model are presented in the Figure 7. RF stands out for its uniform performance in the four evaluated approaches. Approach A1-56 obtains the best performance with 87% and 83% in the cross-validation and evaluation of the model, respectively. In comparison, approach A2-28 achieves the same result, 85% in both tasks. For approaches B1-56 and B2-28, RF achieves the best results. In B1-56, RF achieves 86% during cross-validation and 88% in model testing. In B2-28, RF obtains 87% and 88% accuracy in cross-validation and model evaluation, respectively.
Figure 7. Accuracy comparison in cross-validation and models evaluation. (A) Classification accuracy in approach A-1. (B) Classification accuracy in approach A-2. (C) Classification accuracy in approach B-1. (D) Classification accuracy in approach B-2.
The importance of the features of the A2 approach (28 features) and the B2 approach (28 features) is shown in Figure 8 [Tree SHAP technique (Lundberg et al., 2020)]. Feature relevance is calculated as the decrease in node impurity weighted by the probability of reaching that node. The node probability can be calculated by the number of samples that reach the node, divided by the total number of samples. In both cases, the higher the value, the more important the feature. Looking at the feature sensibility analysis, the relevant features are similar for both techniques. Therefore, future work should include significant features as a small set of electrodes (F3, T7, T8, F7, AF4, etc.) for trying to achieve similar performance with less complexity.
Figure 8. Feature sensibility analysis by SHAP technique. RF best result in approach A2 (upper) and B2 (lower).
To evaluate the relevance of each participant’s dataset on the best model obtained, these are evaluated considering the results of the classification stage. Based on the data, RF is the model that is considered appropriate to evaluate all scenarios with the data per user. It is important to note that during data evaluation of each participant, a similar process is followed for the evaluation of each proposed approach, which consists of selecting the negative and neutral or neutral and positive samples, corresponding to approaches A1–A2 and B1–B2, respectively. The results per participant and the performance of the data on each model are presented in Table 6. It is identified that the performance of the data of participants 1, 2, 3, and 4 are similar, and they obtain accuracy results superior to the rest of the participants. Furthermore, it is possible to visualize that subjects 5, 6, and 7 obtain lower percentages of precision than the rest of the participants.
On the other hand, it is observed that the A2 and B2 models that include 28 features with the alpha and beta frequencies obtain slightly higher results compared to the models with 56 features that include the alpha, beta, theta, and delta frequencies.
Finally, Table 7 includes an accuracy comparison among relevant related research and the approach proposed by the authors (at the bottom of the table).
Discussion
This section discusses the main insights and breakthroughs regarding the results obtained with the framework proposed in this work.
Firstly, this study presents a new scenario for the recognition and classification of emotions in people with visual disabilities, a group of people not previously evaluated. The work defines and implements a framework through a non-invasive BCI (Emotiv Epoc+) with a set of auditory stimuli. From the records obtained, two datasets were formed with the stimuli classified as negatives-neutrals and neutrals-positives. Subsequently, the model’s RF, LR, MLP, KNN, LDA, NB, DT, and NN were configured and evaluated to identify the model with the best performance in recognition and classification tasks emotions from EEG data.
The results show that in the individual evaluation of the stimuli, participants 1, 3, and 4 evaluated more than 50% of the stimuli positively. Instead, participants 2, 5, 6, and 7 had a balanced evaluation toward stimuli. On the other hand, it is possible to observe brain signal variability concerning stimuli considered positively and negatively (see Figure 4). Although the behavior shows similar values of 4,200 μV, outliers were recorded in the data. That could indicate a different reaction of the participants to the presented stimuli. It is important to note that the entire data have been considered during the machine learning tasks. That is, the outliers have not been omitted. Additionally, the data show that RF is the model with the best performance during classification tasks. In evaluating results, the RF model in the A1-56 and A2-28 approaches to negative emotions achieved a classification accuracy of 83% and 85%, respectively. In turn, with the positive emotions of approaches B1-56 and B2-28, he obtained an accuracy of 88% in both cases. The best results obtained from the RF model in approaches A2-28 and B2-28 with negative and positive emotions and analysis of the features’ importance allowed us to recognize that the beta frequencies related to the frontotemporal areas of the brain are important in the decision making of the models. On the other hand, the results show that the algorithms LR, MLP, KNN, LDA, NB, DT, and NN obtain a lower performance compared to RF (see Figure 7). Although, in the validation of the models, it is observed that DT and KNN obtain acceptable results for the classification of positive emotions, this result is not consistent with the identification of negative emotions. Therefore, these models tend to classify toward one type of emotions.
The analysis of the participants’ brain signals’ dataset allows identifying the variability in each subject’s data. This characteristic is relevant because it is considered relative to the perception of each subject toward each stimulus. This agrees with what is stated by Anagnostopoulos et al. (2012), which mentions that people’s emotional perception commonly differs. The evaluation of the different machine learning models and, according to the results obtained from the RF algorithm, their performance coincides with the findings reported in Ackermann et al. (2016). It is stated that RF is a robust algorithm in the processing and recognition of patterns from EEG signals. The test’s precision indicates that RF is an algorithm useful for classifying emotions using EEG signals. Moreover, in the validation of the RF model, it achieves the best results with an accuracy of 85% for negative emotions and 88% for positive emotions. Therefore, the RF classifier shows that it learns in both classes.
Considering the types of frequencies, delta, theta, alpha, and beta, different machine learning models have been trained and evaluated to determine their ability to recognize and classify different affective states of a group of people with visual disabilities. The data show that models that consider alpha and beta frequencies perform slightly better than models that consider all frequencies. The results show that models that consider alpha and beta frequencies perform slightly better than models that consider all frequencies. These results coincide with Ramirez and Vamvakousis (2012), who mention that the most important frequencies are alpha waves (8–12 Hz), which predominate in mental states of relaxation, and beta waves (12–30 Hz), which are active during states with intense mental activity. Similarly, the results show that the frontotemporal brain areas associated with the beta frequency show the greatest contribution to the performance of the models (see Figure 8). Finally, the model’s performance proposed in this research reaches values comparable to other research (Fdez et al., 2021) to classify emotions into two categories (negative and positive).
Due to the limitations of this study in processing brain activity from EGG data, it is important to consider different signal acquisition and processing aspects. The first factor needs to check all the sensors; this avoids errors in the recorded data. On the other hand, signal processing includes considering aspects such as noise generation. Previous tests are necessary to minimize noise generation, ensuring that the participant feels comfortable with the device; this avoids unexpected data generated by involuntary movements.
Finally, one way to respond to the study limitations is to increase the participant population in the experimentation stage. This would result in a greater number of evaluations toward the stimuli. Another criterion is to expand the number of stimuli and experimental sessions to obtain a large amount of information related to people’s emotional perception of different auditory stimuli. In turn, this would allow extending the analysis of the behavior of brain signals and their response to specific stimulus.
As future work, the authors propose to extend the framework with more techniques such as recurrent neural networks (RNN), convolutional neural networks (CNN), or long short-term memory (LSTM). Those techniques could be compared with the current ones for obtaining a more in-depth study of brain waves and emotions in people with visual disability. Additionally, it is proposed to explore the brain regions’ behavior using 2D and 3D maps of the participants’ brain activity. This process will allow recognizing the brain areas that reflect high or low activity during the stimulation process. Future research can also consider incorporating data from other sources; i.e., the framework will have more than one entry at the same time. Adding data from a new source, other than EEG brain signals, will provide more knowledge for the classification of a person’s affective state and could improve the accuracy of the model. Also, as exposed in the section “Discussion,” a reduced dataset can be tested, taking into account the relevant features of the sensitivity analysis. Besides, as stated in the related work, many of the models proposed for the classification of emotions have not been evaluated in real time. Therefore, the authors take into account the assumptions made by Lotte et al. (2018) and propose as future research to adapt and evaluate the framework as a BCI for real-time emotion recognition.
It is remarkable that, although the participants of this research have visual disabilities (population not previously tested), the authors’ proposal reached similar levels of accuracy compared to other research for the classification of people’s emotions.
Conclusion
It should be taken into account that emotions play an essential role in many aspects of our daily lives, including decision making, perception, learning, rational thinking, and actions. Likewise, it should be considered that the study of emotion recognition is indispensable (Pham and Tran, 2012).
In this work, the authors have explored and analyzed a previously unreported scenario, the classification of emotions in people with visual disabilities. The most important aspects of the framework are as follows: (i) It is a twofold framework. The first is mainly focused on data acquisition (signal extraction) with a BCI device using auditory stimuli. The second is concerned with analysis techniques for the modeling of emotions and machine learning models to classify emotions. (ii) The framework can be expanded with more machine learning algorithms, and therefore it increases the flexibility. (iii) Experimentation is focused on people with visual disabilities. Experimentation results show that 28 feature approaches, including alpha and beta frequencies, performed best for emotion recognition and classification. According to these models’ performance, the achieved accuracy is 85 and 88% in the classification of negative and positive emotions, respectively. Therefore, it is considered that feature selection plays a key role in classification performance. Also, an analysis of features illustrates that the brain’s frontotemporal areas linked to beta frequency have the most significant contribution to the proposed models’ performance. Finally, it has been proposed to continue research based on brain signals and to incorporate new sources of information from people with disabilities, to develop new ways of communication and technological interaction that will allow them to integrate into today’s society.
Data Availability Statement
The datasets presented in this article are not readily available because data protection. Requests to access the datasets should be directed to corresponding author.
Ethics Statement
Ethical review and approval was not required for the study on human participants in accordance with the Local Legislation and Institutional Requirements. The patients/participants provided their written informed consent to participate in this study.
Author Contributions
JL-H, IG-C, and JL-C: writing—original draft preparation. JL-H, IG-C, JL-C, and BR-M: writing—review and editing. IG-C and BR-M: supervision and funding acquisition. All authors contributed to the article and approved the submitted version.
Funding
This work was supported by the National Council of Science and Technology of Mexico (CONACyT), through grant number 709656.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
References
Ackermann, P., Kohlschein, C., Bitsch, J. Á, Wehrle, K., and Jeschke, S. (2016). “EEG-based automatic emotion recognition: feature extraction, selection and classification methods,” in Proceedings of the 2016 IEEE 18th International Conference on E-Health Networking, Applications and Services (Healthcom), Munich. doi: 10.1109/HealthCom.2016.7749447
Al Machot, F., Elmachot, A., Ali, M., Al Machot, E., and Kyamakya, K. (2019). A deep-learning model for subject-independent human emotion recognition using electrodermal activity sensors. Sensors 19:1659. doi: 10.3390/s19071659
Ali, M., Mosa, A. H., Al Machot, F., and Kyamakya, K. (2016). “EEG-based emotion recognition approach for e-healthcare applications,” in Proceedings of the 2016 International Conference on Ubiquitous and Future Networks, ICUFN, Vienna, 946–950. doi: 10.1109/ICUFN.2016.7536936
Anagnostopoulos, C. N., Iliou, T., and Giannoukos, I. (2012). Features and classifiers for emotion recognition from speech: a survey from 2000 to 2011. Artif. Intell. Rev 43, 155–177. doi: 10.1007/s10462-012-9368-5
Asghar, M. A., Khan, M. J., Fawad, Amin, Y., Rizwan, M., Rahman, M., et al. (2019). EEG-based multi-modal emotion recognition using bag of deep features: an optimal feature selection approach. Sensors 19:5218. doi: 10.3390/s19235218
Aydemir, O., and Kayikcioglu, T. (2014). Decision tree structure based classification of EEG signals recorded during two dimensional cursor movement imagery. J. Neurosci. Methods 229, 68–75. doi: 10.1016/j.jneumeth.2014.04.007
Banerjee, A., Sanyal, S., Patranabis, A., Banerjee, K., Guhathakurta, T., Sengupta, R., et al. (2016). Study on brain dynamics by non linear analysis of music induced EEG signals. Phys. A 444, 110–120. doi: 10.1016/j.physa.2015.10.030
Belgiu, M., and Drǎgu, L. (2016). Random forest in remote sensing: a review of applications and future directions. ISPRS J. Photogramm. Remote Sens. 114, 24–31. doi: 10.1016/j.isprsjprs.2016.01.011
Chaudhary, U., Birbaumer, N., and Ramos-Murguialday, A. (2016). Brain–computer interfaces in the completely locked-in state and chronic stroke. Prog. Brain Res. 228, 131–161. doi: 10.1016/bs.pbr.2016.04.019
Cousineau, D., and Chartier, S. (2010). Outliers detection and treatment: a review. Int. J. Psychol. Res. 3, 58–67. doi: 10.21500/20112084.844
Cutler, A., Cutler, D. R., and Stevens, J. R. (2012). “Random forests - ensemble machine learning: methods and applications,” in Ensemble Machine Learning, Vol. 45, (Boston, MA: Springer), 157–175. doi: 10.1007/978-1-4419-9326-7_5
Ekman, P., and Friesen, W. V. (1969). Nonverbal leakage and clues to deception. Psychiatry 32, 88–106. doi: 10.1080/00332747.1969.11023575
Ekman, P., O’Sullivan, M., Friesen, W. V., and Scherer, K. R. (1991). Invited article: face, voice, and body in detecting deceit. J. Nonverb. Behav. 15, 125–135. doi: 10.1007/BF00998267
El Kaliouby, R., Picard, R., and Baron-Cohen, S. (2006). Affective computing and autism. Ann. N. Y. Acad. Sci. 1093, 228–248. doi: 10.1196/annals.1382.016
Emotiv Epoc. (2019). Available online at: https://www.emotiv.com/epoc/ (accessed December, 2020).
Fan, Y., Bai, J., Lei, X., Zhang, Y., Zhang, B., Li, K. C., et al. (2020). Privacy preserving based logistic regression on big data. J. Netw. Comput. Appl. 171:102769. doi: 10.1016/j.jnca.2020.102769
Fang, Y., Yang, H., Zhang, X., Liu, H., and Tao, B. (2021). Multi-feature input deep forest for EEG-based emotion recognition. Front. Neurorobot. 14:617531. doi: 10.3389/fnbot.2020.617531
Fdez, J., Guttenberg, N., Witkowski, O., and Pasquali, A. (2021). Cross-subject EEG-based emotion recognition through neural networks with stratified normalization. Front. Neurosci. 15:626277. doi: 10.3389/fnins.2021.626277
Felt, J. M., Castaneda, R., Tiemensma, J., and Depaoli, S. (2017). Using person fit statistics to detect outliers in survey research. Front. Psychol. 8:863. doi: 10.3389/fpsyg.2017.00863
Fratello, M., and Tagliaferri, R. (2018). “Decision trees and random forests,” in Encyclopedia of Bioinformatics and Computational Biology: ABC of Bioinformatics, Vol. 1–3, (Milan: Elsevier), 374–383. doi: 10.1016/B978-0-12-809633-8.20337-3
Gonzalez-Carrasco, I., Garcia-Crespo, A., Ruiz-Mezcua, B., Lopez-Cuadrado, J. L. J. L., and Colomo-Palacios, R. (2014). Towards a framework for multiple artificial neural network topologies validation by means of statistics. Expert Syst. 31, 20–36. doi: 10.1111/j.1468-0394.2012.00653.x
Irish, M., Cunningham, C. J., Walsh, J. B., Coakley, D., Lawlor, B. A., Robertson, I. H., et al. (2006). Investigating the enhancing effect of music on autobiographical memory in mild Alzheimer’s disease. Dement. Geriatr. Cogn. Disord. 22, 108–120. doi: 10.1159/000093487
Iwata, K., Nakashima, T., Anan, Y., and Ishii, N. (2018). Detecting outliers in terms of errors in embedded software development projects using imbalanced data classification. Stud. Comput. Intell. 726, 65–80. doi: 10.1007/978-3-319-63618-4_6
Kaundanya, V. L., Patil, A., and Panat, A. (2015). “Performance of k-NN classifier for emotion detection using EEG signals,” in Proceedings of the 2015 International Conference on Communication and Signal Processing (ICCSP), Melmaruvathur, 1160–1164. doi: 10.1109/ICCSP.2015.7322687
Kennedy, P. R., and Adams, K. D. (2003). A decision tree for brain-computer interface devices. IEEE Trans. Neural. Syst. Rehabil. Eng. 11, 148–150. doi: 10.1109/TNSRE.2003.814420
Koelsch, S. (2010). Towards a neural basis of music-evoked emotions. Trends Cogn. Sci. 14, 131–137. doi: 10.1016/j.tics.2010.01.002
Kotsiantis, S. B. (2013). Decision trees: a recent overview. Artif. Intell. Rev. 39, 261–283. doi: 10.1007/s10462-011-9272-4
Kumar, N., Khaund, K., and Hazarika, S. M. (2016). Bispectral analysis of EEG for emotion recognition. Procedia Comput. Sci. 84, 31–35. doi: 10.1016/j.procs.2016.04.062
Leung, K. S. C., Lam, K. N. A., and To, Y. Y. E. (2020). Visual Disability Detection System Using Virtual Reality. U.S. Patent No. 10,568,502.
Leys, C., Ley, C., Klein, O., Bernard, P., and Licata, L. (2013). Detecting outliers: do not use standard deviation around the mean, use absolute deviation around the median. J. Exp. Soc. Psychol. 49, 764–766. doi: 10.1016/j.jesp.2013.03.013
Liu, Y., Sourina, O., and Nguyen, M. L. S. K. N. (2011). “Real-Time EEG-based emotion recognition and its applications transactions on computational science XII,” in Lecture Notes in Computer Science (LNCS), Vol. 6670, (Berlin: Springer), 256–277. doi: 10.1007/978-3-642-22336-5_13
López-Hernández, J. L., González-Carrasco, I., López-Cuadrado, J. L., and Ruiz-Mezcua, B. (2019). Towards the recognition of the emotions of people with visual disabilities through brain-computer interfaces. Sensors 19:2620. doi: 10.3390/s19112620
Lotte, F., Bougrain, L., Cichocki, A., Clerc, M., Congedo, M., Rakotomamonjy, A., et al. (2018). A review of classification algorithms for EEG-based brain-computer interfaces: a 10 year update. J. Neural Eng. 15:031005. doi: 10.1088/1741-2552/aab2f2
Lundberg, S. M., Erion, G., Chen, H., DeGrave, A., Prutkin, J. M., Nair, B., et al. (2020). From local explanations to global understanding with explainable AI for trees. Nat. Mach. Intell. 2, 56–67. doi: 10.1038/s42256-019-0138-9
Mccallum, A., and Nigam, K. (1998). “A Comparison of Event Models for Naive Bayes Text Classification,” in Proceedings of the AAAI Workshop AAAI Workshop Learning for Text Categorization, Menlo Park, 41–48.
Mehmood, R. M., and Lee, H. J. (2015). “Emotion classification of EEG brain signal using SVM and KNN,” in Proceedings of the 2015 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Turin, 1–5. doi: 10.1109/ICMEW.2015.7169786
Molnar, C. (2019). Interpretable Machine Learning. Available online at: https://christophm.github.io/interpretable-ml-book/ (accessed December, 2020).
Naraei, P., Abhari, A., and Sadeghian, A. (2017). “Application of multilayer perceptron neural networks and support vector machines in classification of healthcare data,” in Proceedings of Future Technologies Conference, Vancouver, BC, 848–852. doi: 10.1109/FTC.2016.7821702
Naser, D. S., and Saha, G. (2021). Influence of music liking on EEG based emotion recognition. Biomed. Signal Process. Control 64:102251. doi: 10.1016/j.bspc.2020.102251
Nijboer, F., Furdea, A., Gunst, I., Mellinger, J., McFarland, D. J., Birbaumer, N., et al. (2008). An auditory brain-computer interface (BCI). J. Neurosci. Methods 167, 43–50. doi: 10.1016/j.jneumeth.2007.02.009
Nineuil, C., Dellacherie, D., and Samson, S. (2020). The impact of emotion on musical long-term memory. Front. Psychol. 11:2110. doi: 10.3389/fpsyg.2020.02110
Orhan, U., Hekim, M., and Ozer, M. (2011). EEG signals classification using the K-means clustering and a multilayer perceptron neural network model. Expert Syst. Appl. 38, 13475–13481. doi: 10.1016/j.eswa.2011.04.149
Patil, A., and Behele, K. (2018). “Classification of human emotions using multiclass support vector machine,” in Proceedings of the 2017 International Conference on Computing, Communication, Control and Automation, ICCUBEA, Pune. doi: 10.1109/ICCUBEA.2017.8463656
Pattnaik, P. K., and Sarraf, J. (2018). Brain Computer Interface issues on hand movement. J. King Saud Univ. Comput. Inf. Sci. 30, 18–24. doi: 10.1016/j.jksuci.2016.09.006
Peiffer, C., and Armytage, R. (2019). Searching for success: a mixed methods approach to identifying and examining positive outliers in development outcomes. World Dev. 121, 97–107. doi: 10.1016/j.worlddev.2019.04.013
Peretz, I., Gagnon, L., and Bouchard, B. (1998). Music and emotion: perceptual determinants, immediacy, and isolation after brain damage. Cognition 68, 111–141. doi: 10.1016/S0010-0277(98)00043-2
Pham, T. D., and Tran, D. (2012). “Emotion recognition using the emotiv EPOC device,” in Neural Information Processing. ICONIP 2012 Lecture Notes in Computer Science, eds T. Huang, Z. Zeng, C. Li, and C. S. Leung (Berlin: Springer), 394–399. doi: 10.1007/978-3-642-34500-5_47
Picard, R. W. (2003). Affective computing: challenges. Int. J. Hum. Comput. Stud. 59, 55–64. doi: 10.1016/S1071-5819(03)00052-1
Picard, R. W. (2010). Affective computing: from laughter to IEEE. IEEE Trans. Affect. Comput. 1, 11–17. doi: 10.1109/T-AFFC.2010.10
Polat, K., and Güneş, S. (2007). Classification of epileptiform EEG using a hybrid system based on decision tree classifier and fast Fourier transform. Appl. Math. Comput. 187, 1017–1026. doi: 10.1016/j.amc.2006.09.022
Probst, P., and Boulesteix, A. L. (2018). To tune or not to tune the number of trees in random forest. J. Mach. Learn. Res. 18, 1–8.
Quinlan, J. R. (1987). Simplifying decision trees. Int. J. Man Mach. Stud. 27, 221–234. doi: 10.1016/S0020-7373(87)80053-6
Ramirez, R., and Vamvakousis, Z. (2012). “Detecting Emotion from EEG signals using the Emotive Epoc Device,” in Proceedings of the 2012 International Conference on Brain Informatics, Macau, 175–184. doi: 10.1007/978-3-642-35139-6
Rubin, G. S., Bandeen-Roche, K., Huang, G. H., Muñoz, B., Schein, O. D., Fried, L. P., et al. (2001). The association of multiple visual impairments with self-reported visual disability: SEE project. Invest. Ophthalmol. Vis. Sci. 42, 64–72.
Sakharov, D. S., Davydov, V. I., and Pavlygina, R. A. (2005). Intercentral relations of the human EEG during listening to music. Hum. Physiol. 31, 392–397. doi: 10.1007/s10747-005-0065-5
Sánchez-Reolid, R., García, A., Vicente-Querol, M., Fernández-Aguilar, L., López, M., and González, A. (2018). Artificial neural networks to assess emotional states from brain-computer interface. Electronics 7:384. doi: 10.3390/electronics7120384
Sharbrough, F., Chatrian, G. E., Lesser, R. P., Luders, H., Nuwer, M., and Picton, T. W. (1991). American Electroencephalographic Society guidelines for standard electrode position nomenclature. J. Clin. Neurophysiol. 8, 200–202.
Soleymani, M., Lichtenauer, J., Pun, T., and Pantic, M. (2012). A multimodal database for affect recognition and implicit tagging. IEEE Trans. Affect. Comput. 3, 42–55. doi: 10.1109/T-AFFC.2011.25
Subasi, A. (2007). EEG signal classification using wavelet feature extraction and a mixture of expert model. Expert Syst. Appl. 32, 1084–1093. doi: 10.1016/j.eswa.2006.02.005
Torkkola, K. (2001). “Linear Discriminant analysis in document classification,” in Proceedings of the International Conference Data Mining Workshop on Text Mining, San Jose, CA, 1–10.
Tsangaratos, P., and Ilia, I. (2016). Comparison of a logistic regression and Naïve Bayes classifier in landslide susceptibility assessments: the influence of models complexity and training dataset size. Catena 145, 164–179. doi: 10.1016/j.catena.2016.06.004
Vaid, S., Singh, P., and Kaur, C. (2015). Classification of human emotions using multiwavelet transform based features and random forest technique. Indian J. Sci. Technol. 8:7. doi: 10.17485/ijst/2015/v8i28/70797
Vamvakousis, Z., and Ramirez, R. (2015). “Is an auditory P300-based brain-computer musical interface feasible??,” in Proceedings of the 1st International BCMI Workshop, Plymouth, 1–8.
Van Der Zwaag, M. D., Janssen, J. H., and Westerink, J. H. D. M. (2013). Directing physiology and mood through music: validation of an affective music player. IEEE Trans. Affect. Comput. 4, 57–68. doi: 10.1109/T-AFFC.2012.28
Ye, J., Janardan, R., and Li, Q. (2005). Two-Dimensional Linear Discriminant Analysis. Available online at: https://papers.nips.cc/paper/2004/file/86ecfcbc1e9f1ae5ee2d71910877da36-Paper.pdf (accessed December, 2020).
Yuan, H., and He, B. (2014). Brain–computer interfaces using sensorimotor rhythms: current state and future perspectives. IEEE Transact. Biomed. Eng. 61, 1425–1435. doi: 10.1109/TBME.2014.2312397
Zhang, S., Li, X., Zong, M., Zhu, X., and Cheng, D. (2017). Learning k for kNN Classification. ACM Trans. Intell. Syst. Technol. 8, 1–19. doi: 10.1145/2990508
Zhang, S., Li, X., Zong, M., Zhu, X., and Wang, R. (2018). Efficient kNN classification with different numbers of nearest neighbors. IEEE Trans. Neural Netw. Learn. Syst. 29, 1774–1785. doi: 10.1109/TNNLS.2017.2673241
Keywords: emotion classification algorithm, brain–computer interface, machine learning, visual disabilities, affective computing
Citation: López-Hernández JL, González-Carrasco I, López-Cuadrado JL and Ruiz-Mezcua B (2021) Framework for the Classification of Emotions in People With Visual Disabilities Through Brain Signals. Front. Neuroinform. 15:642766. doi: 10.3389/fninf.2021.642766
Received: 16 December 2020; Accepted: 26 March 2021;
Published: 07 May 2021.
Edited by:
Giovanni Ottoboni, University of Bologna, ItalyReviewed by:
Stavros I. Dimitriadis, Greek Association of Alzheimer’s Disease and Related Disorders, GreeceMichela Ponticorvo, University of Naples Federico II, Italy
Copyright © 2021 López-Hernández, González-Carrasco, López-Cuadrado and Ruiz-Mezcua. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Israel González-Carrasco, igcarras@inf.uc3m.es