- Escola Tècnica Superior d'Arquitectura de Barcelona, Universitat Politècnica de Catalunya, Barcelona, Spain
Electroencephalography (EEG) stands as a pioneering tool at the intersection of neuroscience and technology, offering unprecedented insights into human emotions. Through this comprehensive review, we explore the challenges and opportunities associated with EEG-based emotion recognition. While recent literature suggests promising high accuracy rates, these claims necessitate critical scrutiny for their authenticity and applicability. The article highlights the significant challenges in generalizing findings from a multitude of EEG devices and data sources, as well as the difficulties in data collection. Furthermore, the disparity between controlled laboratory settings and genuine emotional experiences presents a paradox within the paradigm of emotion research. We advocate for a balanced approach, emphasizing the importance of critical evaluation, methodological standardization, and acknowledging the dynamism of emotions for a more holistic understanding of the human emotional landscape.
Introduction
In the rapidly evolving field of neuroscience and technological integration, Electroencephalography (EEG) has emerged as a pivotal tool, offering profound insights into human brain functions, especially in the study of emotional responses. Emotions, as complex amalgamations of physiological and cognitive reactions, deeply influence our daily interactions, decision-making processes, and even our experiences within architectural settings (Brunner-Sperdin et al., 2012; Nummenmaa et al., 2012; Lehman et al., 2015; Zhang et al., 2022). Emotions play a significant role in shaping human perception and interaction. This makes the study of emotional responses important in various fields, including healthcare and architecture. Understanding these responses is not just academically intriguing; it also has practical implications in these sectors.
Although EEG is increasingly favored due to its non-invasiveness and high temporal resolution, the in-depth analysis of the data it generates is particularly crucial. Numerous studies have underscored the methods of using EEG in emotional research (Chen et al., 2015; Christensen and Abdullah, 2018; Suhaimi et al., 2020; Dadebayev et al., 2021; Li et al., 2022). However, discussions around the quality of EEG emotional measurements are less prominent. While a careful review acknowledges the contributions of these studies, it also unveils oversights, particularly in addressing the intricacies and challenges of EEG data quality. For instance, Cohen et al. (2023) highlights the need for rigorous validation of EEG data in emotional research. These concerns suggest that the allure of high success rates and groundbreaking discoveries might sometimes overshadow the inherent complexities and limitations of EEG, necessitating a more critical examination of its use in emotional studies.
This article explores the complex aspects of emotion research using EEG. It critically examines the claims of high accuracy in the field and discusses the fundamental nature of emotions. The objective is to offer a comprehensive analysis that encompasses the technical challenges associated with EEG data, the impact of equipment and data source variability on application, and the paradoxes faced in conducting emotion research in controlled settings. The paper aims to contribute to a detailed and nuanced understanding of emotional science exploration.
Methodology
To gain a comprehensive understanding of the challenges and dilemmas in EEG emotion measurement, we embarked on an in-depth literature review in August 2023, focusing our efforts on three pivotal academic databases: Web of Science, Scopus, and ProQuest. These databases were chosen for their extensive coverage in neuroscience and psychological research. Using the keyword combination of “EEG” and “emotion”, we aimed to collate core research papers central to our topic. This initial search spanning the years 2022 to 2023 yielded a significant count of 3,741 articles. Recognizing the voluminous nature of our initial pool and the need for stringent quality control, we employed the MMAT (Mixed Methods Appraisal Tool) (Hong et al., 2018) to assess and sieve these articles. The MMAT, renowned for its capability to critically assess mixed methods research, served as a foundational filter in our process.
The subsequent phase of our methodology centered on refining our selection based on the robustness of experimental results. We pivoted our attention toward articles that reported an emotion recognition accuracy rate exceeding 90%, a threshold we deemed crucial as it represents a benchmark in EEG-based emotion recognition that suggests robust and replicable findings. This rigorous criterion distilled our initial list down to 22 quintessential articles poised for an in-depth analysis. To further elevate the quality of our review and to draw comprehensive insights, we employed the bibliometrix packages in R studio for data visualization analysis (Aria and Cuccurullo, 2017). This sophisticated analytical tool enabled us to intricately map co-citation relationships, track research keywords, and intuitively navigate the intricate web of research relationships and emergent themes in the domain, all of which is exemplified in Figure 1.
The reality behind high EEG accuracy rates
Recent EEG-based emotion recognition research claims strikingly high accuracy rates of 90–99% (see Table 1). These figures are particularly remarkable when juxtaposed with the 75–80% accuracy rate achieved by current facial emotion recognition technology (Naga et al., 2023). However, the majority of these EEG studies employ binary (Valence, Arousal) or ternary (Positive, Negative, Neutral) emotional models, which inherently simplify the classification task and can lead to inflated accuracy rates. For instance, Zhong et al. (2019) reported a drop in accuracy from 79.14–91.67% to 68.26–80.14% when their binary model of emotion was expanded to include four dimensions. This is particularly telling considering that standard models of basic emotions are often based on six or seven categories, suggesting that a more nuanced approach might yield significantly different accuracy rates.
In discussions of EEG-based emotion recognition accuracy, it becomes evident that models relying on subject-dependent data often report higher accuracy rates. This is because when a model is trained and tested on data from the same participant, it can more effectively capture specific characteristics of that individual, leading to increased accuracy. For instance, the BiDANN model by Li et al. (2022) demonstrated an accuracy range of 86.15–96.89% on the SEED dataset (Zheng and Lu, 2015) in a subject-dependent scenario, while the accuracy dropped to 74.52–91.04% in subject-independent settings. Similarly, the BiDCNN model by Huang et al. (2021) achieved emotion recognition accuracy rates of 94.38–94.72% using subject-dependent strategies, which significantly declined to 68.14–63.94% under subject-independent conditions. This phenomenon highlights the substantial reduction in overall accuracy rates when shifting from subject-dependent to subject-independent validation strategies. Independent subject validation poses a greater challenge as it requires the model to generalize to data from unseen participants. This generalization often reveals limitations of the model, as it must capture broader and more universal emotional features rather than merely adapting to specific training data. Therefore, when assessing the true performance of EEG emotion recognition technologies, special attention should be paid to the accuracy rates reported under subject-independent conditions.
The choice of k-value in cross-validation significantly impacts the model's generalization ability and accuracy. Although most studies employ 10- or 15-fold cross-validation, which helps to reduce the risk of overfitting to specific data splits, there are situations, such as the use of 5-fold cross-validation, especially when data splits are not sufficiently random or the dataset size is small, where the model may still face the risks of overfitting and artificially inflated accuracy rates. Moreover, the use of windowing and segmentation as data augmentation strategies warrants attention in these studies. When data is augmented through windowing and segmentation, the model may become overly proficient at recognizing repetitive or similar data segments, leading to seemingly improved performance during testing. While this data augmentation strategy can enhance model performance in some cases, it may also lead to poor generalization on new, unseen data, thereby significantly compromising accuracy in practical applications.
Therefore, while it seems that EEG emotion analysis has made significant progress due to deep learning, the high accuracy rates often overshadow the selective presentation of results by researchers under publication pressures.
The challenge of application
In the field of EEG-based emotion recognition, although there are technical challenges such as physiological disturbances to brain signals from sources like body movement, muscle electrical interference, eye movements, and heartbeats (Fatourechi et al., 2004), the greater challenge lies in applying these techniques to practical scenarios. Unlike facial emotion recognition, which primarily uses images, the data sources for EEG-based emotion detection are more intricate and varied. A notable observation is that most EEG-based emotion recognition studies rely on pre-existing datasets [such as SEED, DEAP (Koelstra et al., 2011), DREAMER (Katsigiannis and Ramzan, 2017)] rather than collecting data independently (Kumari et al., 2022). Furthermore, there seems to be a scarcity of research in applying self-designed and trained models to real-world emotion measurement experiments. Several reasons contribute to this scenario:
• Lack of standardization in EEG devices: While image-based facial recognition technologies can often normalize variations through advanced processing, EEG data faces greater challenges in standardization due to the lack of consistent device standards. EEG products can be classified based on the number of electrodes: more than 32 electrodes (high-density systems), 8–32 electrodes (medium-density systems), and fewer than eight electrodes (low-density systems). Additionally, the type of electrode—be it wet, dry, or specialized—adds to the complexity (Hernandez-Pavon et al., 2023). While wet electrodes using conductive gels or saline might offer better signal quality, they can be more cumbersome and intrusive, potentially affecting the emotional state of the participant. On the other hand, consumer-grade EEG devices, typically designed more for entertainment than research, often use fewer electrodes (1–16) and may not be suitable for rigorous scientific study.
• Challenges in data collection: Collecting EEG data presents a unique set of challenges, primarily due to the need for specialized knowledge and high-quality capture techniques (Boudewyn et al., 2023). Unlike more accessible fields like video tracking or facial emotion recognition, EEG data collection demands specialized expertise and sophisticated equipment. Here, “regular researchers” refers to those who might not specialize in neuroscientific methods or lack access to advanced EEG data collection facilities. These general researchers, typically found in broader disciplines without a focus on neurology or bioengineering, often face challenges in acquiring and processing EEG data due to limited facilities or expertise. Such limitations in data gathering can impede the development of robust machine learning models, affecting their generalizability and effectiveness.
Given these challenges, it's clear that emotion recognition based on EEG is not just a matter of recognition technique but also involves the availability of data and equipment. The fragmented nature of data sources, combined with the diversity of EEG devices and the complexities of data collection, poses significant obstacles to the advancement and generalization of EEG-based emotion recognition research.
The paradox of emotional research
The disconnect between laboratory environments and genuine emotional experiences is a key issue in emotion studies. Laboratory settings may be too rigid and artificial, failing to reflect real emotional experiences in daily life (Russell, 1994). Limitations of laboratory-based emotional research include the simulation of real-life scenarios and individual differences in emotional experience (Rottenberg and Gross, 2007). We need to balance laboratory setups and real-life emotional experiences (Healey et al., 2010).
Experimental equipment may interfere with participants' natural emotional experiences. Advanced devices like virtual reality present potential value and challenges in emotional research, including potential interference with participant emotions (Bohil et al., 2011). Especially with EEG devices, the requirement for participants to remain still to avoid movement interference may restrict natural emotional expression (Wilhelm and Grossman, 2010).
The mismatch between the dynamism of emotions and the fixed nature of experimental design may limit research accuracy and relevance. Emotions' fluidity and change have been emphasized as core parts of research (Kuppens et al., 2010). Time scale issues in emotional research, especially in capturing dynamic emotional changes, have been explored (Davidson, 2010). The complexity of emotional research is revealed with the tension between dynamism and stillness, suggesting EEG devices might be unsuitable as emotional measuring tools.
Discussion
The use of EEG in emotion recognition research presents a host of possibilities and pitfalls. This confluence of neuroscience and technological integration represents a frontier in our understanding of human emotional processes, yet as our review suggests, researchers need to tread cautiously. The lure of high accuracy rates, the challenges of generalization, and the inherent paradoxes of emotional research have proven to be formidable challenges.
There's no doubt that the temporal resolution of EEG offers a granular view of neural processes as emotions unfold (MacNamara et al., 2022). However, as our review suggests, the translation from granular neural data to accurate emotional recognition is far from straightforward. A primary concern is the reported accuracy rates, which in some instances seem too good to be true. These astonishingly high rates challenge our understanding of the complexities of both EEG data and the nuanced nature of human emotion. High accuracy rates, when not critically examined, can lead to a false sense of progress in the field.
The challenge of application further exacerbates these concerns. Emotion, by its nature, is influenced by myriad factors, from the immediate environment to an individual's past experiences (Jani and Han, 2015). Thus, relying on standard datasets, while practical, may not capture the full gamut of human emotion. The diversity in EEG devices, data sources, and collection methodologies can introduce variability that complicates generalization. Additionally, as mentioned, the lack of standardization in EEG devices makes the reproducibility of research findings a formidable challenge (Keil et al., 2014).
Furthermore, the paradoxical nature of emotional research, especially the gulf between lab conditions and real-world emotional experiences, cannot be understated. Emotions are not static entities to be captured in controlled environments but are dynamically intertwined with our ever-changing contexts (Barrett et al., 2006; Wilhelm and Grossman, 2010). The very act of measuring emotion in a lab setting might alter the nature of the emotion itself, akin to the observer effect in physics (Russell, 2003). The complexities associated with dynamic emotional changes underline the importance of methodological flexibility and the need for tools that can capture the rich tapestry of human emotions.
Conclusion
In this comprehensive review, we have shed light on the intricacies of EEG-based emotion recognition. While the allure of high accuracy rates in EEG emotion research paints an optimistic picture, it is essential to approach these claims with cautious optimism. The challenges in generalizing these methods, the inherent discrepancies in laboratory environments, and the dynamic nature of emotions are significant cautions for those eager to adopt EEG in emotion recognition. We must recognize that although EEG holds great potential as a tool in emotion research, its ability to fully understand human emotions is not yet perfect and requires further development and refinement.
Limitations, implications, and further directions of research
In our comprehensive review of EEG-based emotion recognition, we identified several key limitations and implications for the broader scientific community. Our review, concentrated on literature from 2018 to 2022, may have missed broader developments in the field. Our focus on high-accuracy studies could also have overshadowed valuable insights from moderate or lower-accuracy research. Further, relying on databases like Web of Science, Scopus, and ProQuest could have led us to overlook significant work housed in niche repositories. The highlighted accuracies in studies emphasize the need for a critical evaluation regarding methodological rigor and the clear divergence between laboratory and real-world settings, calling for more standardized and applicable tools and methodologies. Moving forward, there's an evident need to expand the scope of the literature review, hone in on the most effective methodologies, integrate multiple biometric tools for a more comprehensive emotion assessment, prioritize studies with higher ecological validity, and deeply investigate the influence of diverse emotion theories on EEG data interpretation. As we delve deeper into this domain, embracing these considerations could lead to a more nuanced understanding of the human emotional landscape through EEG research.
Author contributions
ZZ: Conceptualization, Methodology, Visualization, Writing – original draft. JF: Conceptualization, Data curation, Writing – review & editing. LG: Conceptualization, Methodology, Writing – review & editing.
Funding
The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Aria, M., and Cuccurullo, C. (2017). bibliometrix: an r-tool for comprehensive science mapping analysis. J. Informetr. 11, 959–975. doi: 10.1016/j.joi.2017.08.007
Barrett, L. F., Mesquita, B., Ochsner, K. N., and Gross, J. J. (2006). The experience of emotion. Annu. Rev. Psychol. 58, 373–403. doi: 10.1146/annurev.psych.58.110405.085709
Bohil, C. J., Alicea, B., and Biocca, F. A. (2011). Virtual reality in neuroscience research and therapy. Nat. Rev. Neurosci. 12, 752–762. doi: 10.1038/nrn3122
Boudewyn, M. A., Erickson, M. A., Winsler, K., Ragland, J. D., Yonelinas, A., Frank, M., et al. (2023). Managing eeg studies: how to prepare and what to do once data collection has begun. Psychophysiology 60, e14365. doi: 10.1111/psyp.14365
Brunner-Sperdin, A., Peters, M., and Strobl, A. (2012). It is all about the emotional state: managing tourists' experiences. Int. J. Hosp. Manag. 31, 23–30. doi: 10.1016/j.ijhm.2011.03.004
Chen, J., Hu, B., Moore, P., Zhang, X., and Ma, X. (2015). Electroencephalogram-based emotion assessment system using ontology and data mining techniques. Appl. Soft Comput. 30, 663–674. doi: 10.1016/j.asoc.2015.01.007
Cheng, J., Chen, M., Li, C., Liu, Y., Song, R., Liu, A., et al. (2021). Emotion recognition from multi-channel eeg via deep forest. IEEE J. Biomed. Health Inf. 25, 453–464. doi: 10.1109/JBHI.2020.2995767
Cho, J., and Hwang, H. (2020). Spatio-temporal representation of an electoencephalogram for emotion recognition using a three-dimensional convolutional neural network. Sensors 20, 3491. doi: 10.3390/s20123491
Christensen, L. R., and Abdullah, M. A. (2018). “EEG emotion detection review,” in 2018 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB) (St. Louis, MO), 1–7. doi: 10.1109/CIBCB.2018.8404976
Cohen, S., Zantvoord, J., Wezenberg, B., Daams, J., Bockting, C., Denys, D., et al. (2023). Electroencephalography for predicting antidepressant treatment success: a systematic review and meta-analysis. J. Affect. Disord. 321, 201–207. doi: 10.1016/j.jad.2022.10.042
Cui, H., Liu, A., Zhang, X., Chen, X., Wang, K., and Chen, X. (2020). Eeg-based emotion recognition using an end-to-end regional-asymmetric convolutional neural network. Knowl. Based Syst. 205, 106243. doi: 10.1016/j.knosys.2020.106243
Dadebayev, D., Goh, W. W., and Tan, E. X. (2021). Eeg-based emotion recognition: review of commercial eeg devices and machine learning techniques. J. King Saud Univ. Comput. Inf. Sci. 34, 4385–4401. doi: 10.1016/j.jksuci.2021.03.009
Davidson, R. J. (2010). Affective style and affective disorders: perspectives from affective neuroscience. APA PsycNet 12, 307–330. doi: 10.1080/026999398379628
Fatourechi, M., Mason, S. G., Birch, G. E., and Ward, R. K. (2004). “A wavelet-based approach for the extraction of event related potentials from eeg,” in 2004 IEEE International Conference on Acoustics, Speech, and Signal Processing (Montreal, QC: IEEE), ii–737. doi: 10.1109/ICASSP.2004.1326363
Fdez, J., Guttenberg, N., Witkowski, O., and Pasquali, A. (2021). Cross-subject eeg-based emotion recognition through neural networks with stratified normalization. Front. Neurosci. 15, 626277. doi: 10.3389/fnins.2021.626277
He, H., Tan, Y., Ying, J., and Zhang, W. (2020). Strengthen eeg-based emotion recognition using firefly integrated optimization algorithm. Appl. Soft Comput. 94, 106426. doi: 10.1016/j.asoc.2020.106426
Healey, J., Nachman, L., Subramanian, S., Shahabdeen, J., and Morris, M. (2010). “Out of the lab and into the fray: towards modeling emotion in everyday life,” in Pervasive Computing. Pervasive 2010. Lecture Notes in Computer Science, Vol. 6030, eds P. Floréen, A. Krüger, and M. Spasojevic (Berlin; Heidelberg: Springer). doi: 10.1007/978-3-642-12654-3_10
Hernandez-Pavon, J. C., Veniero, D., Bergmann, T. O., Belardinelli, P., Bortoletto, M., Casarotto, S., et al. (2023). Tms combined with eeg: Recommendations and open issues for data collection and analysis. Brain Stimul. 16, 567–593. doi: 10.1016/j.brs.2023.02.009
Hong, Q. N., Fàbregues, S., Bartlett, G., Boardman, F., Cargo, M., Dagenais, P., et al. (2018). The mixed methods appraisal tool (mmat) version 2018 for information professionals and researchers. Educ. Inf. 34, 285–291. doi: 10.3233/EFI-180221
Huang, D., Chen, S., Liu, C., Zheng, L., Tian, Z., and Jiang, D. (2021). Differences first in asymmetric brain: a bi-hemisphere discrepancy convolutional neural network for eeg emotion recognition. Neurocomputing 448, 140–151. doi: 10.1016/j.neucom.2021.03.105
Jani, D., and Han, H. (2015). Influence of environmental stimuli on hotel customer emotional loyalty response: Testing the moderating effect of the big five personality factors. Int. J. Hosp. Manag. 44, 48–57. doi: 10.1016/j.ijhm.2014.10.006
Katsigiannis, S., and Ramzan, N. (2017). Dreamer: a database for emotion recognition through eeg and ecg signals from wireless low-cost off-the-shelf devices. IEEE J. Biomed. Health Inf. 22, 98–107. doi: 10.1109/JBHI.2017.2688239
Keil, A., Debener, S., Gratton, G., Junghöfer, M., Kappenman, E. S., Luck, S. J., et al. (2014). Committee report: Publication guidelines and recommendations for studies using electroencephalography and magnetoencephalography. Psychophysiology 51, 1–21. doi: 10.1111/psyp.12147
Koelstra, S., Muhl, C., Soleymani, M., Lee, J.-S., Yazdani, A., Ebrahimi, T., et al. (2011). Deap: a database for emotion analysis; using physiological signals. IEEE Transact. Affect. Comp. 3, 18–31. doi: 10.1109/T-AFFC.2011.15
Kumari, N., Anwar, S., and Bhattacharjee, V. (2022). Time series-dependent feature of eeg signals for improved visually evoked emotion classification using emotioncapsnet. Neural Comp. Appl. 34, 13291–13303. doi: 10.1007/s00521-022-06942-x
Kuppens, P., Oravecz, Z., and Tuerlinckx, F. (2010). Feelings change: accounting for individual differences in the temporal dynamics of affect. J. Pers. Soc. Psychol. 99, 1042–1060. doi: 10.1037/a0020962
Lehman, B. J., Cane, A. C., Tallon, S. J., and Smith, S. F. (2015). Physiological and emotional responses to subjective social evaluative threat in daily life. Anxiety Stress Coping 28, 321–339. doi: 10.1080/10615806.2014.968563
Li, Y., Zheng, W., Cui, Z., Zhang, T., and Zong, Y. (2018). “A novel neural network model based on cerebral hemispheric asymmetry for eeg emotion recognition,” in Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI-18 (Stockholm: International Joint Conferences on Artificial Intelligence Organization), 1561–1567. doi: 10.24963/ijcai.2018/216
Li, Y., Zheng, W., Wang, L., Zong, Y., and Cui, Z. (2022). From regional to global brain: a novel hierarchical spatial-temporal neural network model for eeg emotion recognition. IEEE Transact. Affect. Comp. 13, 568–578. doi: 10.1109/TAFFC.2019.2922912
Liu, Y., Ding, Y., Li, C., Cheng, J., Song, R., Wan, F., et al. (2020). Multi-channel eeg-based emotion recognition via a multi-level features guided capsule network. Comput. Biol. Med. 123, 103927. doi: 10.1016/j.compbiomed.2020.103927
Luo, Y., Zhu, L. Z., Wan, Z. Y., and Lu, B. L. (2020). Data augmentation for enhancing eeg-based emotion recognition with deep generative models. J. Neural Eng. 17, 056021. doi: 10.1088/1741-2552/abb580
Ma, J., Tang, H., Zheng, W.-L., and Lu, B.-L. (2019). “Emotion recognition using multimodal residual LSTM network,” in Proceedings of the 27th ACM International Conference on Multimedia (MM '19) (Nbookew York, NY: Association for Computing Machinery), 176–183. doi: 10.1145/3343031.3350871
MacNamara, A., Joyner, K., and Klawohn, J. (2022). Event-related potential studies of emotion regulation: a review of recent progress and future directions. Int. J. Psychophysiol. 176, 73–88. doi: 10.1016/j.ijpsycho.2022.03.008
Naga, P., Marri, S. D., and Borreo, R. (2023). Facial emotion recognition methods, datasets and technologies: a literature survey. Materials Today 80, 2824–2828. doi: 10.1016/j.matpr.2021.07.046
Nummenmaa, L., Glerean, E., Viinikainen, M., Jääskeläinen, I. P., Hari, R., and Sams, M. (2012). Emotions promote social interaction by synchronizing brain activity across individuals. Proc. Nat. Acad. Sci. U. S. A. 109, 9599–9604. doi: 10.1073/pnas.1206095109
Rottenberg, J., and Gross, J. J. (2007). Emotion and emotion regulation: a map for psychotherapy researchers. Clin. Psychol.: Sci. Pract. 14, 323–328. doi: 10.1111/j.1468-2850.2007.00093.x
Russell, J. A. (1994). Is there universal recognition of emotion from facial expression? A review of the cross-cultural studies. Psychol. Bull. 115, 102–141. doi: 10.1037/0033-2909.115.1.102
Russell, J. A. (2003). Core affect and the psychological construction of emotion. Psychol. Rev. 110, 145–172. doi: 10.1037/0033-295X.110.1.145
Song, T., Zheng, W., Lu, C., Zong, Y., Zhang, X., and Cui, Z. (2019). Mped: a multi-modal physiological emotion database for discrete emotion recognition. IEEE Access 7, 12177–12191. doi: 10.1109/ACCESS.2019.2891579
Suhaimi, N. S., Mountstephens, J., and Teo, J. (2020). Eeg-based emotion recognition: a state-of-the-art review of current trends and opportunities. Comput. Intell. Neurosci. 2020, 8875426. doi: 10.1155/2020/8875426
Tao, W., Li, C., Song, R., Cheng, J., Liu, Y., Wan, F., et al. (2023). Eeg-based emotion recognition via channel-wise attention and self attention. IEEE Transact. Affect. Comp. 14, 382–393. doi: 10.1109/TAFFC.2020.3025777
Wang, F., Wu, S., Zhang, W., Xu, Z., Zhang, Y., Wu, C., et al. (2020). Emotion recognition with convolutional neural network and eeg-based efdms. Neuropsychologia 146, 107506. doi: 10.1016/j.neuropsychologia.2020.107506
Wilhelm, F. H., and Grossman, P. (2010). Emotions beyond the laboratory: theoretical fundaments, study design, and analytic strategies for advanced ambulatory assessment. Biol. Psychol. 84, 552–569. doi: 10.1016/j.biopsycho.2010.01.017
Yang, Y., Wu, Q., Qiu, M., Wang, Y., and Chen, X. (2018). “Emotion recognition from multi-channel eeg through parallel convolutional recurrent neural network,” in 2018 International Joint Conference on Neural Networks (IJCNN) (Rio de Janeiro: IEEE). doi: 10.1109/IJCNN.2018.8489331
Yin, Y., Zheng, X., Hu, B., Zhang, Y., and Cui, X. (2021). Eeg emotion recognition using fusion model of graph convolutional neural networks and lstm. Appl. Soft Comput. 100, 106954. doi: 10.1016/j.asoc.2020.106954
Zhang, G., and Etemad, A. (2023). Distilling eeg representations via capsules for affective computing. Pattern Recognit. Lett. 171, 99–105. doi: 10.1016/j.patrec.2023.05.011
Zhang, T., Cui, Z., Xu, C., Zheng, W., and Yang, J. (2020). Variational pathway reasoning for eeg emotion recognition. Proc. AAAI Conf. Artif. Intell. 34, 2709–2716. doi: 10.1609/aaai.v34i03.5657
Zhang, T., Zheng, W., Cui, Z., Zong, Y., and Li, Y. (2019). Spatial-temporal recurrent neural network for emotion recognition. IEEE Trans. Cybern. 49, 939–947. doi: 10.1109/TCYB.2017.2788081
Zhang, Z., Mir, J. M. F., and Mateu, L. G. (2022). The effects of white versus coloured light in waiting rooms on people's emotions. Buildings 12, 1356. doi: 10.3390/buildings12091356
Zheng, W.-L., and Lu, B.-L. (2015). Investigating critical frequency bands and channels for eeg-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 7, 162–175. doi: 10.1109/TAMD.2015.2431497
Zhong, P., Wang, D., and Miao, C. (2019). Eeg-based emotion recognition using regularized graph neural networks. IEEE Transact. Affect. Comp. 13, 1290–1301. doi: 10.1109/TAFFC.2020.2994159
Keywords: EEG, emotional measurement, challenges, emotional dynamics, recognition
Citation: Zhang Z, Fort JM and Giménez Mateu L (2024) Mini review: Challenges in EEG emotion recognition. Front. Psychol. 14:1289816. doi: 10.3389/fpsyg.2023.1289816
Received: 07 September 2023; Accepted: 11 December 2023;
Published: 04 January 2024.
Edited by:
Frank A. Russo, Toronto Metropolitan University, CanadaReviewed by:
Monica Leba, Universitatea Din Petrosani, RomaniaFrancesca Gasparini, University of Milano-Bicocca, Italy
Copyright © 2024 Zhang, Fort and Giménez Mateu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Zhihui Zhang, emhpaHVpLnpoYW5nJiN4MDAwNDA7dXBjLmVkdQ==