- 1School of Art, Zhejiang International Studies University, Hangzhou, China
- 2School of Education, Hangzhou Normal University, Hangzhou, China
- 3School of Arts and Media, Beijing Normal University, Beijing, China
- 4College of Science, Zhejiang University of Technology, Hangzhou, China
Music is an archaic form of emotional expression and arousal that can induce strong emotional experiences in listeners, which has important research and practical value in related fields such as emotion regulation. Among the various emotion recognition methods, the music-evoked emotion recognition method utilizing EEG signals provides real-time and direct brain response data, playing a crucial role in elucidating the neural mechanisms underlying music-induced emotions. Artificial intelligence technology has greatly facilitated the research on the recognition of music-evoked EEG emotions. AI algorithms have ushered in a new era for the extraction of characteristic frequency signals and the identification of novel feature signals. The robust computational capabilities of AI have provided fresh perspectives for the development of innovative quantitative models of emotions, tailored to various emotion recognition paradigms. The discourse surrounding AI algorithms in the context of emotional classification models is gaining momentum, with their applications in music therapy, neuroscience, and social activities increasingly coming under the spotlight. Through an in-depth analysis of the complete process of emotion recognition induced by music through electroencephalography (EEG) signals, we have systematically elucidated the influence of AI on pertinent research issues. This analysis offers a trove of innovative approaches that could pave the way for future research endeavors.
1 Introduction
Music serves as a unique medium for people to express their emotions and also can arouse strong emotional responses. Previous studies have shown that the emotional changes induced by appropriate music can relieve listeners’ mental stress (Nawaz et al., 2019; Colin et al., 2023), promote emotional expression ability (Palazzi et al., 2021; Micallef Grimaud and Eerola, 2022; Zhang et al., 2022), improve learning ability (Bergee and Weingarten, 2021; Luo et al., 2023), and so on. Moreover, it also can be applied in the regulation of mood-related disorders such as autism (Carpente et al., 2022; Geretsegger et al., 2022), depression (Geipel et al., 2022; Hartmann et al., 2023), and anxiety (Contreras-Molina et al., 2021; Lu et al., 2021). With the extensive applications of music-induced emotions in medical (Liang et al., 2021), neuroscience (Thaut et al., 2021; Fedotchev et al., 2022), and music retrieval fields (Gomez-Canon et al., 2021), the study of music-induced emotion recognition has received much attention in recent years.
Empirical research on the effects of music on emotions has been discussed for more than three millennia (Perlovsky, 2012), while modern evidence-based work on the effects of music on emotions has its roots in the early 20th century (Humphreys, 1998). Western psychologists and musicians primarily conducted pioneering empirical research on music-induced emotions. A representative example is the experimental research conducted by the American psychologist and music educator Carl Emil Seashore on the emotional expression of music and the emotional impact of music on the listener, combining experiments and psychological tests and proposing the “theory of musical expression,” which emphasizes how elements such as melodies, rhythms, and harmonies of music affect people’s emotional experiences, and lays the foundation for the subsequent development of related work (Metfessel, 1950). With the development of psychology neuroscience and other fields, people gradually realized that the study of music’s induction of emotions also requires an understanding of auditory perception, emotion discrimination, and neural mechanism, which is an interdisciplinary research work (Cui et al., 2022; Ryczkowska, 2022). Musicologists have mainly studied the influence of music on emotion induction from the perspective of different musical features of music (Panda et al., 2023), including analyzing the influence of music on the listener’s emotion from the perspective of musical elements (Ruth and Schramm, 2021), quantitatively analyzing the emotional features of music to find the relationship between the features and emotion (Salakka et al., 2021), developing emotion recognition algorithms based on musical features (Pandey and Seeja, 2022), and exploring cross-cultural emotional understanding of and response to specific musical features (Wang et al., 2022). These studies have made it possible to help people get a better understanding of the relationship between musical features and emotions, and provide theoretical support and practical guidance for the fields of music psychology, music therapy, and creativity, but researchers have also put forward different viewpoints on the individual differences in musical emotional responses and on the objective evaluative validity of emotions as in Figure 1.
With the development of brain science and technology, researchers have found that signals generated by the central nervous system, such as electroencephalography (EEG), magnetoencephalography (MEG), and functional magnetic resonance imaging (MRI) are more objective and reliable in the field of emotion research (Alarcao and Fonseca, 2019; Egger et al., 2019; Saganowski et al., 2023). Among various central nervous system signals, the monitoring of emotions using EEG signals is characterized by the convenience of noninvasive measurements, real-time measurements, and good objectivity. Research on emotion recognition based on EEG signals has been widely used in many disciplines in recent years and has received extensive attention from researchers as in Figure 1. Artificial intelligence (AI) techniques that integrate EEG signals for identifying emotions elicited by music leverage AI’s robust capabilities in data analytics, pattern recognition, and learning, alongside the distinctive benefits of EEG for real-time, non-invasive monitoring of brain activity. AI-enabled EEG recognition of music-induced emotions can accurately and in real-time identify emotions, which has broad applications in many areas including music therapy, education, entertainment, and so on.
How to accurately identify music-induced emotions has always been a difficult research problem due to the subjectivity, abstractness, and individualized differences of music-induced emotions. Researchers have explored a variety of physiological signals to carry out emotion recognition studies, in which using the signal characteristics of facial expressions, researchers have classified emotions including fear, sadness, disgust, surprise, and joy, and the accuracy of obtaining emotion discrimination can be as high as 81% or more, but there are inconsistencies between different cultures in the understanding of facial expressions and the way of expression of facial expressions, which affect the generalizability of the results of the study (Tcherkassof and Dupré, 2021; Witkower et al., 2021). Physiological parameters such as galvanic skin response, heart rate, temperature, blood pressure, and respiration rate have also been utilized for emotion recognition, but these methods are relatively inaccurate for emotion discrimination and highly influenced by other factors (Egger et al., 2019; Saganowski et al., 2023).
In this study, the research methods, processes, and characteristics of EEG in music-induced emotion recognition have been analyzed. The potential future development directions of music-induced emotion based on EEG also have been discussed, which can promote the development of fundamental and application research on music-induced emotion.
2 EEG signal and emotions
Measurement of EEG signals is capable of non-invasive, continuous recording of brain activity with a temporal resolution of a few milliseconds. Based on the characteristic waveform signals from different brain regions, EEG signals are widely used in cognitive neuroscience to research emotion regulation and processing, and the results of the related studies provide an important reference for further research on music-induced emotion recognition (Apicella et al., 2022b; Pandey and Seeja, 2022).
2.1 EEG signals and acquisition method
The activity of the central nervous system of the brain is closely related to human emotions, mainly realized through electrical communication between neurons (Ahmad et al., 2022). When neurons are stimulated, the membrane potential will rise and fall to form weak electrical pulse signals and emotional changes can be monitored by recording and analyzing EEG signals. EEG signals have the characteristics of small amplitude (10–100 μV), many interference sources, and high uncertainty. To analyze the feature information of EEG, as shown in Table 1, EEG signals are generally classified into δ-band (1–4 Hz), θ-band (4–8 Hz), α-band (8–13 Hz), β-1-band (13–22 Hz), β-2-band (22–30 Hz) and γ-band (30–64 Hz) (Alarcao and Fonseca, 2019), and the frequency bands of the bands are divided into slightly different bands by different researchers.
Research has shown that the five bands of EEG signals mentioned above are directly or indirectly related to human emotions. While early studies suggested that the δ-band was not connected to people’s emotions, recent research has found that δ wave is closely associated with the emotional state of individuals following emotional regulation and holds promise for use in areas such as music therapy for emotion regulation (Lapomarda et al., 2022).
The acquisition method of EEG signals is mainly categorized into invasive and non-invasive techniques. Invasive measurements require surgical implantation of electrodes to obtain clearer EEG signals, but this method is traumatized to the human body and difficult to widely apply, which is mainly used in clinical medical treatment. Non-invasive is to fit the electrodes to the surface of the head to collect brain signals.
Previous research has demonstrated the significance of the limbic system (as in Figure 2A) in regulating human emotions, making it a pivotal area of interest in the field of emotion research (Rolls, 2015). To obtain more comprehensive brain signals, the internationally recognized 10/20 system, shown in Figure 2B, is generally used in the arrangement of electrodes, i.e., the actual distance of adjacent electrodes is 10% or 20% of the distance of the brain skull (Silverman, 1963). In the field of emotion recognition, multi-channel EEG acquisition is commonly utilized, featuring electrode channels ranging from 36 to 64, and a sampling frequency of 500 or 1,000 Hz (Wu et al., 2024). Traditional EEG acquisition system devices are often cumbersome and expensive, which hinders their widespread adoption and use. With the advancement of open source technologies like OpenBCI as in Figure 2C and other EEG acquisition devices, more affordable, user-friendly, and portable options have emerged. Recently, these devices have become increasingly popular in EEG emotion recognition research (Aldridge et al., 2019). To investigate the mechanisms of timely emotional response, both the stimulus source and the acquisition system are typically equipped with time-synchronization devices (Pei et al., 2024).
Figure 2. (A) The limbic system for emotion, (B) the international 10/20 system with nine cortical regions labeled with different colors, (C) structure of a typical EEG system from OpenBCI.
In music emotion recognition research, non-invasive acquisition schemes are commonly employed. In recent years, wireless wearable non-invasive EEG measurement devices have greatly facilitated EEG-based emotion recognition research (Apicella et al., 2022a; Saganowski et al., 2023). The emergence of these novel EEG acquisition protocols has significant implications for expanding the scope of EEG emotional applications.
2.2 EEG signal bands corresponding to different emotions
The brain exhibits diverse EEG response patterns for different emotions, and establishing the relationship between EEG signal bands and various emotional states is a crucial foundation for developing effective classification and recognition models. This correspondence serves as one of the key scientific challenges in the domain of artificial intelligence-based recognition of music-induced emotions. Around 1980, researchers found that EEG’s characteristic signals correlate with human emotional states (Davidson and Fox, 1982), as in Table 1. Subsequently, researchers have investigated the relationship between distinct brainwave frequency bands and diverse emotional states. In 2001, Louis A. Schmidt et al. presented that emotions within valence can be distinguished by evaluating the asymmetry and overall power of α-band (8–13 Hz) from frontal brain EEG signals (Schmidt and Trainor, 2001). In 2007, Daniela Sammler et al. conducted a systematic analysis of the correlation between various EEG frequency bands and emotions. Their findings revealed that θ-band (4–8 Hz) power in the prefrontal lobe is more prominent during happy music stimulation, indicating its significance in emotional processing (Sammler et al., 2007). With the continuous advancement of EEG analysis technology, it has become increasingly apparent that the intricate nature of the brain’s emotional processes makes it challenging to establish precise correlations between different emotions and signals derived from a single brain region or waveform. Certainly, in some specific scenarios, researchers continue to explore and identify the most prominent EEG frequency bands to simplify the challenges associated with emotion recognition.
The application of artificial intelligence in the emotional recognition of music-induced electroencephalography (EEG) holds significant value in two primary aspects. On one hand, the utilization of AI algorithms assists researchers in discerning and selecting the appropriate frequency bands amidst a multitude of options. On the other hand, the deployment of AI algorithms facilitates the exploration of additional effective frequency bands, enhancing the depth and breadth of research in this domain. With the development of artificial intelligence and deep learning technologies, emotion recognition by utilizing various frequency band features from different brain regions has emerged as a prominent and contemporary approach (Pei et al., 2024; Xu et al., 2024). Machine learning based Support Vector Machine (SVM) (Bagherzadeh et al., 2023), Naïve Bayes (NB) (Oktavia et al., 2019), and K Nearest Neighbors (KNN) (Sari et al., 2023) classifier methods have been applied in this field. Deep learning based classification methods such as Convolutional Neural Networks (CNN) (Yang et al., 2019), Recurrent Neural Networks (RNN) (Zhong et al., 2023), Long-Short-Term Memory (LSTM) (Du et al., 2022), and other classification methods have also been used in EEG recognition studies.
3 Preprocessing and feature extraction of EEG signals
Extracting effective emotional state information from raw EEG signals is a highly challenging task, given that the signal is a multi-frequency non-stationary signal. EEG preprocessing and feature extraction are essential steps in the recognition algorithms for effective analysis and interpretation of the EEG signals. The purpose of preprocessing EEG signals is to eliminate human motion interference and environmental noise that are unrelated to emotion pattern recognition. This is essential for enhancing the accuracy and robustness of the recognition algorithm.
3.1 EEG preprocessing
Noise removal is a crucial objective of EEG signal preprocessing. EEG signals are often vulnerable to interference from various sources such as environmental electromagnetic signals (~50–60 Hz), eye movements (~4 Hz), electrocardiogram signals (ECG, ~1.2 Hz), and so on.
The removal of these noises can significantly enhance the robustness of the EEG model. Usually, these disturbing signals can be filtered out with band-pass filters, wavelet packet filtering, or independent component analysis (ICA) methods as in Table 2. However, researchers have different opinions regarding the signal filtering methods in preprocessing. Some argue that these methods do not eliminate interfering noise, while others believe that these techniques remove noise at the expense of potentially discarding valuable EEG information. To further improve the denoising performance, the artifact subspace reconstruction (ASR) method can be applied to remove the artificial signals. What’s more, the average value of overall electrodes can be applied to subtract from each channel to reduce the system noise (Katsigiannis and Ramzan, 2018). Compared to classical machine learning algorithms, deep learning classification methods for emotion recognition are less influenced by the effects of preprocessing techniques.
The most popular open source toolbox for EEG preprocessing is EEGLAB running in the MATLAB environment (Martínez-Saez et al., 2024; Pei et al., 2024; Wu et al., 2024). This interactive toolbox can be applied to process continuous and event-related EEG signals. Moreover, the artifacts from eye movements can be removed with the run independent component analysis (RunICA) algorithm incorporated in EEGLAB based on the independent component analysis (ICA) method (Wu et al., 2024). The expansion of artificial intelligence has led to the integration of EEG signal preprocessing algorithms into a growing array of commercial AI development platforms, including Python, Brainstorm, and Curry.
3.2 Time domain feature extraction
Emotional changes in the brain can be influenced by musical stimulation, leading to observable effects on EEG signals. These EEG signals exhibit various time-dependent features, which can be analyzed in the time domain. Time-domain features provide intuitive insights and are relatively easy to obtain. Some categories of time-domain features in EEG analysis include event-related potentials (ERPs), statistical features (such as mean, average, standard deviation, skewness, kurtosis, etc.), rise and fall times, and burst suppression (Stancin et al., 2021; Li et al., 2022b).
Time domain features can intuitively capture changes in brain states following music-induced emotions. The active regions corresponding to these emotions can typically be promptly identified through intuitive brain area distribution maps, as in Figure 3, offering valuable insights for the improvement of recognition algorithms. Time domain features are generally preferred in emotion recognition research.
Figure 3. Topographical maps of EEG signals for different types of music (Xu J. et al., 2023).
Moreover, event-related potentials (ERPs) are specific patterns of EEG activity that are time-locked to particular sensory, cognitive, or motor events (Martins et al., 2022). They reflect the brain’s response to stimuli and provide valuable information about cognitive processes, which is very helpful in studying the dynamic processes of emotion change with music stimuli. Rise and fall times refer to the duration it takes for the EEG signal to rise from its baseline to its peak (rise time) or fall back to the baseline (fall time). These measures provide insights into the speed of neural activation or deactivation. Currently, there is a relatively limited body of research on the speed, duration, and recovery time of human emotions stimulated by music. It is important to dedicate attention to these aspects in future studies to gain a deeper understanding of the relevant phenomena with the time domain feature of rise and fall times.
By examining these time-domain features of EEG signals, researchers can gain a better understanding of the temporal dynamics of brain activity related to emotional responses to music. The deployment of artificial intelligence algorithms enables the real-time identification of emotions induced by music via EEG signals. Making well-informed choices and applying time-domain features effectively is essential for advancing these studies.
3.3 Frequency domain feature extraction
As crucial parameters in EEG emotion recognition algorithms, frequency domain features offer more intricate emotion-related information, including the distribution of energy across different frequency bands. For instance, the energy distribution in high-frequency bands (such as β-band and γ-band waves) tends to increase during pleasurable and exciting emotional states (Li et al., 2018). Analyzing the phase synchronization degree of signals can provide insights into changes in information θ-band wave patterns between brain regions during different emotional states. For example, theta synchronization between the frontal and temporal lobes is associated with pleasant emotions (Ara and Marco Pallarés, 2020). Frequency domain features allow for the analysis of interactions between various brain regions. By calculating correlation features between different brain regions at different frequencies, changes in information exchange patterns between brain regions during different emotional states can be observed (Maffei, 2020). Based on the inter-correlation maps of δ, α and γ-band waves stimulated by six different scenarios, the widest topographical distribution is δ-band, while the narrowest is α-band (Maffei, 2020).
Various techniques are commonly employed for extracting frequency domain features as in Table 3. These include the following methods: Fourier transform, wavelet transform, independent component analysis, and matrix decomposition (Torres et al., 2020; Zhou et al., 2022; Li et al., 2023; Mahmoud et al., 2023). Fourier transform is utilized to convert a time domain signal into a frequency domain signal, providing spectral information such as frequency components and amplitude details (Mahmoud et al., 2023). Frequency domain feature extraction techniques based on the Fourier transform encompass power spectral density (PSD), average power spectral density (APSD), and related features. PSD is usually evaluated within a specific frequency band, considered the most commonly applied feature for classical emotion classifiers (Xu et al., 2024).
Wavelet transform offers a more versatile and multi-scale approach to signal analysis, delivering both frequency and time information (Bagherzadeh et al., 2023). Frequency domain feature extraction methods associated with wavelet transform involve wavelet packet decomposition (WPD), wavelet packet energy features, and similar characteristics. Independent component analysis serves as a signal decomposition method grounded in independence assumptions, yielding independent frequency domain components post-decomposition (Shu et al., 2018). Frequency domain feature extraction techniques stemming from independent component analysis include frequency band energy distribution, phase synchronization degree, and more. Matrix decomposition is an algebraic signal decomposition method that disentangles the original signal into distinct frequency domain components (Hossain et al., 2023). These techniques enable the extraction of diverse frequency domain features such as spectral characteristics, phase synchronization degrees, correlation features, and so forth. In emotion classification applications, a tailored selection and adjustment of methods and feature combinations can be made based on specific requirements.
The capabilities of artificial intelligence algorithms in mining large-scale data sets not only enable the automatic extraction of frequency characteristics from EEG signals but also reveal the underlying connections between frequency domain signals and emotions.
3.4 Time-frequency domain feature extraction
Time-frequency feature extraction methods involve analyzing EEG signal changes in both time and frequency to extract characteristic parameters that capture the dynamic nature of the signal (Bagherzadeh et al., 2023). Common techniques of time-frequency domain features include wavelet transform (Khare and Bajaj, 2021) and short-time Fourier transform (STFT) (Pei et al., 2024). These methods enable the extraction of information across various time scales and frequency ranges, unveiling how signals evolve and frequency as in Figure 4, which also has been applied by our group (Li et al., 2022a).
Figure 4. Time-frequency plots before and after stimulation were used by the authors’ research group.
By extracting time-frequency features, a more comprehensive description of the signal’s dynamic characteristics can be achieved, laying the groundwork for subsequent signal processing and emotion classification analysis.
Time-frequency plots typically encompass a vast array of data points, representing a high-dimensional dataset. The application of artificial intelligence algorithms can automatically discern time-frequency patterns associated with various emotions. This capacity for autonomous learning and data mining enhances the efficacy and reliability of time-frequency plots in the identification of emotions induced by music.
3.5 Other advanced features
The development of new emotion-recognition features has been significantly influenced by researchers’ profound insights into the brain’s response to emotions.
Prior physiological and psychological studies have demonstrated that emotions, being intricate mental states, can be discerned by detecting the status of connections between brain regions. In recent years, scholars have advocated for the establishment of a network of brain regions using phase-locked values and the extraction of features from multiple brain functional connectivity networks through the application of the Hilbert transform. These graph features are then fused to facilitate emotion recognition (Li et al., 2019). Based on this concept, researchers have introduced a novel feature called asPLV (averaged sub-frequency phase locking value), which is derived from the Morlet transform method. This feature effectively mitigates the impact of the brain’s inherent frequency oscillations induced by the cognitive processing of emotional fluctuations, thereby enhancing the accuracy of recognizing mood changes induced by music. The calculation process for asPLV is outlined as in Table 4.
In recent years, scholars have discovered that the spatiotemporal characteristics of EEG play a crucial role in emotion recognition. Many studies have introduced novel spatiotemporal features based on self-attention mechanisms (Zhou and Lian, 2023). As our comprehension of the neural mechanisms underlying emotional responses deepens, these new features are critical for enhancing the accuracy of emotion recognition.
Other than these commonly applied features already discussed, artificial intelligence algorithms excel in processing multidimensional data, enabling the discovery of innovative feature metrics. These algorithms hold great promise in identifying individual-specific traits, and crafting features that are sensitive to the distinctive attributes of each individual.
4 Emotion data source and modeling
Auto-emotion recognition can be realized by integrating various data sources and emotion models. This is important for the development of music-induced emotion recognition and its application areas. In the realm of music-induced emotion recognition, emotional data sources form the foundation for acquiring emotion related in sights, while models serve as the essential tools for processing and analyzing this valuable information (Saganowski et al., 2023; Xu J. et al., 2023).
4.1 Data sources for music-evoked emotion classification
In recent years, in order to promote research on music-induced emotions, a series of databases of music-triggered emotions have been established, with emotion labels provided by psychologists. Although these databases can be used for music-triggered emotion research, they lack a unified criterion. Based on the EEG method of emotion discrimination, researchers also have established emotion databases containing EEG signals. These open source databases are not only important resources for conducting research on music-triggered emotions, but can also be used to evaluate the performance of different EEG algorithms. Table 5 shows some common open source music emotion databases and their characteristics.
The AMG1608 database is a database containing acoustic features extracted from 1,608 music clips of 30s as well as emotion annotations provided by 665 subjects, consisting of 345 females and 320 males. The database used a dimensional emotion model with validity and arousal (VA) as the coordinates in the emotion annotation, and the subjects annotated the emotional state of each music clip. The dataset contains two subsets of emotion annotations from National Taiwan University and Amazon Turkish Robotics and is characterized by a large amount of data capable of being publicly accessible and can be used for music emotion recognition research.
The CAL500 database is composed of 500 Western songs’ clips written by 500 different artists. For the emotion annotation of the music, 174 music-related semantic keywords were used, and at least three subjects annotated keywords for each song. These annotated words were also post-processed algorithmically to constitute a vector of annotated words and weights, ensuring the reliability of the annotation labels. The dataset is able to satisfy the fine granularity and differentiation required in music emotion recognition research.
The DEAM database, which labels musical emotions in terms of valence and arousal (VA) coordinates, has 1,802 songs licensed under the Creative Commons (CC) license. This music library contains categories such as rock, pop, soul, blues, electronic, classical, hip-hop, international, experimental, ethnic, jazz, country, and pop. The emotion annotations for these songs were made by 21 active teams from all over the world, and these annotations were statistically processed to form a database that can be used for music emotion research.
The emoMusic database contains 1,000 audio tracks in MP3 format licensed under the Creative Commons (CC) License in eight different genres: blues, electronica, rock, classical, folk, jazz, country, and pop, with 125 tracks in each genre. The emotion labeling of the music was evaluated using valence and arousal (VA) model, where valence indicates positive and negative emotions and arousal indicates emotional intensity. The database collects time-varying (per second) continuous VA rating data, with each song containing at least 10 thematic annotations. The database can be utilized for the conduct of research related to music emotion annotation and other related studies.
The Emotify database contains 100 pieces of music from each of the four genres of classical, rock, pop, and electronic music randomly selected from a collection of music containing 241 different albums by 140 performers. The database used the Geneva Emotional Music Scale (GEMS), in which subjects labeled the emotions of the music using a Likert scale using a scale of 1–5. The database provides case studies and information on the effects of other factors on evoked emotions (gender, mood, music preference).
The DEAP database is a music emotion recognition database based on an EEG emotion recognition method, which was built together by a consortium of four universities from the UK, the Netherlands, and Sweden, and records EEG and physiological information from 32 subjects who watched a series of forty 1-min music video clips. The database was selected as a semi-automatic stimulus selection method based on emotional labeling is open access to academics and can facilitate research related to emotional stimulation in modern music.
The IADS database is the International Emotionally Digitized Sound Database, which is divided into two distinct phases. The initial Phase I database, established in 1999, contains a modest collection of data that has seen limited use in contemporary times. In contrast, Phase II is an expansive compilation of 167 digitally captured ambient sounds that are frequently encountered in everyday life, such as the joyful laughter of an infant, the rhythmic sounds of cooking, and the dramatic rumble of thunderstorms, with each sound clip precisely lasting 6 s. The collection is meticulously annotated, with each piece of digital audio being evaluated by participants through a self-assessment approach that utilizes the affective dimensions of the Valence-Arousal (VA) model.
At present, these databases of music emotions are mainly based on foreign music libraries, the situation is related to the importance of music in the relevant regions, and the establishment of music databases based on Chinese musical compositions is yet to be carried out. Due to the complexity of the signal measurement and classification of EEG in the early stage, there are fewer studies for EEG music-induced emotion recognition. Enabled by artificial intelligence, EEG-based music emotion recognition can help to expand the establishment of databases as soon as possible, and it can help more researchers to apply the established databases. Currently, besides the above common music-induced emotion databases, many researchers also use their own designed libraries to carry out personalized research in the study of music-induced emotions with EEG. As artificial intelligence technology advances, it facilitates the integration of EEG data with other modalities of data, thereby enriching the dimensions of information within the database. The application of data augmentation techniques helps to enhance the generalization capability of models built from the database. With the progression of research in EEG-induced emotion recognition, artificial intelligence can also assist in the automatic consolidation and updating of databases, providing technical support for the construction of more comprehensive, accurate, and holistic datasets.
4.2 Emotion classification models
To address the challenge of quantifying the emotions elicited by music, researchers have developed a variety of models specifically designed to capture the nuances of music-induced emotions. These models aim to provide a structured approach to understanding and measuring the complex emotional responses that music can evoke. The classical music emotion models for analysis are shown in Figure 5, including the Hevner model (Hevner, 1936), the Pleasure Arousal Dominance (PAD) model (Russell, 1980), and the Thayer model (Thayer and McNally, 1992). With the development of artificial intelligence technology, some algorithm-based emotion classification approaches also have been proposed.
Figure 5. Schematic diagrams of common emotion recognition models. (A) Hevner’s model, (B) PAD model, (C) Thayer’s model.
In computerized categorization studies of musical emotions, the first psychological Hevner emotion classification model was proposed in 1936 (Hevner, 1936). This model classifies music emotion states into eight categories: Solemn, Sad, Dreamy, Quiet, Graceful, Happy, Excited, and Powerful as in Figure 5A. Each category can be further subdivided into more detailed and extensive emotion words, with a total of 67. This emotion classification model was set up considering musicology and psychology and is more abundant in the selection of emotion keywords, which is helpful for the research of emotion recognition in musical works. Hevner is a discrete emotion classification model that is often used as an emotion label for songs in music-induced emotion recognition research. However, due to the large number of labeling categories of the model and the relatively low variability of physiological properties of some categories, this model is seldom applied in EEG-based music emotion recognition studies, but it can be considered in relevant studies for featured music.
Among the classification models of musical emotions, PAD is a three-dimensional measurement model that was proposed in 1980. As in Figure 5B, this model establishes three main dimensions of Pleasure-Displeasure, Arousal-Nonarousal, and Dominance-Submissiveness, which indicate the direction of the emotion, the degree of neurophysiological activation, and the degree of individual feeling, respectively. This categorization provides a continuous quantitative evaluation method, which has been widely used in psychological and emotional brand evaluation (Yang et al., 2020) but has not been used much in the actual evaluation of musical emotions. Thayer’s model is a two-dimensional model that suggests different emotions are classified based on two-dimensional underlying dimensions, i.e., Energy awakening and Tension awakening. Using stress as the horizontal coordinate and energy as the vertical coordinate, emotions are categorized into four zones: vitality, anxiety, contentment, and depression as in Figure 5C. This model is proposed based on a psychological perspective and describes the music mood according to quantitative thinking, which is often used to classify the mood of audiophile music in MP3 and WAV formats (Brotzer et al., 2019). Since this model has fewer classification dimensions compared with other mentioned models and is more visible on the emotional response, they are well suitable to be used for EEG recognition of music-induced emotions. Besides the above classical emotion classification models, some researchers have also used probability distribution (Kim et al., 2022), neural network method (Yang, 2021), linear regression (Griffiths et al., 2021), and inverse word pairs (Liu et al., 2019) approaches to characterize the emotions of music. The probability distribution method describes the emotions corresponding to the song in the emotion description space, which gives a more comprehensive and intuitive description of the song. The ranking is utilized to order the emotion descriptions of songs according to the degree of relevance of the emotions expressed, which is convenient to provide a quick description method. The antonym pairs can give a relatively objective description of the mood of the music. Several researchers have currently extended discrete and multidimensional models for music emotion description based on these ideas.
These new classification models are related to the development of emotion categorization algorithms and have a large potential for application in the field of EEG music-induced emotions. Different emotion models can be used to describe the classification of emotions in different states, meanwhile, there are some intersections between these different models. For different practical applications, people need to choose the appropriate emotion classification models according to the research scenarios and artificial intelligence algorithms.
4.3 Emotional intensity model
Emotional intensity models were applied to quantitatively delineate the depth of emotions experienced by individuals in specific circumstances, forming the cornerstone for emotion recognition and prediction models. The discourse on quantifying emotional intensity emerged around 1990, advocating for a shift from solely focusing on subjective emotional aspects to incorporating physiological and observable behaviors as metrics of intensity (Russell, 1980; Sonnemans and Frijda, 1994). In 1994, Frijida introduced a five-dimensional model for scrutinizing subjective emotions, encompassing dimensions such as the frequency and intensity of re-collected and re-experienced emotions, latency and duration of emotions, intensity of actions and propensities, as well as actual behaviors, beliefs, and behavioral changes (Sonnemans and Frijda, 1994). In recent years, researchers have explored the use of emotional intensity modeling to study the instantaneous dynamic processes in the brain under external stimuli, as in Figure 6A (Gan et al., 2024). This innovative approach provides a new approach to the study of the neural mechanisms and processes of music-induced emotions.
Figure 6. (A) The instantaneous emotional intensity of stimulated emotion dynamic process (Gan et al., 2024), (B) schematic of a fine-grained emotional division.
Despite offering a theoretical framework for objectively describing emotional states, the model’s impact was limited due to the scarcity of physiological emotion measures at that time.
A prevalent theoretical framework in recent years for elucidating emotional intensity is Russell, 1980 proposition that emotional experiences can be depicted in a two-dimensional space defined by emotional valence (positive vs. negative emotions) and arousal levels (high vs. low) (Russell, 1980). Building upon this framework, researchers have delved into refining each dimension to achieve a nuanced portrayal of emotions illustrated in Figure 6B, laying the groundwork for leveraging artificial intelligence in digitally characterizing emotions (Reisenzein, 1994; Zhang et al., 2023). Physiological emotional intensity indices such as EEG, ECG, and MRI are not only valuable for emotion recognition but also serve as essential tools for studying the dynamic processes and physiological mechanisms underlying music-induced emotional changes (Ueno and Shimada, 2023; Wu et al., 2024).
5 Artificial intelligence algorithms for EEG emotion recognition
Music-induced EEG-based emotion classification research can be considered an artificial intelligence classification task, where the selection of appropriate classification algorithms is a crucial element in the current research on EEG-based emotion classification. These algorithms are not only the topicality of emotion classification research in EEG but also serve as an important foundation for further research into music-induced emotions (Khabiri et al., 2023).
5.1 Classical machine learning algorithms
Based on the EEG feature signals of music-induced emotions, various classical machine learning classification methods have been commonly used to achieve relatively good classification accuracy. These methods include classical classifiers such as Bayes classifier (BC) (Koelstra et al., 2012), linear regression (LR) (Zheng and Lu, 2015), support vector machines (SVM) (Bagherzadeh et al., 2023), K Nearest Neighbor (KNN) (Khabiri et al., 2023), and random forests (RF) (Pei et al., 2024), as in Table 6.
Table 6. Several classical machine learning methods applied in music-induced emotion classification.
These algorithms have been extensively employed in the field of emotion classification research and have shown promising results in accurately classifying mu-sic-induced emotions (Dadebayev et al., 2022). One of the commonly used supervised classification algorithms for music sentiment is the K Nearest Neighbor (KNN) algorithm. KNN, as a supervised learning algorithm, is highly versatile and easy to understand. It is robust to outliers and has a simple principle. However, the computational results of the KNN algorithm can be influenced by the training set samples as well as the value of K, which represents the number of nearest neighbors considered for classification. It is important to carefully select the appropriate value of K and ensure the representativeness and quality of the training set to achieve accurate classification results in music sentiment analysis. Another commonly used classical classifier for music sentiment analysis is the Support Vector Machine (SVM). When using SVM for classification, the choice of the kernel function has a significant impact on its performance. By mapping the features nonlinearly to a high-dimensional space using the kernel function, SVM improves the robustness of the music emotion recognition algorithm (Cai et al., 2022). SVM is particularly effective in handling high-dimensional data, making it suitable for achieving better classification results in music EEG emotion recognition compared to KNN.
Classical machine learning algorithms exhibit strong interpretability, high data efficiency, and low computational resource requirements in music emotion recognition research. These characteristics are highly desirable for studying the neural mechanisms of music-induced mood changes. However, in practical applications of music emotions, these models often suffer from poor generalization performance and require improved accuracy.
5.2 Deep learning algorithms
Although machine learning algorithms have been used for emotion recognition and have shown improvements, there are still challenges such as feature extraction, stability, and accuracy. However, the emergence of deep learning methods in recent years has provided a promising approach for EEG-based music emotion recognition research. Deep learning algorithms, characterized by their strong learning ability, have shown great potential in this field.
One notable deep learning algorithm applied in EEG-based music emotion recognition research is Convolutional Neural Networks (CNN). CNN extends the network structure of artificial neural networks by incorporating convolutional layers and pooling layers between the input layer and the fully connected layer. This architecture allows CNN to automatically learn and extract relevant features from the input data, making it suitable for analyzing complex patterns in EEG signals. By leveraging deep learning techniques, researchers can enhance the performance of music emotion recognition systems. Deep learning algorithms can effectively handle the high-dimensional and time-varying nature of EEG signals, leading to improved accuracy and stability in emotion recognition tasks. Moreover, with the ability to capture hierarchical representations, CNN can capture both local and global features in EEG data, enabling a more comprehensive analysis of music-induced emotions (Moctezuma et al., 2022; Mahmoud et al., 2023). With the development of deep learning algorithms, a variety of deep learning models have been developed and applied to EEG-based music-induced emotion recognition, including recurrent neural networks (RNN) (Dar et al., 2022), long and short-term memory networks (LSTM) (Du et al., 2022), gated recurrent neural networks (GRNN) (Weerakody et al., 2021) and autoencoder (AE) (Liu et al., 2020). The properties and applications of these reported deep learning algorithms are summarized in the following Table 7.
Table 7. Properties of typical deep learning methods applied in music-induced emotion classification.
The deep learning algorithms employed in EEG recognition of music-induced emotions demonstrate excellent generalization capabilities and data insensitivity, essential for the practical application of such emotions. While deep learning algorithms typically lack interpretability, recent advancements like GRNN (Weerakody et al., 2021) and RNN (Dar et al., 2022) can effectively capture the temporal aspects of EEG data, offering a novel approach to investigating the transient characteristics of music-induced emotions.
5.3 Model optimization and fusion strategies
Previous studies have demonstrated that classical machine learning algorithms as well as deep learning algorithms each possess their unique strengths and weaknesses in EEG-based music-induced emotion recognition research. To address the research and application requirements in related domains, researchers have investigated fusion strategies involving diverse algorithms.
To enhance the precision of emotion recognition, researchers have delved into a hybrid deep learning framework combining gated recurrent unit (GRU) and CNN to leverage the strengths of both methodologies. The conventional GRU model is excellent in handling time series data, while the CNN model is adept at capturing spatial features within the data. During the implementation phase, researchers opted to retain all feature information outputted by the GRU and extract spatial information from the temporal features using the CNN model. Ultimately, they achieved a recognition average accuracy of 87.04% (Xu G. et al., 2023). Based on the brain’s functional network of emotional activity, researchers proposed a multi-feature fusion method combining energy activation, spatial distribution, and brain functional connectivity network features. In the study, the SVM model-based fusion of power activation features of differential entropy (DE), spatial features of common spatial patterns (CSP), five frequency features, and phase synchronization features of EEG phase-locked values (PLV) achieved classification results with an average accuracy around 85% (Pan et al., 2022).
To realize the fusion between different machine learning algorithms, it can be achieved by combining multiple basic classifiers for better performance, fusing different algorithmic training models for model fusion of multiple ones, and also by joint training of multiple neural network models for fusion of different algorithms, etc. In addition to these fusion approaches mentioned above, some researchers have also considered about the optimization method from a music-induced emotions perspective. A pentameter-based EEG music model was proposed. The model constructs a multi-channel EEG sensor network and records the EEG of individuals in various emotional states to establish a mapping library of EEG and emotions. Subsequently, the music pentameter model is employed to adaptively segment the time-domain EEG signal, transforming the EEG signal. The time-frequency features of the EEG signal, such as amplitude, contour, and signal frequency, are quantitatively represented in a normalized musical space (Li and Zheng, 2023). The arithmetic modeling process is described as in Figure 7.
5.4 Algorithm comparison and evaluation
Evaluating these various algorithms used for music-induced emotion EEG is very difficult. There is no clear consensus on an optimal algorithm, and several metrics can be employed to evaluate the selected algorithmic model to satisfy the requirements for various applications. The accuracy of emotion recognition is a fundamental evaluation metric, reflecting the model’s performance in correctly predicting samples compared to the total number of samples and overall classifying emotions. Precision and recall, as evaluation metrics for binary classification problems, aid in assessing the model’s performance across different categories. With the expansion of applications and the development of algorithms, developing new criteria is also an important part of future research.
6 Application examples and analysis
EEG-based music emotion recognition has emerged as a multidisciplinary research area at the intersection of medicine, psychology, and computer science. This field explores the use of EEG signals to detect and classify the emotional responses evoked by music. The insights gained from EEG-based music emotion recognition have profound implications across various domains.
6.1 Music therapy
Music therapy is a therapeutic approach that harnesses the power of music and musical activities to promote both physical and mental well-being. The use of music for healing purposes has a long history, dating back to ancient civilizations like Greece, Rome, Egypt, and China. In more recent times, the formal recognition of music as a legitimate form of therapy began with the establishment of the American Music Therapy Association in 1944 (Taylor, 1981). This marked a significant milestone in acknowledging music therapy as a valid and effective treatment modality within modern society. To broaden the scope of music therapy, the American Music Therapy Association took a significant step in 2005 by introducing the Research Strategic Priority (RSP) program. The primary objective of this initiative is to delve into the physiological evidence supporting the effectiveness of music therapy in both practical and theoretical contexts. In 2013, a team of researchers from Finland conducted a study to investigate the impact of music on the activity of frontotemporal areas during resting state in individuals with depression. The study utilized EEG-based recognition of music-induced emotions as a methodological approach (Fachner et al., 2013). In 2018, a team of Spanish researchers conducted a study to evaluate the emotional response to music in patients with advanced cancer using EEG signals. The study aimed to demonstrate the positive emotional impact of music therapy on these patients (Ramirez et al., 2018). In 2020, a team of Canadian researchers conducted a study to explore the potential of music-induced emotions in alleviating psycho-cognitive symptoms of Alzheimer’s disease. The study involved combining EEG analysis methods to investigate how music activates the brain system, reduces negative emotions, and increases positive emotions. By analyzing EEG signals, the researchers were able to assess the emotional states induced by music. They found that music had a significant impact on the participants’ emotional well-being, with the potential to reduce negative emotions and increase positive emotions (Byrns et al., 2020).
6.2 Neuroscience
Brain science is a rapidly growing field of research that offers valuable insights into human thinking, behavior, and consciousness. One area of study within brain science is the investigation of how music stimulates the brain, which has been recognized as a notable example of this research. In 1992, French and German scientists conducted a groundbreaking EEG analysis study to examine the effects of music stimulation on the brain. The study revealed a fascinating phenomenon: different types of music had varying impacts on the intensity of EEG signals across different frequency bands (Steinberg et al., 1992). In 2016, a team of Indian researchers conducted a study using EEG to investigate the effects of Hindustani music on brain activity in a relaxed state. The results of the study revealed that Hindustani music had a significant effect on the listeners’ arousal levels in all activities stimulated by the music. The EEG analysis indicated an increase in brain activity in response to the music, suggesting that it had a stimulating effect on the listeners (Banerjee et al., 2016). In 2019, a group of Indian scholars delved into research on the reverse inversion of brain sounds by utilizing Hindustani classical music. They recognized the profound emotional impact of this music and sought to explore the correlation between EEG signals and musical stimuli. By leveraging the real-time recording capability of EEG, researchers from the fields of psychology and neurology conducted studies to analyze the neural mechanisms underlying the stimulation of music, both in positive and negative contexts. These investigations have significantly contributed to the advancement of brain science research (Sanyal et al., 2019).
In the early stages, EEG, as a direct signal of brain activity, was employed by neuroscientists to conduct exploratory studies on functional brain regions associated with impaired musical ability caused by brain dysfunction. This utilization of EEG monitoring technology has played a pivotal role in advancing our understanding of the brain’s mechanisms involved in music processing (Vuust et al., 2022). These initial findings laid the technical groundwork for subsequent research on EEG-based music emotion recognition. With a focus on music-induced emotions, researchers have endeavored to further investigate the realm of music-induced emotions using EEG technology (Gomez-Canon et al., 2021). From the perspective of music therapy, the utilization of EEG signals offers direct evidence regarding the process of music-induced emotions. By analyzing EEG signals from various brain regions corresponding to different emotions, researchers can obtain more detailed physiological information that aids in the interpretation of the brain mechanisms involved in music therapy. This application of EEG signals provides valuable insights into understanding the effects of music on emotional states and enhances our knowledge of the therapeutic potential of music (Byrns et al., 2020; Fedotchev et al., 2022).
6.3 Others
Emotions play a crucial role in human experiences, behaviors, health, and social interactions. Music, a language of the human mind, has the power to vividly and imaginatively express various emotions such as happiness, sadness, and more, and can greatly influence listeners’ emotional state. In recent years, there has been significant progress in understanding music-induced emotions and their psychological and neurological mechanisms.
In clinical medicine, this research can contribute to the development of personalized music therapy interventions for mental health disorders, neurorehabilitation, and stress management. It can also aid in diagnosing and monitoring emotional disorders such as depression and anxiety. In the realm of brain science, EEG-based music emotion recognition provides valuable insights into the neural mechanisms underlying emotional processing and music perception. These findings can enhance our understanding of how the brain responds to music and its impact on emotional well-being. Moreover, in the field of music information, this research can improve music recommendation systems, enhance user experiences, and facilitate music healing approaches. By tailoring music selections based on an individual’s emotional responses, music platforms can offer personalized and therapeutic listening experiences. Overall, EEG-based music emotion recognition holds immense potential for diverse applications in fields like clinical medicine, brain science, and music information. It represents a promising avenue for advancing our understanding of the complex relationship between music and emotions and harnessing music’s therapeutic benefits.
Furthermore, for some long-term music healing processes, the real-time sensitivity of EEG to emotional signals induced by music stimulation can provide evidence for the effectiveness of certain therapeutic methods. This evidence can facilitate the development, correction, and smooth dissemination of related therapeutic techniques. By monitoring changes in EEG signals throughout the music therapy process, researchers can evaluate the effectiveness of different therapeutic methods and fine-tune them accordingly. This approach enhances the precision and efficacy of music therapy, allowing for optimized treatment plans that cater to individual needs (Byrns et al., 2020). For music researchers, the individual variability in EEG emotion detection allows for personalized categorization and annotation of musical emotions. This capability is crucial not only for music composition and information retrieval but also for guiding the development of more immersive multimedia environments. By leveraging EEG data to understand how individuals uniquely experience and respond to musical emotions, researchers can enhance the creation of tailored musical experiences and enrich the design of multimedia environments that resonate with diverse emotional responses (Yu et al., 2022).
7 Discussion and conclusions
Based on the mentioned model, researchers were able to carry out systematic research on the study of the emotional impacts of the same music on different listeners, the study of the emotional impact of various types of music on the same listener, and the key parameters of music-stimulated emotions. Previous researchers have conducted various studies in terms of music-induced emotion classification models, music-induced datasets, training and classification of emotion models, and so on.
As an interdisciplinary challenge, research on EEG-based music-induced emotion recognition has emerged as a valuable approach for real-time and effective assessment of emotional responses to music. This innovative research not only offers new technical tools for studying music-related emotions but also provides a controllable research paradigm applicable to brain science and other fields. In recent years, researchers from various disciplines have made significant progress in addressing this complex issue. By approaching the problem from different perspectives, they have achieved notable results. However, during these investigations, several limitations have also been identified.
Compared to other signals commonly used for music emotion recognition, such as audio signals, facial expressions, heart rate, and respiration, EEG signals have distinct advantages. EEG signals belong to physiological signals of the central nervous system, which are typically not under conscious control. Consequently, they can provide information about the current emotional state of an individual that cannot be deliberately concealed. Furthermore, EEG signals offer several benefits when compared to other methods of detecting physiological signals of the central nervous system. EEG is a relatively mature technology that has been extensively studied and validated. It is also portable, non-invasive, and cost-effective, making it practical for use in various research and real-world settings.
As EEG monitoring hardware and recognition algorithm software technology continue to evolve, the advantages of personalization, real-time analysis, and the convenience of using EEG to recognize music-induced emotions will be further explored in various application fields. The growing sophistication of EEG technology opens up new possibilities for research and practical implementation of music-based therapies, multimedia environments, and personalized music experiences. As such, the continued development and refinement of EEG-based music emotion recognition has the potential to revolutionize our understanding of the impact of music on human emotions and behavior. Advancements in EEG monitoring hardware and recognition algorithm software technology have opened up new avenues for exploring the potential applications of EEG-based music-induced emotion recognition. With these technological improvements, the advantages of personalization, real-time, and convenience in recognizing music-induced emotions through EEG can be further explored in various fields.
At present, the labeling basis of the training set in EEG emotion recognition algorithms largely relies on psychological scales and emotion labels from databases. However, these conventional labeling methods are inherently subjective and discrete. Therefore, there is a pressing need for extensive research to establish a standardized library of emotions based on EEG signals themselves. To address this challenge, musicologists from diverse cultural backgrounds have embarked on initial research into the emotional labeling of music within their respective cultures. As the accuracy of EEG signals for emotion recognition continues to improve, there has been increasing mention of establishing direct EEG signal discrimination for personalized emotion recognition. This advancement holds promise for enhancing our understanding of how individuals from different cultural backgrounds experience and interpret emotions in music, paving the way for more nuanced and culturally sensitive approaches to music emotion recognition (Du et al., 2023).
Improving the accuracy of music-induced emotion recognition can be a challenging problem that demands long-term research, and the advent of deep learning algorithms in recent years has provided a more effective means of addressing this challenge compared to traditional machine learning approaches. With deep learning algorithms, researchers can leverage large amounts of data to train neural networks that can learn to recognize complex patterns and relationships in music-induced emotions. This approach has shown great promise in improving the accuracy of music emotion recognition, allowing researchers to gain deeper insights into how music affects human emotions and behavior. However, ongoing research and development are still required to further refine and optimize these algorithms for use in practical applications (Pandey and Seeja, 2022). As artificial intelligence algorithms continue to undergo continuous optimization and enhancement, new concepts, approaches, and research findings will undoubtedly emerge, offering fresh perspectives and advancing the field of music-induced emotion recognition.
The research and development of music-induced emotion recognition based on EEG relies on the continuous expansion of the application field for such technology. Currently, there are some notable examples of music-induced emotion applications in clinical treatment (Byrns et al., 2020), neuroscience (Luo et al., 2023), and music information retrieval (Li et al., 2023). However, there is still a need for further development of related technical products that can be scaled up and made accessible to the general public, allowing them to better understand and benefit from this technology. This requires ongoing efforts to bridge the gap between research and practical implementation, fostering the creation of user-friendly tools and platforms that can effectively harness the potential of music-induced emotion recognition for broader applications and public engagement. In the realm of consumer applications, there is still much to be explored regarding the potential combination of EEG and personalized music to develop emotional regulation technology and products for users.
Author contributions
YS: Conceptualization, Data curation, Funding acquisition, Investigation, Methodology, Writing – original draft, Writing – review & editing. YL: Resources, Supervision, Validation, Writing – original draft. YX: Data curation, Resources, Supervision, Validation, Writing – review & editing. JM: Formal analysis, Visualization, Writing – original draft. DL: Conceptualization, Formal analysis, Project administration, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This study was supported by the Zhejiang International Studies University’s key project “A comparative study of Chinese and English art education from the perspective of nationalization” (No. 090500112016), the first-class curriculum construction project from the Department of Education of Zhejiang Province (No. 080830302022), and Zhejiang Natural Science Foundation project (No. LQ21E060006).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Aftanas, L. I., Reva, N. V., Savotina, L. N., and Makhnev, V. P. (2006). Neurophysiological correlates of induced discrete emotions in humans: an individually oriented analysis. Neurosci. Behav. Physiol. 36, 119–130. doi: 10.1007/s11055-005-0170-6
Ahmad, J., Ellis, C., Leech, R., Voytek, B., Garces, P., Jones, E., et al. (2022). From mechanisms to markers: novel noninvasive EEG proxy markers of the neural excitation and inhibition system in humans. Transl. Psychiatry 12:467. doi: 10.1038/s41398-022-02218-z
Alarcao, S. M., and Fonseca, M. J. (2019). Emotions recognition using EEG signals: a survey. IEEE Trans. Affect. Comput. 10, 374–393. doi: 10.1109/TAFFC.2017.2714671
Aldridge, A., Barnes, E., Bethel, C. L., Carruth, D. W., Kocturova, M., Pleva, M., et al. (2019). “Accessible electroencephalograms (EEGs): a comparative review with OpenBCI’s ultracortex mark IV headset” in 2019 29th international conference radioelektronika (RADIOELEKTRONIKA) (Pardubice, Czech Republic: IEEE), 1–6.
Aljanaki, A., Yang, Y.-H., and Soleymani, M. (2017). Developing a benchmark for emotional analysis of music. PLoS One 12:e0173392. doi: 10.1371/journal.pone.0173392
Apicella, A., Arpaia, P., Frosolone, M., Improta, G., Moccaldi, N., and Pollastro, A. (2022a). EEG-based measurement system for monitoring student engagement in learning 4.0. Sci. Rep. 12:5857. doi: 10.1038/s41598-022-09578-y
Apicella, A., Arpaia, P., Isgro, F., Mastrati, G., and Moccaldi, N. (2022b). A survey on EEG-based solutions for emotion recognition with a low number of channels. IEEE Access 10, 117411–117428. doi: 10.1109/ACCESS.2022.3219844
Ara, A., and Marco-Pallarés, J. (2020). Fronto-temporal theta phase-synchronization underlies music-evoked pleasantness. NeuroImage 212:116665. doi: 10.1016/j.neuroimage.2020.116665
Bagherzadeh, S., Maghooli, K., Shalbaf, A., and Maghsoudi, A. (2023). A hybrid EEG-based emotion recognition approach using wavelet convolutional neural networks and support vector machine. Basic Clin. Neurosci. 14, 87–102. doi: 10.32598/bcn.2021.3133.1
Balasubramanian, G., Kanagasabai, A., Mohan, J., and Seshadri, N. P. G. (2018). Music induced emotion using wavelet packet decomposition—an EEG study. Biomed. Signal Process. Control 42, 115–128. doi: 10.1016/j.bspc.2018.01.015
Banerjee, A., Sanyal, S., Patranabis, A., Banerjee, K., Guhathakurta, T., Sengupta, R., et al. (2016). Study on brain dynamics by non linear analysis of music induced EEG signals. Phys. A Stat. Mech. Appl. 444, 110–120. doi: 10.1016/j.physa.2015.10.030
Bergee, M. J., and Weingarten, K. M. (2021). Multilevel models of the relationship between music achievement and reading and math achievement. J. Res. Music. Educ. 68, 398–418. doi: 10.1177/0022429420941432
Brotzer, J. M., Mosqueda, E. R., and Gorro, K. (2019). Predicting emotion in music through audio pattern analysis. IOP Conf. Ser. Mater. Sci. Eng. 482:12021. doi: 10.1088/1757-899X/482/1/012021
Byrns, A., Abdessalem, H. B., Cuesta, M., Bruneau, M.-A., Belleville, S., and Frasson, C. (2020). EEG analysis of the contribution of music therapy and virtual reality to the improvement of cognition in Alzheimer’s disease. J. Biomed. Sci. Eng. 13, 187–201. doi: 10.4236/jbise.2020.138018
Cai, Q., Cui, G.-C., and Wang, H.-X. (2022). EEG-based emotion recognition using multiple kernel learning. Mach. Intell. Res. 19, 472–484. doi: 10.1007/s11633-022-1352-1
Carpente, J., Casenhiser, D. M., Kelliher, M., Mulholland, J., Sluder, H. L., Crean, A., et al. (2022). The impact of imitation on engagement in minimally verbal children with autism during improvisational music therapy. Nord. J. Music. Ther. 31, 44–62. doi: 10.1080/08098131.2021.1924843
Chen, Y.-A., Yang, Y.-H., Wang, J.-C., and Chen, H. (2015). The AMG1608 dataset for music emotion recognition. In 2015 IEEE international conference on acoustics, speech and signal processing (ICASSP), (South Brisbane, Queensland, Australia: IEEE), 693–697.
Colin, C., Prince, V., Bensoussan, J.-L., and Picot, M.-C. (2023). Music therapy for health workers to reduce stress, mental workload and anxiety: a systematic review. J. Public Health 45, e532–e541. doi: 10.1093/pubmed/fdad059
Contreras-Molina, M., Rueda-Núñez, A., Pérez-Collado, M. L., and García-Maestro, A. (2021). Effect of music therapy on anxiety and pain in the critical polytraumatised patient. Enferm. Intensiva 32, 79–87. doi: 10.1016/j.enfie.2020.03.005
Cui, X., Wu, Y., Wu, J., You, Z., Xiahou, J., and Ouyang, M. (2022). A review: music-emotion recognition and analysis based on EEG signals. Front. Neuroinform. 16:997282. doi: 10.3389/fninf.2022.997282
Dadebayev, D., Goh, W. W., and Tan, E. X. (2022). EEG-based emotion recognition: review of commercial EEG devices and machine learning techniques. J. King Saud Univ. Comput. Inf. Sci. 34, 4385–4401. doi: 10.1016/j.jksuci.2021.03.009
Dar, M., Akram, M., Yuvaraj, R., Khawaja, S., and Murugappan, M. (2022). EEG-based emotion charting for Parkinson’s disease patients using convolutional recurrent neural networks and cross dataset learning. Comput. Biol. Med. 144:105327. doi: 10.1016/j.compbiomed.2022.105327
Davidson, R. J., and Fox, N. A. (1982). Asymmetrical brain activity discriminates between positive and negative affective stimuli in human infants. Science 218, 1235–1237. doi: 10.1126/science.7146906
Du, X., Ma, C., Zhang, G., Li, J., Lai, Y.-K., Zhao, G., et al. (2022). An efficient LSTM network for emotion recognition from multichannel EEG signals. IEEE Trans. Affect. Comput. 13, 1528–1540. doi: 10.1109/TAFFC.2020.3013711
Du, R., Zhu, S., Ni, H., Mao, T., Li, J., and Wei, R. (2023). Valence-arousal classification of emotion evoked by Chinese ancient-style music using 1D-CNN-BiLSTM model on EEG signals for college students. Multimed. Tools Appl. 82, 15439–15456. doi: 10.1007/s11042-022-14011-7
Egger, M., Ley, M., and Hanke, S. (2019). Emotion recognition from physiological signal analysis: a review. Electron. Notes Theor. Comput. Sci. 343, 35–55. doi: 10.1016/j.entcs.2019.04.009
Er, M. B., Çiğ, H., and Aydilek, İ. B. (2021). A new approach to recognition of human emotions using brain signals and music stimuli. Appl. Acoust. 175:107840. doi: 10.1016/j.apacoust.2020.107840
Fachner, J., Gold, C., and Erkkilä, J. (2013). Music therapy modulates fronto-temporal activity in rest-EEG in depressed clients. Brain Topogr. 26, 338–354. doi: 10.1007/s10548-012-0254-x
Fedotchev, A., Parin, S., Polevaya, S., and Zemlianaia, A. (2022). EEG-based musical neurointerfaces in the correction of stress-induced states. Brain-Comput. Interfaces 9, 1–6. doi: 10.1080/2326263X.2021.1964874
Gan, K., Li, R., Zhang, J., Sun, Z., and Yin, Z. (2024). Instantaneous estimation of momentary affective responses using neurophysiological signals and a spatiotemporal emotional intensity regression network. Neural Netw. 172:106080. doi: 10.1016/j.neunet.2023.12.034
Geipel, J., Koenig, J., Hillecke, T. K., and Resch, F. (2022). Short-term music therapy treatment for adolescents with depression – a pilot study. Art. Psychother. 77:101874. doi: 10.1016/j.aip.2021.101874
Geretsegger, M., Fusar-Poli, L., Elefant, C., Mössler, K. A., Vitale, G., and Gold, C. (2022). Music therapy for autistic people. Cochrane Database Syst. Rev. 2022:CD004381. doi: 10.1002/14651858.CD004381.pub4
Gomez-Canon, J. S., Cano, E., Eerola, T., Herrera, P., Hu, X., Yang, Y.-H., et al. (2021). Music emotion recognition: toward new, robust standards in personalized and context-sensitive applications. IEEE Signal Process. Mag. 38, 106–114. doi: 10.1109/MSP.2021.3106232
Griffiths, D., Cunningham, S., Weinel, J., and Picking, R. (2021). A multi-genre model for music emotion recognition using linear regressors. J. New Music Res. 50, 355–372. doi: 10.1080/09298215.2021.1977336
Guo, S., Lu, J., Wang, Y., Li, Y., Huang, B., Zhang, Y., et al. (2020). Sad music modulates pain perception: an EEG study. J. Pain Res. 13, 2003–2012. doi: 10.2147/JPR.S264188
Hartmann, M., Mavrolampados, A., Toiviainen, P., Saarikallio, S., Foubert, K., Brabant, O., et al. (2023). Musical interaction in music therapy for depression treatment. Psychol. Music 51, 33–50. doi: 10.1177/03057356221084368
Hevner, K. (1936). Experimental studies of the elements of expression in music. Am. J. Psychol. 48:246. doi: 10.2307/1415746
Hossain, S., Rahman, M., Chakrabarty, A., Rashid, M., Kuwana, A., and Kobayashi, H. (2023). Emotional state classification from MUSIC-based features of multichannel EEG signals. Bioengineering 10:99. doi: 10.3390/bioengineering10010099
Humphreys, J. T. (1998). Musical aptitude testing: from James McKeen Cattell to Carl Emil seashore. Res. Stud. Music Educ. 10, 42–53. doi: 10.1177/1321103X9801000104
Katsigiannis, S., and Ramzan, N. (2018). DREAMER: a database for emotion recognition through EEG and ECG signals from wireless low-cost off-the-shelf devices. IEEE J. Biomed. Health Inform. 22, 98–107. doi: 10.1109/JBHI.2017.2688239
Khabiri, H., Naseh Talebi, M., Kamran, M. F., Akbari, S., Zarrin, F., and Mohandesi, F. (2023). Music-induced emotion recognition based on feature reduction using PCA from EEG signals. Front. Biomed. Technol. 11, 59–68. doi: 10.18502/fbt.v11i1.14512
Khare, S. K., and Bajaj, V. (2021). Time–frequency representation and convolutional neural network-based emotion recognition. IEEE Trans. Neural Networks Learn. Syst. 32, 2901–2909. doi: 10.1109/TNNLS.2020.3008938
Kim, H., Zhang, D., Kim, L., and Im, C.-H. (2022). Classification of individual’s discrete emotions reflected in facial microexpressions using electroencephalogram and facial electromyogram. Expert Syst. Appl. 188:116101. doi: 10.1016/j.eswa.2021.116101
Koelstra, S., Muhl, C., Soleymani, M., Lee, J.-S., Yazdani, A., Ebrahimi, T., et al. (2012). DEAP: a database for emotion analysis using physiological signals. IEEE Trans. Affect. Comput. 3, 18–31. doi: 10.1109/T-AFFC.2011.15
Lapomarda, G., Valer, S., Job, R., and Grecucci, A. (2022). Built to last: theta and delta changes in resting-state EEG activity after regulating emotions. Brain Behav. 12:e2597. doi: 10.1002/brb3.2597
Li, P., Liu, H., Si, Y., Li, C., Li, F., Zhu, X., et al. (2019). EEG based emotion recognition by combining functional connectivity network and local activations. IEEE Trans. Biomed. Eng. 66, 2869–2881. doi: 10.1109/TBME.2019.2897651
Li, D., Ruan, Y., Zheng, F., Lijuan, S., and Lin, Q. (2022a). Effect of Taiji post-standing on the brain analyzed with EEG signals. J. Taiji Sci. 1, 2–7. doi: 10.57612/2022.jts.01.01
Li, D., Ruan, Y., Zheng, F., Su, Y., and Lin, Q. (2022b). Fast sleep stage classification using cascaded support vector machines with single-channel EEG signals. Sensors 22:9914. doi: 10.3390/s22249914
Li, M., Xu, H., Liu, X., and Lu, S. (2018). Emotion recognition from multichannel EEG signals using K-nearest neighbor classification. Technol. Health Care 26, 509–519. doi: 10.3233/THC-174836
Li, X., Zhang, Y., Tiwari, P., Song, D., Hu, B., Yang, M., et al. (2023). EEG based emotion recognition: a tutorial and review. ACM Comput. Surv. 55, 1–57. doi: 10.1145/3524499
Li, Y., and Zheng, W. (2023). EEG processing in emotion recognition: inspired from a musical staff. Multimed. Tools Appl. 82, 4161–4180. doi: 10.1007/s11042-022-13405-x
Liang, J., Tian, X., and Yang, W. (2021). Application of music therapy in general surgical treatment. Biomed. Res. Int. 2021, 1–4. doi: 10.1155/2021/6169183
Liu, H., Fang, Y., and Huang, Q. (2019). Music emotion recognition using a variant of recurrent neural network., in Proceedings of the 2018 international conference on mathematics, modeling, simulation and statistics application (MMSSA 2018), (Shanghai, China: Atlantis Press).
Liu, J., Wu, G., Luo, Y., Qiu, S., Yang, S., Li, W., et al. (2020). EEG-based emotion classification using a deep neural network and sparse autoencoder. Front. Syst. Neurosci. 14:43. doi: 10.3389/fnsys.2020.00043
Lu, G., Jia, R., Liang, D., Yu, J., Wu, Z., and Chen, C. (2021). Effects of music therapy on anxiety: a meta-analysis of randomized controlled trials. Psychiatry Res. 304:114137. doi: 10.1016/j.psychres.2021.114137
Luo, E., Pan, W., and Fan, X. (2023). Music, language, and autism: neurological insights for enhanced learning. Int. J. Innov. Res. Med. Sci. 8, 398–408. doi: 10.23958/ijirms/vol08-i09/1743
Maffei, A. (2020). Spectrally resolved EEG intersubject correlation reveals distinct cortical oscillatory patterns during free-viewing of affective scenes. Psychophysiology 57:e13652. doi: 10.1111/psyp.13652
Mahmoud, A., Amin, K., Al Rahhal, M., Elkilani, W., Mekhalfi, M., and Ibrahim, M. (2023). A CNN approach for emotion recognition via EEG. Symmetry 15:1822. doi: 10.3390/sym15101822
Martínez-Saez, M., Ros, L., López-Cano, M., Nieto, M., Navarro, B., and Latorre, J. (2024). Effect of popular songs from the reminiscence bump as autobiographical memory cues in aging: a preliminary study using EEG. Front. Neurosci. 17:1300751. doi: 10.3389/fnins.2023.1300751
Martins, I., Lima, C., and Pinheiro, A. (2022). Enhanced salience of musical sounds in singers and instrumentalists. Cogn. Affect. Behav. Ne. 22, 1044–1062. doi: 10.3758/s13415-022-01007-x
Metfessel, M. (1950). Carl emil seashore, 1866-1949. Science 111, 713–717. doi: 10.1126/science.111.2896.713
Micallef Grimaud, A., and Eerola, T. (2022). An interactive approach to emotional expression through musical cues. Music. Sci. 5:205920432110617. doi: 10.1177/20592043211061745
Moctezuma, L., Abe, T., and Molinas, M. (2022). Two-dimensional CNN-based distinction of human emotions from EEG channels selected by multi-objective evolutionary algorithm. Sci. Rep. 12:3523. doi: 10.1038/s41598-022-07517-5
Nawaz, R., Ng, J. T., Nisar, H., and Voon, Y. V. (2019). Can background music help to relieve stress? An EEG analysis., in 2019 IEEE international conference on signal and image processing applications (ICSIPA), Kuala Lumpur, Malaysia: IEEE, 68–72.
Oktavia, N. Y., Wibawa, A. D., Pane, E. S., and Purnomo, M. H. (2019). “Human emotion classification based on EEG signals using naïve bayes method” in 2019 international seminar on application for technology of information and communication (iSemantic) (Semarang, Indonesia: IEEE), 319–324.
Palazzi, A., Meschini, R., and Piccinini, C. A. (2021). NICU music therapy effects on maternal mental health and preterm infant’s emotional arousal. Infant Ment. Health J. 42, 672–689. doi: 10.1002/imhj.21938
Pan, J., Yang, F., Qiu, L., and Huang, H. (2022). Fusion of EEG-based activation, spatial, and connection patterns for fear emotion recognition. Comput. Intell. Neurosci. 2022, 1–11. doi: 10.1155/2022/3854513
Panda, R., Malheiro, R., and Paiva, R. P. (2023). Audio features for music emotion recognition: a survey. IEEE Trans. Affect. Comput. 14, 68–88. doi: 10.1109/TAFFC.2020.3032373
Pandey, P., and Seeja, K. R. (2022). Subject independent emotion recognition from EEG using VMD and deep learning. J. King Saud Univ. Comput. Inf. Sci. 34, 1730–1738. doi: 10.1016/j.jksuci.2019.11.003
Pei, G., Shang, Q., Hua, S., Li, T., and Jin, J. (2024). EEG-based affective computing in virtual reality with a balancing of the computational efficiency and recognition accuracy. Comput. Hum. Behav. 152:108085. doi: 10.1016/j.chb.2023.108085
Perlovsky, L. (2012). Cognitive function, origin, and evolution of musical emotions. Music. Sci. 16, 185–199. doi: 10.1177/1029864912448327
Ramirez, R., Planas, J., Escude, N., Mercade, J., and Farriols, C. (2018). EEG-based analysis of the emotional effect of music therapy on palliative care cancer patients. Front. Psychol. 9:254. doi: 10.3389/fpsyg.2018.00254
Reisenzein, R. (1994). Pleasure-arousal theory and the intensity of emotions. J. Pers. Soc. Psychol. 67, 525–539. doi: 10.1037/0022-3514.67.3.525
Rolls, E. T. (2015). Limbic systems for emotion and for memory, but no single limbic system. Cortex 62, 119–157. doi: 10.1016/j.cortex.2013.12.005
Russell, J. A. (1980). A circumplex model of affect. J. Pers. Soc. Psychol. 39, 1161–1178. doi: 10.1037/h0077714
Ruth, N., and Schramm, H. (2021). Effects of prosocial lyrics and musical production elements on emotions, thoughts and behavior. Psychol. Music 49, 759–776. doi: 10.1177/0305735620902534
Ryczkowska, A. (2022). Positive mood induction through music: the significance of listener age and musical timbre. Psychol. Music 50, 1961–1975. doi: 10.1177/03057356221081164
Saganowski, S., Perz, B., Polak, A. G., and Kazienko, P. (2023). Emotion recognition for everyday life using physiological signals from wearables: a systematic literature review. IEEE Trans. Affect. Comput. 14, 1876–1897. doi: 10.1109/TAFFC.2022.3176135
Salakka, I., Pitkäniemi, A., Pentikäinen, E., Mikkonen, K., Saari, P., Toiviainen, P., et al. (2021). What makes music memorable? Relationships between acoustic musical features and music-evoked emotions and memories in older adults. PLoS One 16:e0251692. doi: 10.1371/journal.pone.0251692
Sammler, D., Grigutsch, M., Fritz, T., and Koelsch, S. (2007). Music and emotion: electrophysiological correlates of the processing of pleasant and unpleasant music. Psychophysiology 44, 293–304. doi: 10.1111/j.1469-8986.2007.00497.x
Sanyal, S., Nag, S., Banerjee, A., Sengupta, R., and Ghosh, D. (2019). Music of brain and music on brain: a novel EEG sonification approach. Cogn. Neurodyn. 13, 13–31. doi: 10.1007/s11571-018-9502-4
Sari, D. A. L., Kusumaningrum, T. D., and Kusumoputro, B. (2023). Non-linear EEG based emotional classification using k-nearest neighbor and weighted k-nearest neighbor with variation of features selection methods. AIP Conf. Proc. 2654:020004. doi: 10.1063/5.0116377
Schmidt, L. A., and Trainor, L. J. (2001). Frontal brain electrical activity (EEG) distinguishes valence and intensity of musical emotions. Cogn. Emot. 15, 487–500. doi: 10.1080/02699930126048
Shu, L., Xie, J., Yang, M., Li, Z., Li, Z., Liao, D., et al. (2018). A review of emotion recognition using physiological signals. Sensors 18:2074. doi: 10.3390/s18072074
Silverman, D. (1963). The rationale and history of the 10-20 system of the international federation. Am. J. EEG Technol. 3, 17–22. doi: 10.1080/00029238.1963.11080602
Soleymani, M., Caro, M. N., Schmidt, E. M., Sha, C.-Y., and Yang, Y.-H. (2013). 1000 songs for emotional analysis of music., in Proceedings of the 2nd ACM international workshop on crowdsourcing for multimedia, (Barcelona, Spain: ACM), 1–6.
Sonnemans, J., and Frijda, N. H. (1994). The structure of subjective emotional intensity. Cogn. Emot. 8, 329–350. doi: 10.1080/02699939408408945
Stancin, I., Cifrek, M., and Jovic, A. (2021). A review of EEG signal features and their application in driver drowsiness detection systems. Sensors 21:3786. doi: 10.3390/s21113786
Steinberg, R., Günther, W., Stiltz, I., and Rondot, P. (1992). EEG-mapping during music stimulation. Psychomusicol. J. Res. Music Cogn. 11, 157–170. doi: 10.1037/h0094123
Tang, Z., Xia, D., Li, X., Wang, X., Ying, J., and Yang, H. (2023). Evaluation of the effect of music on idea generation using electrocardiography and electroencephalography signals. Int. J. Technol. Des. Educ. 33, 1607–1625. doi: 10.1007/s10798-022-09782-x
Taylor, D. B. (1981). Music in general hospital treatment from 1900 to 1950. J. Music. Ther. 18, 62–73. doi: 10.1093/jmt/18.2.62
Tcherkassof, A., and Dupré, D. (2021). The emotion–facial expression link: evidence from human and automatic expression recognition. Psychol. Res. 85, 2954–2969. doi: 10.1007/s00426-020-01448-4
Thaut, M. H., Francisco, G., and Hoemberg, V. (2021). Editorial: the clinical neuroscience of music: evidence based approaches and neurologic music therapy. Front. Neurosci. 15:740329. doi: 10.3389/fnins.2021.740329
Thayer, R. E., and McNally, R. J. (1992). The biopsychology of mood and arousal. Cogn. Behav. Neurol. 5:65.
Torres, E. P., Torres, E. A., Hernández-Álvarez, M., and Yoo, S. G. (2020). EEG-based BCI emotion recognition: a survey. Sensors 20:5083. doi: 10.3390/s20185083
Turnbull, D., Barrington, L., Torres, D., and Lanckriet, G. (2008). Semantic annotation and retrieval of music and sound effects. IEEE Trans. Audio Speech Lang. Process. 16, 467–476. doi: 10.1109/TASL.2007.913750
Ueno, F., and Shimada, S. (2023). Inter-subject correlations of EEG reflect subjective arousal and acoustic features of music. Front. Hum. Neurosci. 17:1225377. doi: 10.3389/fnhum.2023.1225377
Vuust, P., Heggli, O. A., Friston, K. J., and Kringelbach, M. L. (2022). Music in the brain. Nat. Rev. Neurosci. 23, 287–305. doi: 10.1038/s41583-022-00578-5
Wang, X., Wei, Y., and Yang, D. (2022). Cross-cultural analysis of the correlation between musical elements and emotion. Cognit. Comput. Syst. 4, 116–129. doi: 10.1049/ccs2.12032
Weerakody, P. B., Wong, K. W., Wang, G., and Ela, W. (2021). A review of irregular time series data handling with gated recurrent neural networks. Neurocomputing 441, 161–178. doi: 10.1016/j.neucom.2021.02.046
Witkower, Z., Hill, A. K., Koster, J., and Tracy, J. L. (2021). Beyond face value: evidence for the universality of bodily expressions of emotion. Affect. Sci. 2, 221–229. doi: 10.1007/s42761-021-00052-y
Wu, Q., Sun, L., Ding, N., and Yang, Y. (2024). Musical tension is affected by metrical structure dynamically and hierarchically. Cogn. Neurodyn. 18, 1955–1976. doi: 10.1007/s11571-023-10058-w
Xu, G., Guo, W., and Wang, Y. (2023). Subject-independent EEG emotion recognition with hybrid spatio-temporal GRU-conv architecture. Med. Biol. Eng. Comput. 61, 61–73. doi: 10.1007/s11517-022-02686-x
Xu, J., Hu, L., Qiao, R., Hu, Y., and Tian, Y. (2023). Music-emotion EEG coupling effects based on representational similarity. J. Neurosci. Methods 398:109959. doi: 10.1016/j.jneumeth.2023.109959
Xu, J., Qian, W., Hu, L., Liao, G., and Tian, Y. (2024). EEG decoding for musical emotion with functional connectivity features. Biomed. Signal Process. Control 89:105744. doi: 10.1016/j.bspc.2023.105744
Yang, J. (2021). A novel music emotion recognition model using neural network technology. Front. Psychol. 12:760060. doi: 10.3389/fpsyg.2021.760060
Yang, H., Han, J., and Min, K. (2019). A multi-column CNN model for emotion recognition from EEG signals. Sensors 19:4736. doi: 10.3390/s19214736
Yang, K., Kim, H. M., and Zimmerman, J. (2020). Emotional branding on fashion brand websites: harnessing the pleasure-arousal-dominance (P-a-D) model. J. Fash. Mark. Manag. 24, 555–570. doi: 10.1108/JFMM-03-2019-0055
Yang, W., Makita, K., Nakao, T., Kanayama, N., Machizawa, M. G., Sasaoka, T., et al. (2018). Affective auditory stimulus database: an expanded version of the international affective digitized sounds (IADS-E). Behav. Res. Methods 50, 1415–1429. doi: 10.3758/s13428-018-1027-6
Yu, M., Xiao, S., Hua, M., Wang, H., Chen, X., Tian, F., et al. (2022). EEG-based emotion recognition in an immersive virtual reality environment: from local activity to brain network features. Biomed. Signal Process. Control 72:103349. doi: 10.1016/j.bspc.2021.103349
Zentner, M., Grandjean, D., and Scherer, K. R. (2008). Emotions evoked by the sound of music: characterization, classification, and measurement. Emotion 8, 494–521. doi: 10.1037/1528-3542.8.4.494
Zhang, M., Ding, Y., Zhang, J., Jiang, X., Xu, N., Zhang, L., et al. (2022). Effect of group impromptu music therapy on emotional regulation and depressive symptoms of college students: a randomized controlled study. Front. Psychol. 13:851526. doi: 10.3389/fpsyg.2022.851526
Zhang, L., Xia, B., Wang, Y., Zhang, W., and Han, Y. (2023). A fine-grained approach for EEG-based emotion recognition using clustering and hybrid deep neural networks. Electronics 12:4717. doi: 10.3390/electronics12234717
Zheng, W.-L., and Lu, B.-L. (2015). Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 7, 162–175. doi: 10.1109/TAMD.2015.2431497
Zhong, M., Yang, Q., Liu, Y., Zhen, B., Zhao, F., and Xie, B. (2023). EEG emotion recognition based on TQWT-features and hybrid convolutional recurrent neural network. Biomed. Signal Process. Control 79:104211. doi: 10.1016/j.bspc.2022.104211
Zhou, Y., and Lian, J. (2023). Identification of emotions evoked by music via spatial-temporal transformer in multi-channel EEG signals. Front. Neurosci. 17:1188696. doi: 10.3389/fnins.2023.1188696
Keywords: music-induced, emotion recognition, artificial intelligence, personalization, applications
Citation: Su Y, Liu Y, Xiao Y, Ma J and Li D (2024) A review of artificial intelligence methods enabled music-evoked EEG emotion recognition and their applications. Front. Neurosci. 18:1400444. doi: 10.3389/fnins.2024.1400444
Edited by:
Dan Zhang, Tsinghua University, ChinaReviewed by:
Jun Jiang, Shanghai Normal University, ChinaWen Li, Universiti Sains Malaysia, Malaysia
Copyright © 2024 Su, Liu, Xiao, Ma and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Dezhao Li, ZGxpYWVAY29ubmVjdC51c3QuaGs=