- 1State Key Laboratory of Cognitive Neuroscience and Learning and IDG/McGovern Institute for Brain Research, Beijing Normal University, Beijing, China
- 2Center for Collaboration and Innovation in Brain and Learning Sciences, Beijing Normal University, Beijing, China
- 3Beijing Key Laboratory of Brain Imaging and Connectomics, Beijing Normal University, Beijing, China
- 4Beijing Institute of Biomedicine, Beijing, China
- 5School of Psychology, Beijing Normal University, Beijing, China
- 6College of Humanities, Jiangxi University of Chinese Medicine, Nanchang, China
- 7Centre for Cognitive and Brain Sciences and Department of Psychology, University of Macau, Taipa, China
Micro-expressions are fleeting and subtle emotional expressions. As they are spontaneous and uncontrollable by one’s mind, micro-expressions are considered an indicator of genuine emotions. Their accurate recognition and interpretation promote interpersonal interaction and social communication. Therefore, enhancing the ability to recognize micro-expressions has captured much attention. In the current study, we investigated the effects of training on micro-expression recognition with a Chinese version of the Micro-Expression Training Tool (METT). Our goal was to confirm whether the recognition accuracy of spontaneous micro-expressions could be improved through training and brain stimulation. Since the right temporal parietal junction (rTPJ) has been shown to be involved in the explicit process of facial emotion recognition, we hypothesized that the rTPJ would play a role in facilitating the recognition of micro-expressions. The results showed that anodal transcranial direct-current stimulation (tDCS) of the rTPJ indeed improved the recognition of spontaneous micro-expressions, especially for those associated with fear. The improved accuracy of recognizing fear spontaneous micro-expressions was positively correlated with personal distress in the anodal group but not in the sham group. Our study supports that the combined use of tDCS and METT can be a viable way to train and enhance micro-expression recognition.
Introduction
The capacity to use emotional signals is vital for navigation in daily life. Identifying rapidly changing facial expressions to monitor others’ intentions and adjusting reactions accordingly is crucial for establishing, developing and maintaining stable interpersonal relationships (Schönenberg et al., 2014; Alguacil et al., 2017; Hubble et al., 2017). Both macro-expressions and micro-expressions exist in human interactions. Macro-expressions, the full facial expression, lasting from 1/2 s to 4 s, are typical and common in daily life. Less often, when people attempt to conceal or repress their emotions, a short micro-expression (usually lasting from 1/25 s to 1/5 s, up to 1/2 s) is embedded in the flow of expressions. Since micro-expressions are automatic and uncontrollable, they are considered to reveal the genuine feelings that individuals attempt to conceal. The authenticity and objectivity of micro-expression recognition processes make them of great value in diverse fields such as forensic investigation, clinical diagnosis, and national security (Ekman, 2009a; Frank et al., 2009; Weinberger, 2010). Likewise, the ability to “read” others’ true emotional states accurately is also advantageous for ordinary people and has been linked to social communication skills. For instance, individuals with higher micro-expression reading accuracy received better evaluations from their supervisors at the workplace (Matsumoto and Hwang, 2011). However, since micro-expressions are very subtle and rapid, recognizing them in real life is a practical challenge (Ekman et al., 1987; Shen et al., 2016).
The Micro-Expression Training Tool (METT) is a definitive training tool used to enhance the micro-expression recognition ability. Previous research has indicated that the METT can significantly increase the accuracy of recognizing micro-expressions. Using METT for less than 40 min improved a respondent’s recognition accuracy by 30–40% (Ekman, 2009b). It is noteworthy that the micro-expression material used in Ekman’s METT for test and practice is artificial dynamic expression made from three consecutive facial images (Matsumoto et al., 2000). However, in real human interactions, spontaneous micro-expression is not only a dynamic but also a continuously changing expression with a short duration and subtle amplitude. Given the vast differences between artificial and spontaneous micro-expressions, the ideal training tool for micro-expression recognition would use spontaneous micro-expression stimuli to achieve good ecological validity. Therefore, the METT used in the present study to assess the training effect included artificial and spontaneous micro-expressions (Chen et al., 2018). Specifically, artificial micro-expressions were selected from POFA, NimStim, and TFEID (Ekman, 1993; Chen and Yen, 2007; Yan et al., 2014, 2013). Spontaneous micro-expressions were real facial micro-expression videos selected from the Chinese Academy of Sciences Micro-Expression (CASME) database (Yan et al., 2014).
Facial expression recognition consists of the implicit perception and explicit recognition processes. Implicit facial expression perception, occurring relatively quickly, can be made with simple and familiar emotional information input and without consciousness. While explicit facial expression recognition is a process of inferring the emotional state based on facial expression cues, and requires comparison between the currently obtained features and related prior knowledge (Wicker et al., 2008; Rellecke et al., 2012; Xiao et al., 2016). In the training section as main part of the METT, individuals can learn how to read other people’s emotions accurately based on facial information. Specifically, a narrator describes unique facial cues to signal emotions and provides explicit examples of dynamic changes in the regions of the eyes, nose, and mouth. An example is the micro-expression of anger in which—”the mouth is sometimes open and sometimes closed, whereas the eyebrows are always pulled down and close to the inside, the upper eyelids deepen, and the lower eyelids shrink, causing the eyes to glare widely.” (Ekman, 2003). According to the different functions of the implicit and explicit facial processes, this method of describing the unique features of various categories of expressions to help trainees better distinguish emotional expressions mainly works on the explicit route of facial expression recognition (Ekman, 2003; Endres and Laidlaw, 2009). Therefore, presumably, the training effect can be enhanced by improving the explicit recognition ability. Existing neuroimaging studies indicated that regions overlapping with the inferior portion of the temporal parietal junction (TPJ) were linked to the identification of facial expressions (Carter and Huettel, 2013). Critchley et al. (2000) investigated the two parallel pathways of facial expression recognition and found that explicit processing activated the temporal cortex while implicit processing evoked the amygdala. Strong structural and functional links between amygdala and temporal parietal regions have been identified (Bickart et al., 2014). A study in macaques revealed that electrical microstimulation of a face-processing region of the temporal lobe activated the amygdala (Moeller et al., 2008). Amygdala played a central role in processing unconscious stimuli through a subcortical route and was implicated in coarse processing of emotion. TPJ was responsible for the attention reorienting/integration and further processing emotional information deeply (Dzafic et al., 2019). Monk et al. (2010) found that when attention bias to emotional faces was equivalent, autism participants had weak connectivity between amygdala-temporal lobe during emotional face processing, which may imply their obstructions to transmission in emotion processing. Olsson and Ochsner (2008) analyzed the role of social cognition in emotion and suggested that the right TPJ (rTPJ) interconnected with visuospatial centers supported externally generated representations and might code cognitive aspects in this way. Qiao et al. (2020) found that anodal HD-tDCS over the rTPJ significantly increased fixation time and fixation count in emotional cue area and better utilized facial expression information to infer emotional states. Meanwhile, other studies further support the involvement of the rTPJ in the explicit facial expression recognition (Kesler-West et al., 2001; Wicker et al., 2008; Atique et al., 2011).
Transcranial direct current stimulation (tDCS) is a non-invasive neuromodulation technique involving the delivery of small direct currents through two surface electrodes on the scalp to modulate cortical excitability in the underlying brain region. The excitability can be increased by anodal stimulation and decreased by cathodal stimulation (Hogeveen et al., 2014; Chrysikou et al., 2017). In this study, we administered the anodal stimulation over the rTPJ and expected to observe an improvement in explicit expression recognition ability — reflected in micro-expression training and measured through the improved accuracy of micro-expression recognition. Specifically, we assumed that the anodal group would demonstrate higher scores after training compared to the sham group. In addition, compared with the artificial micro-expressions synthesized from three static expressions, spontaneous micro-expressions from the real facial micro-expression video depicted a new and unique expression at each time point, and therefore provided more expression information (social information). When dealing with these more complex, natural, and ecologically valid stimuli, a higher level of social cognitive function is required. As the rTPJ was sensitive to processing social information (Tang et al., 2015; Kelley et al., 2021), it is likely to observe a better training effect on spontaneous micro-expression recognition with anodal stimulation. Previous research also demonstrated that the rTPJ played a role in attentional orientation toward threat information (Sagliano et al., 2019). One study indicated increased fear-specific activation in the rTPJ (Zwanzger et al., 2014). Donaldson et al. (2019) discovered that anodal rTPJ high-definition tDCS (HD-tDCS) improved the facial emotion processing performance of fear, and the stimulation effects depended on the intensity and salience/valence (negative/threat) of the emotion. Therefore, it is possible that a stronger effect of training on fearful expression recognition is observed after anodal stimulation.
Previous research highlighted the close relationship between emotion recognition and empathy (Riess et al., 2012; Zhang et al., 2017; Yang et al., 2018). Empathy has been defined as the understanding of their emotions and internal thoughts (cognitive empathy) and sharing the emotional state of others (affective empathy) (Bloom, 2017; Israelashvili et al., 2020). Facial emotion recognition overlapped widely with empathy in terms of the perceptual inputs of emotional empathy (i.e., visual facial emotional stimuli) and the inferential processes of cognitive empathy (Holland et al., 2021). A previous study supported that empathy improved the recognition of facial emotion expressions, such as those expressing fear (Besel and Yuille, 2010). Emotion perception was correlated to affective and cognitive empathy with different effect sizes (Olderbak and Wilhelm, 2017). Moreover, Israelashvili et al. (2020) found that two facets of affective empathy had opposite relationships to facial expression recognition: empathic concern and personal distress were positively and negatively correlated, respectively, with accurate emotion recognition. Interestingly, personal distress, the “self-oriented” feeling of personal anxiety and unease in response to extreme distress in others, was detrimental to expression recognition because of its overly self-focused reaction (Davis, 1980, 1983). However, anodal stimulation over the rTPJ could inhibit self-centered concern and promote other-oriented concern (Payne and Tsakiris, 2017; Martin et al., 2019) that may improve facial expression recognition accuracy and thus may enhance the training effect of the individuals with personal distress. Nevertheless, considering the complex relationship among empathy, rTPJ and facial expression recognition, we conducted exploratory analyses to investigate the role of all empathy traits with the Interpersonal Reactivity Index (IRI) questionnaire in the effect of anodal stimulation over the rTPJ on training effect.
In the current study, we aimed to examine whether targeting tDCS over the rTPJ enhanced the effect of training on recognizing artificial and spontaneous micro-expressions using the Chinese version of METT. We also sought to understand the role of empathy in this process. We expected a significant training effect for both artificial and spontaneous micro-expression by using the Chinese version of METT. The anodal stimulation would enhance the effect of training on micro-expression recognition, especially for fear spontaneous ones. In light of prior findings regarding the role of rTPJ in facilitating the transformation of self to other representation, it was also hypothesized that anodal stimulation would enhance the effect of training in individuals with personal distress.
Materials and Methods
Participants
Using G*Power 3.1.9.4 software, a priori calculation indicated a required minimum sample size of thirty-four participants to detect adequate power (1 – β = 0.8) and medium-sized effect (f = 0.25). Fifty-eight healthy, right-handed, college students took part in our study (mean age = 21.71 ± 2.75 years; 47% males, 53% females). We used a brief self-report questionnaire to obtain information on the participants’ gender, age, and history of psychopathological disorders, including seizure, tumor, stroke, mood disorder, brain surgery or intracranial metal implantation. None of the participants reported a history of neurological or psychiatric disorders. No one dropped out of the experiment. Based on the reported partial η2 value of 0.07 in the current study that corresponds to an effect size f of 0.27, our sample size of 58 participants divided into two groups who each have two measurements would yield an estimated power of 98.25% to detect significant interaction between stimulation and testing stage for spontaneous micro-expression, which was used as the primary outcome measure. All participants signed a written informed consent form and were paid for their participation. This study was approved by the Institutional Review Board of the State Key Laboratory of Cognitive Neuroscience and Learning at Beijing Normal University.
Procedure
All participants made two visits (see Figure 1A). During the first visit, they received different types of tDCS stimulation depending on whether they belonged to the anodal or the sham group. Participants then completed the Chinese version of METT and filled in the IRI questionnaire. Two weeks later, during the second visit, they performed only the Chinese version of METT. Participants were randomly assigned to the anodal (n = 30; 47% males, 53% females) or the sham (n = 28; 46% males, 54% females) group. The group assignment was double-blind by using the “study mode” feature of the DC-Stimulator Plus to ensure that neither the experimenters nor the participants were aware of the actual stimulation being administered.
Figure 1. Experiment design. (A) Experimental procedure. Two visits were made. In the first visit, participants underwent anodal or sham stimulation and then completed the Chinese version of Micro-Expression Training Tool (METT); in the second visit 2 weeks later, they only finished the Chinese version of METT. The Chinese version of METT included five sections, pre-test, training, practice, review, and post-test. In the sections of pre-test and post-test, participants were asked to choose one of eight emotion labels after seeing the stimuli. The stimuli of pre-test included static expressions, artificial and spontaneous micro-expressions. The stimuli of post-test included artificial and spontaneous micro-expressions. (B) Placement of the anodal electrode for the right temporal parietal junction (rTPJ) between P6 and CP6 regions (top row) and the normalized electric field (NormE) derived from electric field modeling calculations using SimNIBS (bottom row). (C) The time series of artificial and spontaneous micro-expression in disgust. Source: L.F. Chen and Y.S. Yen, Taiwanese facial expression image database, Brain Mapping Laboratory, Institute of Brain Science, National Yang-Ming University, 2007.
The Chinese Version of Micro-Expression Training Tool
The Chinese version of METT expands upon the Ekman METT also includes five sections: pre-test, training, practice, review, and post-test. In the pre-test section, participants’ baseline ability of facial expression recognition was measured. The test materials included 84 static expressions with seven basic emotions (twelve each of anger, disgust, fear, happiness, sadness, surprise, and neutral), 14 artificial micro-expressions selected from POFA (Ekman, 1993) (Caucasian models), NimStim (black models), and TFEID (Chen and Yen, 2007) (Asian models) and with the same seven emotions (two each of anger, disgust, fear, happiness, sadness, surprise and neutral), as well as 9 spontaneous micro-expressions, real facial micro-expression videos, selected from the CASME database (Yan et al., 2013, 2014) with five emotions (two each of disgust, happiness, sadness and surprise, as well as one of fear).
The static expression was presented until participants responded. The artificial micro-expression consisted of three successive facial pictures. The target facial picture was sandwiched between two neutral pictures for 120ms. The spontaneous micro-expression lasted more than 5000 ms, but the target spontaneous micro-expression duration was less than 250 ms (1/4 s). To illustrate, the artificial and spontaneous micro-expressions for disgust were highlighted separately in Figure 1C. All materials were presented in random order and balanced for expresser gender, ethnicity, and emotional dimensions. Participants in the pre-test section were asked to choose a single corresponding emotion from the eight emotional labels including anger, disgust, fear, happiness, sadness, surprise, neutral along with an ‘other’ option to help prevent any artifactual agreement. The accuracy of emotion recognition as a dependent variable for this assessment was computed by coding one when participants matched labels successfully and zero when they selected incorrectly, and was expressed as a percentage.
The next section of the training included several Chinese versions of commentary videos in which the narrator emphasized critical facial features and explained how to recognize and distinguish confusing emotions accurately. In the practice section, participants practiced with feedback to ensure they understood and internalized the knowledge and skills learned in the earlier section. The review section repeated the training section. The final post-test followed the same form as the pre-test, but alternative materials were used to assess the ability of artificial and spontaneous micro-expression recognition after training.
All facial images and videos were presented centrally.
Transcranial Direct-Current Stimulation Manipulation
A battery-driven constant current stimulator device (DC-Stimulator Plus, NeuroConn GmbH, Germany) was used to induce tDCS and deliver low amplitude direct current to the scalp via two saline-soaked sponge electrodes. In this experiment, a 35 cm2 anodal electrode (7 cm × 5 cm) was placed between the CP6 and P6 regions (10–20 EEG system, see Figure 1B), which covered the MNI coordinates [54, –59, 22] of the rTPJ reported in previous fMRI studies (Jurcak et al., 2007). The other reference electrode was placed on the left cheek based on previous reports (Hsu et al., 2011; Reinhart et al., 2016; Shires et al., 2020). For the anodal condition, the stimulation with a current intensity of 1.5 mA was delivered continuously for 20 min, and its fade-in and fade-out time were both 15 s. For the sham condition, all settings were identical to the anodal condition except for the duration of real stimulation. For the sham condition, after continuously delivering the current for 30 s, the current was turned off, but the electrodes were kept on the head. This procedure allowed the sham stimulation to be comparable in sensation to the anodal stimulation (Gandiga et al., 2006). At the end of the experiment, when asked to report their impression of whether they had received a real or sham stimulation, all participants believed that the real stimulation was delivered during the experiment.
Simulation of electric field distributions in the brain for the tDCS (see Figure 1B) was performed using the SimNIBS 3.2 (Thielscher et al., 2015) incorporating the template head model included in the software with the following parameters: 1 mm thick rubber electrodes with rectangular connectors encased in 3 mm thick sponges and default tissue connectivity values (white matter: 0.126 S/m, gray matter: 0.275 S/m, cerebrospinal fluid: 1.654 S/m, bone: 0.010 S/m, scalp: 0.465 S/m).
Empathy Trait Assessment
The IRI questionnaire (Davis, 1980) includes 28 items to assess empathy on a 5-point Likert scale ranging from 0 (does not describe me well) to 4 (describe me very well). The IRI consists of four subscales—empathic concern (α = 0.80), personal distress (α = 0.75), perspective taking (α = 0.79), and fantasy (α = 0.82) —and has good validity and reliability. In this scale, empathic concern (EC) measures the degree of warmth, compassion, and care for others; personal distress (PD) assesses self-oriented discomfort caused by stressful interpersonal situations or emergencies; perspective taking (PT) reflects the tendency to adopt the psychological perspectives of others spontaneously; and fantasy (FS) taps the tendency to imaginatively transpose oneself into fictional situations (Davis, 1983).
Statistical Analysis
Descriptive data of the participants from the anodal and sham groups were shown as mean ± SD. Between-group baseline characteristic comparisons were performed using the independent sample t-tests for continuous variables and the chi-square tests for categorical variables. To examine the effect of tDCS on micro-expression training, we separately ran a mixed two-way repeated ANOVA on the accuracy of artificial and spontaneous micro-expression recognition, with stimulation (anodal/sham) as a between-subject variable and testing stage (pre-test/post-test) as a within-subject variable. To investigate the impact of tDCS on micro-expression recognition training of different emotions, we conducted a mixed two-way repeated ANOVA on the improved accuracy of the micro-expression recognition, with stimulation as a between-subject variable and emotion (disgust, fear, happiness, sadness and surprise) as a within-subject variable. Moreover, we also conducted five independent sample t-tests on the improved accuracy of spontaneous micro-expression recognition between two groups to analyze which emotion training effect can be improved using anodal stimulation on exploratory. To explore how tDCS together with empathy traits (empathic concern, personal distress, perspective taking, and fantasy) contributed to the micro-expression training effect, we tested the relationships between empathy traits and the improved accuracy of micro-expression recognition with significant between-group differences by using Spearman rank correlations in each group. The improved accuracy of the micro-expression recognition was calculated by subtracting pre-test accuracy from post-test accuracy. The statistical significance was set at a two-tailed t-test (p < 0.05). Bonferroni adjustments were used to reduce the risk of Type I error where multiple statistical tests were conducted. However, our results didn’t survive Bonferroni’s correction for multiple comparisons with the statistical threshold adjusted to 0.05/5 = 0.01 for multiple independent sample t-tests and 0.05/10 = 0.005 for multiple Spearman rank correlations. Therefore, we complemented the classical frequentist statistical analyses by Bayesian independent t-tests and Bayesian Spearman’s rank correlations in the JASP (version 0.16: JASP Team, 2021) and R (version 3.6.3: R Core Team, 2020; van Doorn et al., 2020). The resulting Bayes factors (BF10) reflect the ratio of the extent to which the data support the hypothesis, rather than its complement. Therefore, BF10 > 1: more support for the hypothesis than for its complement; BF10 ≈ 1: inconclusive; BF10 < 1: more support for the complement of the hypothesis.
Results
Thirty participants (mean age = 21.83 ± 3.35 years, 47% males) and twenty-eight participants (mean age = 21.57 ± 1.95 years, 46% males) were assigned to the anodal and sham groups, respectively. There was no significant difference between the two groups in terms of age [t(56) = 0.37, p = 0.72, Cohen’s d = 0.01] and gender (χ2 = 0.0003, p = 0.99). Furthermore, we conducted t-tests to discern differences between the two groups in baseline characteristics, including empathy traits and primary recognition accuracy of static expressions, artificial micro-expressions and spontaneous micro-expressions, and found no significant differences (ps > 0.18). Baseline characteristics and corresponding statistics for the two stimulation groups are represented in Table 1.
Repeated ANOVA indicated a training effect on artificial micro-expression recognition. The mean accuracy of artificial micro-expression recognition for each condition is displayed in Figure 2A. The main effect of the testing stage was significant for the first visit [F(1,56) = 146.91, pcorrected < 0.001, η2p = 0.72] and second visit [F(1,56) = 21.47, pcorrected < 0.001, η2p = 0.28], indicating that the accuracy of artificial micro-expression recognition was higher post-test than pre-test in both the first and the second visits (first visit: 81.96 ± 12.48 vs. 57.39 ± 13.79; second visit: 80.40 ± 16.53 vs. 69.46 ± 14.77). However, during these two training programs, both the main effect of stimulation [first visit: F(1,56) = 0.99, pcorrected = 0.33, η2p = 0.02; second visit: F(1,56) = 0.01, pcorrected = 0.94, η2p = 0.00] and stimulation × testing stage interaction [first visit: F(1,56) = 1.11, pcorrected = 0.30, η2p = 0.02; second visit: F(1,56) = 0.98, pcorrected = 0.33, η2p = 0.17] did not reach significance.
Figure 2. Mean accuracy of artificial and spontaneous micro-expression recognition in pre-test and post-test for the anodal and sham group. (A) Mean accuracy of artificial micro-expression recognition. (B) Mean accuracy of spontaneous micro-expression recognition. The interaction of stimulation and study was significant in the first visit and marginally significant in the second visit. The anodal group performed better than the sham group after training in the first visit. The anodal group rather than the sham group could improve the accuracy of spontaneous micro-expression recognition through training, even if no stimulation was applied in the second visit. Pre_Anodal and Pre_Sham refer to the mean accuracy of the micro-expression recognition in pre-test for the anodal and sham group. Post_Anodal and Post_Sham refer to the mean accuracy of the micro-expression recognition in post-test for the anodal and sham group. The significance levels are marked by different symbols. “**”, “*”, and “†” indicate p < 0.01, p < 0.05, and p < 0.1 (two-tailed), respectively. Error bars represent standard error of mean (SEM).
Similar training effects were found in spontaneous micro-expression recognition (see Figure 2B). The main effect of the testing stage was significant (first visit: F(1,56) = 23.78, pcorrected < 0.001, η2p = 0.30, post-test accuracy: 53.35 ± 17.25 vs. pre-test accuracy: 40.71 ± 18.01; second visit: F(1,56) = 10.76, pcorrected < 0.01, η2p = 0.16, post-test accuracy: 55.90 ± 16.48 vs. pre-test accuracy: 47.41 ± 15.23). This noteworthy improvement was highly consistent with our observations in the artificial micro-expression recognition test, although it was much more difficult to identify a spontaneous micro-expression correctly. Furthermore, this training effect can be sustained for a significant period, as the accuracy of the pre-test in the second visit after 2 weeks was higher than the baseline created during the first visit [t(57) = 2.40, pcorrected < 0.05]. The main effect of stimulation was not significant [F(1,56) = 2.59, pcorrected = 0.11, η2p = 0.04]. Of note, stimulation × testing stage interaction was significant in the first visit [F(1,56) = 4.17, pcorrected < 0.05, η2p = 0.07]. A simple effect analysis showed a higher accuracy of spontaneous micro-expression recognition for the anodal condition (M = 58.80, SD = 15.74) compared with the sham condition (M = 47.52, SD = 17.14) in the post-test (pcorrected < 0.05). Meanwhile, there was no significant difference between the anodal (M = 41.11, SD = 18.94) and sham (M = 40.28, SD = 17.28) groups in the pre-test (pcorrected = 0.86). We also found marginally significant interaction in the second visit [F(1,56) = 3.11, pcorrected = 0.08, η2p = 0.05]. The simple effect analysis showed an improvement in the accuracy of spontaneous micro-expression recognition through training in the anodal group even if no stimulation was applied in the second visit (anodal group: post-test accuracy: 59.11 ± 15.45 vs. pre-test accuracy: 46.30 ± 17.04, pcorrected < 0.01; sham group: post-test accuracy: 52.47 ± 17.12 vs. pre-test accuracy: 48.61 ± 13.23, pcorrected = 0.30). As all the results until this stage showed no stimulation effect from training on artificial micro-expression, we did not consider it in the next analysis.
To investigate the impact of tDCS on micro-expression recognition training of different emotions, we conducted a mixed two-way repeated ANOVA on the improved accuracy of the micro-expression recognition. For the first visit, the main effect of stimulation was significant, F(1,56) = 5.11, pcorrected < 0.05, η2p = 0.08. The main effect of emotion was marginally significant, F(4,53) = 2.26, pcorrected = 0.08, η2p = 0.15. Stimulation × emotion interaction was not significant, F(4,53) = 1.24, pcorrected = 0.31, η2p = 0.09. For the second visit, the main effect of stimulation was marginally significant, F(1,56) = 3.35, pcorrected = 0.07, η2p = 0.06. Neither the main effect of emotion [F(4,53) = 0.41, pcorrected = 0.80, η2p = 0.03] nor the interaction between stimulation and emotion [F(4,53) = 0.65, pcorrected = 0.63, η2p = 0.05] was significant. Furthermore, we also conducted five independent sample t-tests (classical frequentist statistical analysis) and Bayesian independent t-tests on the improved accuracy of spontaneous micro-expression recognition between two groups to analyze which emotion training effect can be improved using anodal stimulation on exploratory. The results showed the significantly higher improved accuracy in the anodal group for only the spontaneous micro-expression of fear [t(56) = 2.20, puncorrected < 0.05, Cohen’s d = 0.58, BF10 = 1.9] and sadness [t(56) = 2.00, puncorrected = 0.051, Cohen’s d = 0.525, BF10 = 1.4] in the first visit (see Figure 3). We found no significant difference between the two groups for the improved accuracy of spontaneous micro-expressions of disgust [t(56) = −0.71, puncorrected = 0.48, Cohen’s d = −0.19, BF10 = 0.33], happiness [t(56) = 1.13, puncorrected = 0.26, Cohen’s d = 0.30, BF10 = 0.46], and surprise [t(56) = 1.24, puncorrected = 0.22, Cohen’s d = 0.32, BF10 = 0.50] in the first visit. We didn’t find similar significant results in the second visit (ps > 0.28, BF10 < 0.44).
Figure 3. Mean improved accuracy on distinct emotions of spontaneous micro-expressions for anodal and sham group in the first visit. Comparisons of the differences of distinct emotions between the two groups showed that the anodal group improved more than the sham group on spontaneous micro-expressions of fear and sadness. The significance levels are marked by different symbols. “*” and “†” indicate p < 0.05 and p < 0.1 (two-tailed), respectively. Error bars represent standard error of mean (SEM).
In addition, we tested whether empathy played a role in the effect of tDCS on the improved accuracy of spontaneous micro-expression recognition. Spearman rank correlations were performed separately in the anodal and sham groups to examine the relationship between empathy traits and the improved accuracy of overall spontaneous micro-expressions and the spontaneous micro-expressions of fear and sadness. We found different relationships between the improved accuracy of fear spontaneous micro-expressions and personal distress, with a positive correlation in the anodal group (Spearman’s rho = 0.43, puncorrected = 0.02, BF10 = 3.20) but no correlation in the sham group (Spearman’s rho = 0.13, puncorrected = 0.50, BF10 = 0.43). There was no significant correlation for other empathy traits (empathic concern, perspective taking, and fantasy) in the two groups (ps > 0.11, BF10 < 0.88). Furthermore, no significant correlation was found between empathy traits and the improved accuracy of overall spontaneous micro-expression or spontaneous micro-expression of sadness in the anodal or sham group (ps > 0.16, BF10 < 0.43) (see Table 2).
Table 2. Correlations between empathic traits and mean improved accuracy of overall spontaneous micro-expressions, and mean improved accuracy of spontaneous micro-expression of fear and sadness in the anodal and sham group.
Discussion
Training Effect for Both Artificial and Spontaneous Micro-Expression
In the present study, we administered anodal stimulation over the rTPJ and used the Chinese version of METT to investigate the combined effect on artificial and spontaneous micro-expressions as well as the role of empathy in this process. We first expected a significant training effect for both artificial and spontaneous micro-expression using the Chinese version of METT. Results revealed that the accuracy of the participants’ detection of artificial and spontaneous micro-expressions improved after completing the training. Furthermore, the learning efficiency could be observed 2 weeks after training. Results were consistent with previous studies. Hurley’s (2012) randomly assigned participants who completed the METT detected micro-expression more accurately than those in a control group. Russell et al. (2006) determined that the ability to recognize emotional facial expressions and micro-expressions improved significantly in patients with schizophrenia using METT. Besides, concomitant changes in visual attention on facial emotion sustained and could be observed 1 week after METT training (Russell et al., 2008). Hurley et al. (2014) showed that learning efficiency could be retained 6−20 months after training. However, the learning effects could not be expected to last for such a long time in the present study, because there was only a 2-week duration between two visits in the experiment. As a whole, we believed that as a training tool, METT had a training effect not only on artificial micro-expressions but also on spontaneous micro-expressions, and this training effect could last for sustained periods.
Enhanced Effect for Spontaneous Rather Than Artificial Micro-Expression
As hypothesized, there were stark differences in the training effect between spontaneous and artificial micro-expressions under anodal stimulation over the rTPJ. Our data suggested that the stimulation over the rTPJ was associated with an enhanced micro-expression recognition training effect, but only for spontaneous (and not artificial) micro-expressions. The artificial micro-expression in Ekman’s METT was artificially synthesized by presenting an emotional expression with both a forward and backward neutral expression. However, the spontaneous micro-expression in our study was presented via the video recording that showed the complete spontaneous process from the formation to the changes of the facial micro-expression. The former was dynamic but discrete, whereas the latter was continuous and more in line with the actual social context in which the rTPJ played an important role (Loscalzo, 2011; Tang et al., 2015). Moreover, the emotions of the expressions sandwiched between the two neutral expression pictures in the artificial micro-expressions were basic, obvious, and intense, while the emotions of spontaneous micro-expressions were suppressed and partial though also basic. The facial cues provided by artificial micro-expressions were simpler and more evident than those generated by spontaneous micro-expressions. Therefore, the recognition of artificial micro-expression was more like the response to the familiar and rapid information, in which implicit perception processing was a predominant driver (Rellecke et al., 2012). Given we focused on enhancing explicit identification ability to improve the training effect in the present study, anodal stimulation had a feeble effect. Furthermore, in contrast to artificial micro-expressions, the facial information presented by spontaneous micro-expressions was subtle and ambiguous. Therefore, successful recognition of the spontaneous micro-expressions required the higher capability to attribute a person’s intentions and emotional state based on current cues. Additionally, decades of research provide evidence that the rTPJ played a specific role in this ability, namely ToM, which incorporated the information received into a context to represent and interpret the ideas of others (Mitchell, 2008; Carter and Huettel, 2013; Mai et al., 2016). Therefore, anodal stimulation promoted the training effect of spontaneous micro-expression recognition. Above all, anodal stimulation over the rTPJ could enhance the training effect of spontaneous micro-expressions rather than artificial micro-expressions.
Enhanced Specific Emotions Under Right Temporal Parietal Junction in People With Different Empathy Traits
We found significantly higher improved accuracy of fear spontaneous micro-expression in the anodal group than in the sham group. Fear is a visual cue associated with fast-approaching danger and requires an immediate response so that fear always attracts salient attention (Jack et al., 2014). The rTPJ engaged in attentional orienting toward threat information could contribute to the increased training effect of stimulation on the spontaneous micro-expressions of fear (Sagliano et al., 2019). This finding was consistent with previous research. Donaldson et al. (2019) found that the rTPJ anodal HD-tDCS could influence fear facial emotion recognition. Besides, they suggested that the rTPJ involvement in facial emotion processing depended on the intensity and salience/valence (negativity/threat) of the emotion. Surprisingly, we also found an increased training effect from anodal stimulation over the rTPJ for the spontaneous micro-expression of sadness. When recognizing expressions of sadness, eye cues had greater significance than other cues and tended to capture most attentional resources (Eisenbarth and Alpers, 2011). rTPJ was proved to be linked to eye gaze and sensitive to eye information (Kelley et al., 2021), and might therefore lead to an improved effect of training on spontaneous micro-expressions of sadness.
We also explored the role of empathy in the effect of tDCS over the rTPJ on micro-expression training. We found different relationships between the improved accuracy of spontaneous micro-expressions of fear and personal distress, with a positive correlation in the anodal group but no correlation in the sham group. Personal distress is an index of the individual’s own feelings of fear, apprehension, and discomfort when they witness the negative experiences of others. Personal distress may be positively associated with sensitivity to dangerous signals, but it is relatively more self-focused that it does not contribute to and is even adverse to emotional reaction (Davis, 1980, 1983). Therefore, personal distress does not influence the training effect in the sham group. Based on the development pattern of empathic tendencies set forth by Hoffman, personal distress seems to be an automatic imitation of the state observed in those who suffer from a disaster. It appears to be the preliminary empathy phase in early development, a phase in which children cannot differentiate between themselves and others effectively. As self-centered empathic distress is transformed into other-oriented concern, the ability to apprehend others’ mental and emotional states develops (Hoffman, 1976). Existing neuroimaging findings suggested that the rTPJ region played a crucial role in the mechanism that enabled the control of automatic imitative responses and allowed representation of the self to be inhibited and that of others to be enhanced (Brass et al., 2009; Spengler et al., 2009). Furthermore, causal evidence for the role of the rTPJ in self-other control was derived from studies measuring the effects of magnetic or electric stimulation over this area. Repetitive transcranial magnetic stimulation (rTMS) targeted to the rTPJ has been shown to be involved in switching between self and other representations (Sowden and Catmur, 2015). Additionally, anodal tDCS over the rTPJ could inhibit self-centered concern and promote other-oriented concern, facilitating the transformation of self to other representations (Sowden et al., 2015). Therefore, in the current study, the anodal stimulation over the rTPJ suppressed self-centered empathic distress and promoted other-oriented concern to improve the ability to read the emotional states of others, which was embodied in the enhanced effect of training on fear spontaneous micro-expressions.
Limitations and Future Directions
This study has several limitations and there are some possible directions for further study. One limitation was that emotional dimensions between artificial and spontaneous micro-expression stimuli were inconsistent. Spontaneous micro-expressions of anger were lacking compared to artificial micro-expressions because of unsatisfactory labels and the difficulty in eliciting the emotion in a laboratory context (Yan et al., 2014). Furthermore, although the study was motivated based on the specific role of the rTPJ in emotional recognition, it couldn’t be denied that the anodal stimulation activated other brain regions that might play a similar or synergistic role. Therefore, further work is required to establish a more refined neural mechanism of micro-expression training using tDCS combined with fMRI. Besides, in the current study, we applied anodal stimulation over the rTPJ that played an important role in the explicit process of facial emotion recognition to enhance micro-expression training effect. However, an issue that was not addressed in this study was whether micro-expression recognition could be improved by manipulating the implicit perception process of facial emotion recognition. Micro-expressions are presented as brief durations, so judgments are thought to rely in part on intuitive, biologically based emotion processing. Therefore, the causal relationship between perception and emotional recognition is also in need of further investigation (Svetieva and Frank, 2016). In addition, other individual characteristics, such as “the Big Five,” that may contribute to modifying the effect of tDCS over the right TPJ on micro-expression training, are a worthy focus for future studies. Moreover, in a meta-analysis across studies including populations characterized as psychopathic, abusive, conduct-disordered, aggressive, unsocialized, criminal and antisocial traits/behaviors were most consistently associated with deficits in recognizing the facial expression of fear. In a future study, we plan to explore the training effect of the combined use of tDCS over the rTPJ and the Chinese version of METT in clinical populations.
Conclusion
This study provided convincing evidence that METT can improve both artificial and spontaneous micro-expression recognition. Anodal stimulation over the rTPJ was associated with an enhanced micro-expression recognition training effect, particularly for spontaneous micro-expressions related to fear. The improved accuracy of recognizing fear spontaneous micro-expressions was positively correlated with personal distress in the anodal group but not in the sham group. The results are promising but preliminary. The combined use of tDCS and METT can be a promising method for teaching and enhancing micro-expression recognition.
Data Availability Statement
The original contributions presented in this study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.
Ethics Statement
The studies involving human participants were reviewed and approved by the Institutional Review Board of the State Key Laboratory of Cognitive Neuroscience and Learning at Beijing Normal University. The patients/participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.
Author Contributions
CL conceived the study and acquired the funding. RS, ZL, JL, and ST conducted the experiments with the help of YG and XS. RS analyzed the data with critical comments from CL. RS and YG wrote the manuscript with help from CL, HW, and XS. All authors helped to refine the final experimental design and commented on the manuscript.
Funding
This work was supported by the Open Research Fund of the State Key Laboratory of Cognitive Neuroscience and Learning (CNLYB2002), the National Natural Science Foundation of China (NSFC) (31871094, 31960180, and U1736125), the Major Project of National Social Science Foundation (19ZDA363), and the Jiangxi University of Chinese Medicine Science and Technology Innovation Team Development Program (CXTD22016).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Alguacil, S., Madrid, E., Espín, A. M., and Ruz, M. (2017). Facial identity and emotional expression as predictors during economic decisions. Cogn. Affect. Behav. Neurosci. 17, 315–329. doi: 10.3758/s13415-016-0481-9
Atique, B., Erb, M., Gharabaghi, A., Grodd, W., and Anders, S. (2011). Task-specific activity and connectivity within the mentalizing network during emotion and intention mentalizing. NeuroImage 55, 1899–1911. doi: 10.1016/j.neuroimage.2010.12.036
Besel, L. D. S., and Yuille, J. C. (2010). Individual differences in empathy: the role of facial expression recognition. Personali. Indiv. Diff. 49, 107–112. doi: 10.1016/j.paid.2010.03.013
Bickart, K. C., Dickerson, B. C., and Barrett, L. F. (2014). The amygdala as a hub in brain networks that support social life. Neuropsychologia 63, 235–248. doi: 10.1016/j.neuropsychologia.2014.08.013
Bloom, P. (2017). Empathy and its discontents. Trends Cogn. Sci. 21, 24–31. doi: 10.1016/j.tics.2016.11.004
Brass, M., Ruby, P., and Spengler, S. (2009). Inhibition of imitative behaviour and social cognition. Philos. Trans. R. Soc. B Biol. Sci. 364, 2359–2367. doi: 10.1098/rstb.2009.0066
Carter, R. M., and Huettel, S. A. (2013). A nexus model of the temporal–parietal junction. Trends Cogn. Sci. 17, 328–336. doi: 10.1016/j.tics.2013.05.007
Chen, L., Shen, X., Yang, H., and Sui, H. (2018). “Micro expression recognition training in college students,” in Proceeding of the 2018 First Asian Conference on Affective Computing and Intelligent Interaction (ACII Asia), (IEEE), 1–3.
Chen, L.-F., and Yen, Y.-S. (2007). Taiwanese Facial Expression Image Database. Taipei, Taiwan: Brain Mapping Laboratory, Institute of Brain Science, National Yang-Ming University.
Chrysikou, E. G., Gorey, C., and Aupperle, R. L. (2017). Anodal transcranial direct current stimulation over right dorsolateral prefrontal cortex alters decision making during approach-avoidance conflict. Soc. Cogn. Affect. Neurosci. 12, 468–475. doi: 10.1093/scan/nsw140
Critchley, H., Daly, E., Phillips, M., Brammer, M., Bullmore, E., Williams, S., et al. (2000). Explicit and implicit neural mechanisms for processing of social information from facial expressions: a functional magnetic resonance imaging study. Hum. Brain Mapp. 9, 93–105. doi: 10.1002/(SICI)1097-0193(200002)9:2<93::AID-HBM4<3.0.CO;2-Z
Davis, M. H. (1980). A multidimensional approach to individual differences in empahty. Catal. Selected Doc. Psychol. 40, 3480. doi: 10.1207/s15327906mbr0302_7
Davis, M. H. (1983). Measuring individual differences in empathy: evidence for a multidimensional approach. J. Personali. Soc. Psychol. 44, 113–126. doi: 10.1037/0022-3514.44.1.113
Donaldson, P. H., Kirkovski, M., Rinehart, N. J., and Enticott, P. G. (2019). A double-blind HD-tDCS/EEG study examining right temporoparietal junction involvement in facial emotion processing. Soc. Neurosci. 14, 681–696. doi: 10.1080/17470919.2019.1572648
Dzafic, I., Oestreich, L., Martin, A. K., Mowry, B., and Burianová, H. (2019). Stria terminalis, amygdala, and temporoparietal junction networks facilitate efficient emotion processing under expectations. Hum. Brain Mapp. 40, 5382–5396. doi: 10.1002/hbm.24779
Eisenbarth, H., and Alpers, G. W. (2011). Happy mouth and sad eyes: scanning emotional facial expressions. Emotion 11, 860–865. doi: 10.1037/a0022758
Ekman, P. (2009b). Telling Lies: Clues to Deceit in the Marketplace, Politics, and Marriage (Revised Edition). New York, NY: WW Norton & Company.
Ekman, P., Friesen, W. V., O’sullivan, M., Chan, A., Diacoyanni-Tarlatzis, I., Heider, K., et al. (1987). Universals and cultural differences in the judgments of facial expressions of emotion. J. Personali. Soc. Psychol. 53:712.
Endres, J., and Laidlaw, A. (2009). Micro-expression recognition training in medical students: a pilot study. BMC Med. Educ. 9:47. doi: 10.1186/1472-6920-9-47
Frank, M., Herbasz, M., Sinuk, K., Keller, A., Kurylo, A., and Nolan, C. (2009). “I see how you feel: training laypeople and professionals to recognize fleeting emotions,” in Proceeding of the Annual Meeting of the International Communication Association.
Gandiga, P. C., Hummel, F. C., and Cohen, L. G. (2006). Transcranial DC stimulation (tDCS): a tool for double-blind sham-controlled clinical studies in brain stimulation. Clin. Neurophysiol. 117, 845–850. doi: 10.1016/j.clinph.2005.12.003
Hoffman, M. L. (1976). “Empathy, role-taking, guilt, and development of altruistic motives,” in Moral Development and Behavior: Theory, Research, and Social Issues, ed. T. Lickona (New York, NY: Holt, Rinehart, & Winston).
Hogeveen, J., Obhi, S. S., Banissy, M. J., Santiesteban, I., Press, C., Catmur, C., et al. (2014). Task-dependent and distinct roles of the temporoparietal junction and inferior frontal cortex in the control of imitation. Soc. Cogn. Affect. Neurosci. 10, 1003–1009. doi: 10.1093/scan/nsu148
Holland, A. C., O’Connell, G., and Dziobek, I. (2021). Facial mimicry, empathy, and emotion recognition: a meta-analysis of correlations. Cogn. Emot. 35, 150–168. doi: 10.1080/02699931.2020.1815655
Hsu, T. Y., Tseng, L. Y., Yu, J. X., Kuo, W. J., Hung, D. L., Tzeng, O. J. L., et al. (2011). Modulating inhibitory control with direct current stimulation of the superior medial frontal cortex. NeuroImage 56, 2249–2257. doi: 10.1016/j.neuroimage.2011.03.059
Hubble, K., Daughters, K., Manstead, A. S. R., Rees, A., Thapar, A., and van Goozen, S. H. M. (2017). Oxytocin reduces face processing time but leaves recognition accuracy and eye-gaze unaffected. J. Int. Neuropsychol. Soc. 23, 23–33. doi: 10.1017/S1355617716000886
Hurley, C. M. (2012). Do you see what i see? Learning to detect micro expressions of emotion. Motivat. Emot. 36, 371–381. doi: 10.1007/s11031-011-9257-2
Hurley, C. M., Anker, A. E., Frank, M. G., Matsumoto, D., and Hwang, H. C. (2014). Background factors predicting accuracy and improvement in micro expression recognition. Motivat. Emot. 38, 700–714. doi: 10.1007/s11031-014-9410-9
Israelashvili, J., Sauter, D., and Fischer, A. (2020). Two facets of affective empathy: concern and distress have opposite relationships to emotion recognition. Cogn. Emot. 34, 1112–1122. doi: 10.1080/02699931.2020.1724893
Jack, R. E., Garrod, O. G. B., and Schyns, P. G. (2014). Dynamic facial expressions of emotion transmit an evolving hierarchy of signals over time. Curr. Biol. 24, 187–192. doi: 10.1016/j.cub.2013.11.064
Jurcak, V., Tsuzuki, D., and Dan, I. (2007). 10/20, 10/10, and 10/5 systems revisited: their validity as relative head-surface-based positioning systems. NeuroImage 34, 1600–1611. doi: 10.1016/j.neuroimage.2006.09.024
Kelley, M. S., Noah, J. A., Zhang, X., Scassellati, B., and Hirsch, J. (2021). Comparison of human social brain activity during eye-contact with another human and a humanoid robot. Front. Robot. AI 7:1–11. doi: 10.3389/frobt.2020.599581
Kesler-West, M. L., Andersen, A. H., Smith, C. D., Avison, M. J., Davis, C. E., Kryscio, R. J., et al. (2001). Neural substrates of facial emotion processing using fMRI. Cogn. Brain Res. 11, 213–226. doi: 10.1016/S0926-6410(00)00073-2
Loscalzo, D. E. (2011). A nexus model of the temporal-parietal junction. Bone 23, 1–7. doi: 10.1016/j.tics.2013.05.007.A
Mai, X., Zhang, W., Hu, X., Zhen, Z., Xu, Z., Zhang, J., et al. (2016). Using tDCS to explore the role of the right temporo-parietal junction in theory of mind and cognitive empathy. Front. Psychol. 7:1–7. doi: 10.3389/fpsyg.2016.00380
Martin, A. K., Huang, J., Hunold, A., and Meinzer, M. (2019). Dissociable roles within the social brain for self-other processing: a HD-tDCS study. Cerebral Cortex 29, 3642–3654. doi: 10.1093/cercor/bhy238
Matsumoto, D., and Hwang, H. S. (2011). Evidence for training the ability to read microexpressions of emotion. Motiv. Emot. 35, 181–191. doi: 10.1007/s11031-011-9212-2
Matsumoto, D., LeRoux, J., Wilson-Cohn, C., Raroque, J., Kooken, K., Ekman, P., et al. (2000). A new test to measure emotion recognition ability: matsumoto and ekman’s Japanese and caucasian brief affect recognition test (JACBART). J. Nonverbal Behav. 24, 179–209. doi: 10.1023/A:1006668120583
Mitchell, J. P. (2008). Activity in right temporo-parietal junction is not selective for theory-of-mind. Cerebral Cortex 18, 262–271. doi: 10.1093/cercor/bhm051
Moeller, S., Freiwald, W. A., and Tsao, D. Y. (2008). Patches with links: a unified system for processing faces in the macaque temporal lobe. Science 320, 1355–1359. doi: 10.1126/science.1157436
Monk, C. S., Weng, S. J., Wiggins, J. L., Kurapati, N., Louro, H. M. C., Carrasco, M., et al. (2010). Neural circuitry of emotional face processing in autism spectrum disorders. J. Psychiatry Neurosci. 35, 105–114. doi: 10.1503/jpn.090085
Olderbak, S., and Wilhelm, O. (2017). Emotion perception and empathy: an individual differences test of relations. Emotion 17, 1092s–1106s. doi: 10.1037/emo0000308
Olsson, A., and Ochsner, K. N. (2008). The role of social cognition in emotion. Trends Cogn. Sci. 12, 65–71. doi: 10.1016/j.tics.2007.11.010
Payne, S., and Tsakiris, M. (2017). Anodal transcranial direct current stimulation of right temporoparietal area inhibits self-recognition. Cogn. Affect. Behav. Neurosci. 17, 1–8. doi: 10.3758/s13415-016-0461-0
Qiao, Y., Hu, Q., Xuan, R., Guo, Q., Ge, Y., Chen, H., et al. (2020). High-definition transcranial direct current stimulation facilitates emotional face processing in individuals with high autistic traits: a sham-controlled study. Neurosci. Lett. 738:135396. doi: 10.1016/j.neulet.2020.135396
R Core Team (2020). R: A Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing.
Reinhart, R. M. G., Xiao, W., McClenahan, L. J., and Woodman, G. F. (2016). Electrical stimulation of visual cortex can immediately improve spatial vision. Curr. Biol. 26, 1867–1872. doi: 10.1016/j.cub.2016.05.019
Rellecke, J., Sommer, W., and Schacht, A. (2012). Does processing of emotional facial expressions depend on intention? Time-resolved evidence from event-related brain potentials. Biol. Psychol. 90, 23–32. doi: 10.1016/j.biopsycho.2012.02.002
Riess, H., Kelley, J. M., Bailey, R. W., Dunn, E. J., and Phillips, M. (2012). Empathy training for resident physicians: a randomized controlled trial of a neuroscience-informed curriculum. J. Gen. Int. Med. 27, 1280–1286. doi: 10.1007/s11606-012-2063-z
Russell, T. A., Chu, E., and Phillips, M. L. (2006). A pilot study to investigate the effectiveness of emotion recognition remediation in schizophrenia using the micro-expression training tool. Br. J. Clin. Psychol. 45, 579–583. doi: 10.1348/014466505X90866
Russell, T. A., Green, M. J., Simpson, I., and Coltheart, M. (2008). Remediation of facial emotion perception in schizophrenia: concomitant changes in visual attention. Schizop. Res. 103, 248–256. doi: 10.1016/j.schres.2008.04.033
Sagliano, L., Vela, M., Trojano, L., and Conson, M. (2019). The role of the right premotor cortex and temporo-parietal junction in defensive responses to visual threats. Cortex 120, 532–538. doi: 10.1016/j.cortex.2019.08.005
Schönenberg, M., Christian, S., Gaußer, A.-K., Mayer, S. V., Hautzinger, M., and Jusyte, A. (2014). Addressing perceptual insensitivity to facial affect in violent offenders: first evidence for the efficacy of a novel implicit training approach. Psychol. Med. 44, 1043–1052. doi: 10.1017/S0033291713001517
Shen, X., Wu, Q., Zhao, K., and Fu, X. (2016). Electrophysiological evidence reveals differences between the recognition of microexpressions. Macroexpressions 7, 1–11. doi: 10.3389/fpsyg.2016.01346
Shires, J., Carrasco, C., and Berryhill, M. E. (2020). No tDCS augmented working memory training benefit in undergraduates rewarded with course credit. Brain Stimulat. 13, 1524–1526. doi: 10.1016/j.brs.2020.08.015
Sowden, S., and Catmur, C. (2015). The role of the right temporoparietal junction in the control of imitation. Cerebral Cortex 25, 1107–1113. doi: 10.1093/cercor/bht306
Sowden, S., Wright, G. R. T., Banissy, M. J., Catmur, C., and Bird, G. (2015). Transcranial current stimulation of the temporoparietal junction improves lie detection. Curr. Biol. 25, 2447–2451. doi: 10.1016/j.cub.2015.08.014
Spengler, S., Von Cramon, D. Y., and Brass, M. (2009). Control of shared representations relies on key processes involved in mental state attribution. Hum. Brain Mapp. 30, 3704–3718. doi: 10.1002/hbm.20800
Svetieva, E., and Frank, M. G. (2016). Empathy, emotion dysregulation, and enhanced microexpression recognition ability. Motivat. Emot. 40, 309–320. doi: 10.1007/s11031-015-9528-4
Tang, H., Mai, X., Wang, S., Zhu, C., Krueger, F., and Liu, C. (2015). Interpersonal brain synchronization in the right temporo-parietal junction during face-to-face economic exchange. Soc. Cogn. Affect. Neurosci. 11, 23–32. doi: 10.1093/scan/nsv092
Thielscher, A., Antunes, A., and Saturnino, G. B. (2015). “Field modeling for transcranial magnetic stimulation: a useful tool to understand the physiological effects of TMS,” in Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 222–225. doi: 10.1109/EMBC.2015.7318340
van Doorn, J., Ly, A., Marsman, M., and Wagenmakers, E. J. (2020). Bayesian rank-based hypothesis testing for the rank sum test, the signed rank test, and Spearman’s ρ. J. Appl. Statist. 47, 2984–3006. doi: 10.1080/02664763.2019.1709053
Weinberger, S. (2010). Airport security: intent to deceive? Nat. News 465, 412–415. doi: 10.1038/465412a
Wicker, B., Fonlupt, P., Hubert, B., Tardif, C., Gepner, B., and Deruelle, C. (2008). Abnormal cerebral effective connectivity during explicit emotional processing in adults with autism spectrum disorder. Soc. Cogn. Affect. Neurosci. 3, 135–143. doi: 10.1093/scan/nsn007
Xiao, R., Li, X., Li, L., and Wang, Y. (2016). Can we distinguish emotions from faces? Investigation of implicit and explicit processes of peak facial expressions. Front. Psychol. 7:1–13. doi: 10.3389/fpsyg.2016.01330
Yan, W., Wu, Q., Liu, Y., Wang, S., and Fu, X. (2013). “Casme database: a dataset of spontaneous micro-expressions collected from neutralized faces,” in Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, 1–7. doi: 10.1109/FG.2013.6553799
Yan, W.-J., Li, X., Wang, S.-J., Zhao, G., Liu, Y.-J., Chen, Y.-H., et al. (2014). CASME II: an improved spontaneous micro-expression database and the baseline evaluation. PLoS One 9:e86041. doi: 10.1371/journal.pone.0086041
Yang, C.-C., Khalifa, N., and Völlm, B. (2018). The effects of repetitive transcranial magnetic stimulation on empathy: a systematic review and meta-analysis. Psychol. Med. 48, 737–750. doi: 10.1017/S003329171700232X
Zhang, X., Chen, L., Zhong, Z., Sui, H., and Shen, X. (2017). “The effects of the micro-expression training on empathy in patients with schizophrenia,” in International Conference on Man-Machine-Environment System Engineering, eds X. Zhang, L. Chen, Z. Zhong, H. Sui, and X. Shen (Singapore: Springer), 189–194. doi: 10.1007/978-981-10-6232-2_23
Zwanzger, P., Steinberg, C., Rehbein, M. A., Br??ckelmann, A. K., Dobel, C., Zavorotnyy, M., et al. (2014). Inhibitory repetitive transcranial magnetic stimulation (rTMS) of the dorsolateral prefrontal cortex modulates early affective processing. NeuroImage 101, 193–203. doi: 10.1016/j.neuroimage.2014.07.003
Keywords: transcranial direct current stimulation, right temporal parietal junction, micro-expression training, artificial micro-expression, spontaneous micro-expression
Citation: Ge Y, Su R, Liang Z, Luo J, Tian S, Shen X, Wu H and Liu C (2022) Transcranial Direct Current Stimulation Over the Right Temporal Parietal Junction Facilitates Spontaneous Micro-Expression Recognition. Front. Hum. Neurosci. 16:933831. doi: 10.3389/fnhum.2022.933831
Received: 01 May 2022; Accepted: 21 June 2022;
Published: 08 July 2022.
Edited by:
Changming Wang, Capital Medical University, ChinaReviewed by:
XiXi Zhao, Capital Medical University, ChinaTong Chen, Southwest University, China
Haining Liu, Chengde Medical University, China
Copyright © 2022 Ge, Su, Liang, Luo, Tian, Shen, Wu and Liu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Xunbing Shen, c2hlbnh1bmJpbmdAeWFob28uY29t; Haiyan Wu, aGFpeWFud3VAdW0uZWR1Lm1v; Chao Liu, bGl1Y2hhb0BibnUuZWR1LmNu
†These authors have contributed equally to this work