- 1Department of Psychology, Korea University, Seoul, South Korea
- 2Department of Psychiatry, Ilsan-Paik Hospital, Inje University, Goyang, South Korea
- 3Department of Psychology, Chonnam National University, Gwangju, South Korea
Individuals with schizophrenia show a reduced ability to integrate facial and vocal information in emotion perception. Although emotion perception has been a target for treatment, no study has yet examined the effect of multimodal training on emotion perception in schizophrenia. In the present study, we developed an audiovisual emotion perception training and test in which a voice and a face were simultaneously presented, and subjects were asked to judge whether the emotions of the voice and the face matched. The voices were either angry or happy, and the faces were morphed on a continuum ranging from angry to happy. Sixteen patients with schizophrenia participated in six training sessions and three test sessions (i.e., pre-training, post-training, and generalization). Eighteen healthy controls participated only in pre-training test session. Prior to training, the patients with schizophrenia performed significantly worse than did the controls in the recognition of anger; however, following the training, the patients showed a significant improvement in recognizing anger, which was maintained and generalized to a new set of stimuli. The patients also improved the recognition of happiness following the training, but this effect was not maintained or generalized. These results provide preliminary evidence that a multimodal, audiovisual training may yield improvements in anger perception for patients with schizophrenia.
Introduction
Impaired emotion perception in schizophrenia has been well-documented. Individuals with schizophrenia have deficits in identifying or discriminating emotions in others from facial expressions (1–3) and tone of the voice (4, 5). Furthermore, they have a reduced ability to integrate facial (visual) and vocal (auditory) emotional information. Specifically, patients with schizophrenia have been shown to not improve to the extent that healthy controls did in emotion recognition tasks when congruent audiovisual stimuli were presented (i.e., face and voice combined) compared to when audio or visual stimuli were presented alone (6, 7), and performed worse than healthy controls in the audiovisual condition [(8–10), cf. 11]. Deficits in the ability to perceive others' emotions are likely trait-like and serve as a risk marker for schizophrenia (12) and remain stable over the course of the illness (13). Moreover, these deficits have an impact on functional outcomes, such as social problem solving, social skills, and community functioning (14, 15). Therefore, addressing emotion perception in schizophrenia has been a target for treatment (16–18).
The most common intervention for emotion perception in schizophrenia is social-cognitive remediation based on Social Cognition and Interaction Training (19), which is a 24-session therapist-led group treatment targeting multiple domains of social cognition including emotion processing, social perception, theory of mind, and attribution bias. More focused intervention on emotion perception has also been developed and the most widely used program is Training of Affect Recognition (20). This program involves a 12-session manualized group training designed to improve facial emotion recognition by using strategies such as verbalization of typical features of emotional faces. More recently, emotion perception in a non-verbal, social context has been gauged by gesture behaviors in schizophrenia patients (21). Although these psychosocial interventions have yielded positive effects on emotion perception (22–24), no study has yet examined the effects of multimodal training on emotion perception in schizophrenia. Given that people with schizophrenia show diminished multimodal integration of facial and vocal emotional stimuli, and as emotion perception in natural circumstances typically requires simultaneous processing of both facial and vocal cues (13, 25, 26), it is important to develop and examine the effects of training that targets multimodal emotion perception abilities in schizophrenia.
In the present study, based on our prior work (27) we developed an audiovisual emotion perception training and test in which a voice and a face were simultaneously presented and trainees were asked to judge whether the emotion of the voice and the emotion of the face were the same or different. We selected angry and happy stimuli because they represent a bipolar dimension of valence (28), and angry and happy faces are the most distinctive with the least overlap in facial visual information processing (29). For the facial stimuli, we used morphed facial photos that varied on a continuum from angry to happy, which allowed us to create ambiguous expressions and to assess the level of difficulty of training. We aimed to report preliminary data regarding first training effects of this program by conducting a six-session intervention and one follow-up session in a group of patients with schizophrenia. We expected that although schizophrenia patients would perform worse than the healthy controls in accuracy of emotion recognition prior to the training, the patients would demonstrate significant improvement following the training. We also aimed to examine if the training has differential effects on the perceptions of angry and happy emotions based on previous findings that schizophrenia patients show more deficits in recognition of negative emotions than positive emotions [e.g., (30)] and expected that the training effects would be more pronounced for anger compared to happiness.
Methods
Subjects
Eighteen schizophrenia patients (SP) were recruited from a long-term care mental institution in South Korea. At the time of enrollment, all patients met the criteria for schizophrenia based on the Structured Clinical Interview for DSM-IV (SCID-IV) (31) and were on stable antipsychotic medication (risperidone or olanzapine). The psychotic symptoms were evaluated using the Positive and Negative Syndrome Scale (PANSS) (32). None of the patients had a history of central nervous system diseases (e.g., epilepsy or cerebrovascular accident), substance abuse, electroconvulsive therapy, mental retardation, head injury, or hearing impairment.
Twenty healthy controls (HC) were recruited via advertisements in local newspapers and flyers. Subjects were excluded if they had any neurological disorder, head injury, or personal or family history of psychiatric diseases. After the initial screening, potential HCs were interviewed using the SCID for Axis II Psychiatric Disorders and were excluded if they had any of these disorders (33). SP and HC groups were matched for age, sex, and education. Finally, 16 SP and 18 HC subjects participated to the end of the study. All subjects gave written informed consent before the experimental procedures commenced. The protocol was approved by the Institutional Review Board of Inje University Ilsan Paik Hospital.
Audiovisual Emotion Perception Training and Tests
As illustrated in Figure 1A, we developed the audiovisual emotion perception training and test. In this paradigm, each trial began with a fixation cross (200 ms) that was followed by the presentation of a morphed face (1,000 ms) and then a mask (black screen; 500 ms) while a voice was simultaneously delivered via earphones. Subjects were asked to judge whether the emotion of the face and the emotion of the voice were the same or different by pressing buttons that were counterbalanced across subjects. If subjects had correct responses, a yellow circle appeared, and if subjects had incorrect responses, a blue X appeared in the center of the screen for 500 ms. Trials of the training and the tests were the same except that feedback was given for subjects' responses only in the training.
Figure 1. (A) Trial sequences and morphed images used at the training and test and (B) schematic illustration of the procedure. SP, schizophrenia patient; HC, healthy control.
The test, which was given prior to the training and after the training, consisted of 840 trials [21 morphed faces × 2 voice emotions × 4 identities (two males, two females) × 5 repetitions], of which 400 trials were congruent audiovisual stimulus pairs (i.e., angry face–angry voice, or happy face–happy voice), 400 trials were incongruent pairs (i.e., angry face–happy voice, or happy face–angry voice), and 40 trials had no correct answer (i.e., 50% morphed face–happy voice, or 50% morphed face–angry voice). The test that was given for the generalization test after the training consisted of 420 trials [21 morphed faces × 2 voice emotions × 2 identities (1 male, 1 female) × 5 repetitions], of which 200 trials were congruent, 200 trials were incongruent, and 20 trials had no correct answer.
The training consisted of 480 trials [6 morphed faces × 2 voice emotions × 4 identities (2 males, 2 females) × 10 repetitions], of which 240 trials were congruent and 240 trials were incongruent. For training purposes, only six images that were emotionally ambiguous from each continuum were used (i.e., 70, 65, 60, 40, 35, and 30% angry face).
For the facial stimuli, we created emotional morphed facial images using a computerized morphing program (Abrosoft FantaMorph version 5.4.1; Abrosoft, USA). We selected two prototypical photographs (i.e., angry and a happy face) of six identities (three males, three females) from a standardized facial stimulus set (Korea University Facial Expression Collection) (34) and created emotional morphed faces by merging the prototypical photographs of each identity in 5% steps, resulting in a morphing continuum in which a total of 21 faces with graded blending of the emotional facial features of the two faces. Finally, six angry–happy morphing continua for each six identities (three males, three females) were created. Among them, four morphing continua were used in the pre-training test session (T0), training session, and post-training test session (T1), and another two morphing continua (one male, one female) were used in the generalization test session (T2).
For the voice stimuli, we recorded valenced voices from amateur actors in a noise-free room. The actors were instructed to pronounce semantically neutral sentences (e.g., “stayed in the house”) either in an angry or happy voice. Sixteen voice stimuli (2 emotions × 2 sex × 4 sentences) were used in the training and tests. The mean length of the 16 voice stimuli was 1.06 s (SD = 59 ms).
Procedure
Three days prior to the study, all subjects were tested on their ability to identify the valence of the auditory stimuli, which was necessary to perform the task. Forty auditory stimuli (2 emotions × 2 sex × 10 sentences) were presented, including the stimuli that were used in the study. Subjects were asked to determine whether the presented auditory stimulus represented anger or happiness. There was no difference in the accuracy between SP group (M = 38.688, SD = 1.537) and HC group (M = 39.105, SD = 1.729), F(1, 33) =.561, p = 0.459.
The procedure of the training and tests is shown in Figure 1B. The SP group participated in a total of nine sessions: T0, six training sessions, T1, and T2. The T0 was to measure baseline that was conducted 7 days prior to the first training session. Training consisted of six sessions lasting over a period of 3 weeks. Each training session was divided into two blocks to reduce burden on subjects. Upon completing each block, subjects received tokens if they had scored 110% (for training sessions 1–3) or 120% (for training sessions 4–6) of the correct responses relative to T0. The number of tokens ranged from 1 to 12, and one token was ~1 US dollar. Five days after the final training session, the T1 was conducted to measure the effects of the training. To ascertain that the training effects would be generalizable, T2 was conducted using two faces that had not been previously used 7 days after T1. The HC group participated only in the T0.
Statistical Analysis
Sensitivity for emotion perception was calculated as d' according to signal detection theory. The calculation of d' (sensitivity index) was based on the formula reported in the paper by Macmillan and Creelman (35). The sensitivity index provides the separation between the means of the signal and the noise distribution, in units of the standard deviation of the noise distribution. An estimate of d' can be found from measurements of the hit rate and false alarm rate. This is calculated as: d' = Z(hit rate) – Z(false alarm rate), where function Z(p), p ∈ [0,1], is the inverse of the cumulative Gaussian distribution. A correction of extreme proportions of hit and false-alarm rates was applied as proposed by Macmillan and Kaplan (36). A higher d' indicates that the signal is more readily detected.
In order to rule out any possible covariates, correlations between emotion perception (d') and demographic/clinical characteristics of the patients, including age, sex, education, age of onset, two WAIS subtest scores, and PANSS scores (total score and five subscale scores), were analyzed using nonparametric correlational analyses (Spearman's rho).
Departure from normal distribution assumption was tested by the Shapiro–Wilk's test. Due to normality of data, descriptive statistics show means and standard deviations. To examine group differences in demographic and clinical information, paired-sample t-tests and chi-square test were conducted. To investigate pre-existing group difference, a mixed ANCOVA on d' was performed with Group (SP, HC) as a between-subjects factor, Emotion (anger, happiness) and Morphing (three levels; lv1: 100, 95, 90 and 10, 5, 0% of angry faces; lv2: 85, 80, 75 and 25, 20, 15% of angry faces; lv3: 70, 65, 60, 55, and 45, 40, 35, 30% of angry faces) as within-subjects factors, and premorbid IQ (WAIS Information, Vocabulary) as covariates. To evaluate the training and generalization effects, a repeated measures ANOVA on d' was performed with Emotion, Morphing, and Time (T0, T1, T2). Paired t tests were conducted if there were any significant results in the repeated measures ANOVA. The Bonferroni correction was used for pairwise comparisons and the Greenhouse–Geisser adjustment was used to correct for violations of sphericity.
Results
Characteristics of Subjects
Demographic and clinical data for the SP and HC groups are summarized in Table 1. There were no group differences in age, t(32) = 0.112, p = 0.912, education, t(32) = 0.489, p = 0.628, or sex ratio, χ2 = 0.133, p =0.744. However, there was a group difference in the WAIS Information subtest, t(32) = 2.267, p = 0.030 and Vocabulary subtest, t(32) = 2.957, p = 0.006. Therefore, analyses of group comparison used the WAIS Information and Vocabulary scores as covariates.
Correlations Between Emotion Perception (d') and Demographic and Clinical Characteristics
Correlations between emotion perception (d') and all demographic and clinical characteristics of the SP group were not significant. These results indicate that demographic and clinical characteristics were not associated with emotion perception abilities in the patients.
Pre-existing Group Differences
Emotion perception performances (d') of the SP and HC groups on T0 are presented in Figures 2, 3A. There was a main effect of Group, F(1, 30) = 27.748, p < 0.001, = 0.480, such that the SP group performed worse than did the HC group. There was also an interaction effect of Group and Emotion, F(1, 30) = 7.262, p = 0.011, = 0.195. The SP group performed worse than did the HC group both for anger, t = 5.232, p < 0.001 and happiness, t = 2.404, p = 0.022. These results indicate that prior to the training, the SP group performed worse than did the HC group in audiovisual perception of anger as well as happiness.
Figure 2. Emotion perception performances (d') of the SP and HC according to time (T0, T1, and T2). Emotion perception performances for (A) anger, (B) happiness, and (C) total. d' = Z(hit rate) – Z(false alarm rate), where function Z(p), p ∈ [0,1]. Error bar indicates standard error. SP, schizophrenia patient; HC, healthy control. T0, pre-training session; T1, post-training session; T2, generalization session.
Figure 3. (A) Pre-existing group differences and (B) training and generalization effects in SP group. d' = Z(hit rate) – Z(false alarm rate), where function Z(p), p ∈ [0,1]. Error bar indicates standard error. SP, schizophrenia patient; HC, healthy control. T0, pre-training session; T1, post-training session; T2, generalization session.
Training and Generalization Effects in SP Group
Emotion perception performances (d') of the SP group on T0, T1, and T2 are presented in Figures 2, 3B. There was a main effect of Time, F(2, 30) = 15.336, p = 0.001, = 0.506, such that the performances of T1 and T2 were better than that of T0, ts > 3.796, ps < 0.002, and there was no difference between T1 and T2, t = 0.478, p = 0.640. There was also a main effect of Emotion, F(1, 15) = 19.617, p < 0.001, = 0.567, such that the perception of happiness was better than that of anger. Furthermore, there was an interaction effect of Time and Emotion, F(2, 30) = 3.406, p = 0.046, = 0.185. For anger, the performances of T1 and T2 were better than T0, ts > 4.335, ps < 0.001, and there was no difference between T1 and T2, t = 0.392, p = 0.701. For happiness, the performance of T1 was better than T0, t = 2.234, p = 0.041, but T2 was not different from T0, t = 1.243, p = 0.233. There was no difference between T1 and T2, t = 1.473, p = 0.161. These results indicate that the training improved and maintained the audiovisual perception for anger. However, the training improved but not maintained the audiovisual perception for happiness in the SP group.
Discussion
This study provides preliminary evidence that the audiovisual emotion perception training we developed may be a promising method for improving emotion perception for people with schizophrenia. The results showed that prior to the training, the schizophrenia group performed significantly worse than did the control group in the recognition of both anger and happiness, and the recognition of anger was much worse than that of happiness. These findings are consistent with previous findings that show that, compared to healthy controls, schizophrenia patients were significantly worse at recognizing angry faces as well as happy faces (37–42). It is also in line with the extant literature suggesting that schizophrenia patients perform more poorly with negative emotions such as anger, fear, or sadness, compared to positive emotions such as happiness or joy [e.g., (30, 43–45)].
Following the training, the schizophrenia group showed significant improvement in recognizing anger. Specifically, compared to their performance during the pre-training, the patients performed better across all levels of morphing after the six sessions of training as well as at the follow-up session with a new set of stimuli. With respect to happiness, although the training increased the patients' accuracy, this effect was not maintained or generalized to a new set of stimuli. These results provide preliminary evidence that the training program may yield improvements in the multimodal recognition of anger and happiness and that the training effect may be more reliable and pronounced for anger perception.
The training effects that were observed might be attributable to cross-modal effects in which information in one modality influences the emotion perception in another modality (46–48). In particular, studies on cross-modal effects found that the less ambiguous modality has a greater impact; for example, the effect of voice became greater as the faces were more morphed (49) or ambiguous (25). Our paradigm had morphed faces on the angry–happy continuum whereas voices were distinctively angry or happy. Thus, it is possible that the auditory cues influenced and guided the visual cues in integrating multimodal cues in our training. Indeed, auditory information has been known to guide eye movements in audiovisual emotional processing, such that an emotional voice yielded longer and more frequent fixations on emotionally congruent faces (50–52) and drew more attention to salient facial features, which could help to improve emotion recognition in schizophrenia patients (11).
To our knowledge, this is the first study of a multimodal training program targeting the integration of vocal and facial emotional information. This program has several strengths. First, the training is relatively brief, self- and computer-administered, and free from confounding effects of group interactions as in other common treatment allowing a more targeted intervention. Second, it uses morphed, ambiguous faces, which have greater ecological validity than the prototypical faces that were used in previous studies, and enabled the training to have increasing levels of task difficulty. Third, it has a follow-up session with a new set of stimuli based on which the generalizability of the findings can be examined.
There are several limitations of the current study that warrant more control conditions to draw a conclusion. First, we did not have post-training data of the healthy control group, which limited the control for the effect of repetition and unspecific components of the training. A randomized controlled trial should be followed to confirm the efficacy of the program. Second, because the tests we used were modified from the training program, “training to the test” effect might have occurred. An independent test on multimodal emotion perception shall be needed to measure the training effects (53). Third, more categories of emotion are required to rule out the possibility that the patients learned to differentiate between the two emotions in the current study. Lastly, it will be important for future studies to investigate the extent to which the changes brought out by this training impact people's everyday life, particularly social functioning.
In conclusion, this study provides an initial evaluation of the effects of the audiovisual emotion perception program. Although more controlled research is needed, the current results provide preliminary evidence that the multimodal training of the audiovisual using faces and vocal information might improve the perception of anger in individuals with chronic schizophrenia. Future research with more rigorous designs and longer-term follow-up with functioning outcomes is needed to confirm the efficacy of the program.
Data Availability Statement
The datasets generated for this study will not be made publicly available. We did not get the IRB approval on this.
Ethics Statement
The studies involving human participants were reviewed and approved by Institutional Review Board of Inje University Ilsan Paik Hospital. The patients/participants provided their written informed consent to participate in this study.
Author Contributions
HK and JJ conceived and designed the study. S-HL provided access to patients and JJ collected data. HL and JJ analyzed and interpreted the data, and wrote the manuscript. All authors read and approved the final manuscript.
Funding
This work was supported by the National Research Foundation of Korea Grant funded by the Korean Government (NRF-2010-32A-B00282).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
The authors thank Jiyoung Kim, Eunhee Chang, and Yeseul Chun for help with data collection.
References
1. Chan RC, Li H, Cheung EF, Gong QY. Impaired facial emotion perception in schizophrenia: a meta-analysis. Psychiatry Res. (2010) 178:381–90. doi: 10.1016/j.psychres.2009.03.035
2. Kohler CG, Walker JB, Martin EA, Healey KM, Moberg PJ. Facial emotion perception in schizophrenia: a meta-analytic review. Schizophr Bull. (2010) 36:1009–19. doi: 10.1093/schbul/sbn192
3. Mehl S, Hesse K, Schmidt AC, Landsberg MW, Soll D, Bechdolf A, et al. Theory of mind, emotion recognition, delusions and the quality of the therapeutic relationship in patients with psychosis - a secondary analysis of a randomized-controlled therapy trial. BMC Psychiatry. (2020) 20:59. doi: 10.21203/rs.2.16385/v2
4. Hoekert M, Kahn RS, Pijnenborg M, Aleman A. Impaired recognition and expression of emotional prosody in schizophrenia: review and meta-analysis. Schizophr Res. (2007) 96:135–45. doi: 10.1016/j.schres.2007.07.023
5. Lin Y, Ding H, Zhang Y. Emotional prosody processing in schizophrenic patients: a selective review and meta-analysis. J Clin Med. (2018) 7. doi: 10.3390/jcm7100363
6. Fiszdon JM, Bell MD. Effects of presentation modality and valence on affect recognition performance in schizophrenia and healthy controls. Psychiatry Res. (2009) 170:114–8. doi: 10.1016/j.psychres.2008.11.014
7. Thaler NS, Strauss GP, Sutton GP, Vertinski M, Ringdahl EN, Snyder JS, et al. Emotion perception abnormalities across sensory modalities in bipolar disorder with psychotic features and schizophrenia. Schizophr Res. (2013) 147:287–92. doi: 10.1016/j.schres.2013.04.001
8. Castagna F, Montemagni C, Maria Milani A, Rocca G, Rocca P, Casacchia M, et al. Prosody recognition and audiovisual emotion matching in schizophrenia: the contribution of cognition and psychopathology. Psychiatry Res. (2013) 205:192–8. doi: 10.1016/j.psychres.2012.08.038
9. de Gelder B, Vroomen J, de Jong SJ, Masthoff ED, Trompenaars FJ, Hodiamont P. Multisensory integration of emotional faces and voices in schizophrenics. Schizophr Res. (2005) 72:195–203. doi: 10.1016/j.schres.2004.02.013
10. de Jong JJ, Hodiamont PP, Van den Stock J, de Gelder B. Audiovisual emotion recognition in schizophrenia: reduced integration of facial and vocal affect. Schizophr Res. (2009) 107:286–93. doi: 10.1016/j.schres.2008.10.001
11. Simpson C, Pinkham AE, Kelsven S, Sasson NJ. Emotion recognition abilities across stimulus modalities in schizophrenia and the role of visual attention. Schizophr Res. (2013) 151:102–6. doi: 10.1016/j.schres.2013.09.026
12. Lavoie MA, Plana I, Bedard Lacroix J, Godmaire-Duhaime F, Jackson PL, Achim AM. Social cognition in first-degree relatives of people with schizophrenia: a meta-analysis. Psychiatry Res. (2013) 209:129–35. doi: 10.1016/j.psychres.2012.11.037
13. Comparelli A, Corigliano V, De Carolis A, Mancinelli I, Trovini G, Ottavi G, et al. Emotion recognition impairment is present early and is stable throughout the course of schizophrenia. Schizophr Res. (2013) 143:65–9. doi: 10.1016/j.schres.2012.11.005
14. Irani F, Seligman S, Kamath V, Kohler C, Gur RC. A meta-analysis of emotion perception and functional outcomes in schizophrenia. Schizophr Res. (2012) 137:203–11. doi: 10.1016/j.schres.2012.01.023
15. Javed A, Charles A. The importance of social cognition in improving functional outcomes in schizophrenia. Front Psychiatry. (2018) 9:157. doi: 10.3389/fpsyt.2018.00157
16. Grant N, Lawrence M, Preti A, Wykes T, Cella M. Social cognition interventions for people with schizophrenia: a systematic review focussing on methodological quality and intervention modality. Clin Psychol Rev. (2017) 56:55–64. doi: 10.1016/j.cpr.2017.06.001
17. Statucka M, Walder DJ. Efficacy of social cognition remediation programs targeting facial affect recognition deficits in schizophrenia: a review and consideration of high-risk samples and sex differences. Psychiatry Res. (2013) 206:125–39. doi: 10.1016/j.psychres.2012.12.005
18. Nijman SA, Veling W, van der Stouwe ECD, Pijnenborg GHM. Social cognition training for people with a psychotic disorder: a network meta-analysis. Schizophr Bull. (2020) 46:1086–103. doi: 10.1093/schbul/sbaa023
19. Penn DL, Roberts DL, Combs D, Sterne A. Best practices: The development of the Social Cognition and Interaction Training program for schizophrenia spectrum disorders. Psychiatr Serv. (2007) 58:449–51. doi: 10.1176/ps.2007.58.4.449
20. Frommann N, Streit M, Wolwer W. Remediation of facial affect recognition impairments in patients with schizophrenia: a new training program. Psychiatry Res. (2003) 117:281–4. doi: 10.1016/S0165-1781(03)00039-8
21. Walther S, Stegmayer K, Sulzbacher J, Vanbellingen T, Müri R, Strik W, et al. Nonverbal social communication and gesture control in schizophrenia. Schizophr Bull. (2015) 41:338–45. doi: 10.1093/schbul/sbu222
22. Bordon N, O'Rourke S, Hutton P. The feasibility and clinical benefits of improving facial affect recognition impairments in schizophrenia: Systematic review and meta-analysis. Schizophr Res. (2017) 188:3–12. doi: 10.1016/j.schres.2017.01.014
23. Kurtz MM, Richardson CL. Social cognitive training for schizophrenia: a meta-analytic investigation of controlled research. Schizophr Bull. (2012) 38:1092–104. doi: 10.1093/schbul/sbr036
24. Aloi M, de Filippis R, Grosso Lavalle F, Chiappetta E, Vigano C, Segura-Garcia C, et al. Effectiveness of integrated psychological therapy on clinical, neuropsychological, emotional and functional outcome in schizophrenia: a RCT study. J Ment Health. (2020) 29:524–31. doi: 10.1080/09638237.2018.1521948
25. Collignon O, Girard S, Gosselin F, Roy S, Saint-Amour D, Lassonde M, et al. Audio-visual integration of emotion expression. Brain Res. (2008) 1242:126–35. doi: 10.1016/j.brainres.2008.04.023
26. Lin Y, Ding H, Zhang Y. Multisensory integration of emotion in schizophrenic patients. Multisens Res. (2020) 1:1–37. doi: 10.1163/22134808-bja10016
27. Jeong JW, Wendimagegn TW, Chang E, Chun Y, Park JH, Kim HJ, et al. Classifying schizotypy using an audiovisual emotion perception test and scalp electroencephalography. Front Hum Neurosci. (2017) 11:450. doi: 10.3389/fnhum.2017.00450
28. Hamann S. Mapping discrete and dimensional emotions onto the brain: controversies and consensus. Trends Cogn Sci. (2012) 16:458–66. doi: 10.1016/j.tics.2012.07.006
29. Smith ML, Cottrell GW, Gosselin F, Schyns PG. Transmitting and decoding facial expressions. Psychol Sci. (2005) 16:184–9. doi: 10.1111/j.0956-7976.2005.00801.x
30. Edwards J, Jackson HJ, Pattison PE. Emotion recognition via facial expression and affective prosody in schizophrenia: a methodological review. Clin Psychol Rev. (2002) 22:789–832. doi: 10.1016/S0272-7358(02)00130-7
31. First M, Spitzer R, Gibbon M, Williams J. Structured Clinical Interview for DSM-IV Axis I Disorders (SCID-I), Clinician Version 1997. New York, NY: New York State Psychiatric Institute (1997).
32. Emsley R, Rabinowitz J, Torreman M, Group R-I-EPGW. The factor structure for the Positive and Negative Syndrome Scale (PANSS) in recent-onset psychosis. Schizophr Res. (2003) 61:47–57. doi: 10.1016/S0920-9964(02)00302-X
33. Kim M, Choi J, Cho Y. The Korea University Facial Expression Collection (KUFEC) and semantic differential ratings of emotion. Korean J Psychol Gen. (2011) 30:1189–2111.
34. First MB, Gibbon M. The Structured Clinical Interview for DSM-IV Axis I Disorders (SCID-I) and the Structured Clinical Interview for DSM-IV Axis II Disorders (SCID-II). Washington, DC: American Psychiatric Publishing (2004).
35. Macmillan NA, Creelman CD. Response bias: characteristics of detection theory, threshold theory, and “nonparametric” indexes. Psychol Bull. (1990) 107:401–13. doi: 10.1037/0033-2909.107.3.401
36. Macmillan NA, Kaplan HL. Detection theory analysis of group data: estimating sensitivity from average hit and false-alarm rates. Psychol Bull. (1985) 98:185–99. doi: 10.1037/0033-2909.98.1.185
37. Tseng H-H, Chen S-H, Liu C-M, Howes O, Huang Y-L, Hsieh MH, et al. Facial and prosodic emotion recognition deficits associate with specific clusters of psychotic symptoms in schizophrenia. PLoS One. (2013) 8:e66571. doi: 10.1371/journal.pone.0066571
38. Mendoza R, Cabral-Calderin Y, Domínguez M, Garcia A, Borrego M, Caballero A, et al. Impairment of emotional expression recognition in schizophrenia: a Cuban familial association study. Psychiatry Res. (2011) 185:44–8. doi: 10.1016/j.psychres.2009.10.006
39. Huang CL-C, Hsiao S, Hwu H-G, Howng S-L. Are there differential deficits in facial emotion recognition between paranoid and non-paranoid schizophrenia? A signal detection analysis. Psychiatry Res. (2013) 209:424–30. doi: 10.1016/j.psychres.2013.03.026
40. Bediou B, Asri F, Brunelin J, Krolak-Salmon P, D'Amato T, Saoud M, et al. Emotion recognition and genetic vulnerability to schizophrenia. Br J Psychiatry. (2007) 191:126–30. doi: 10.1192/bjp.bp.106.028829
41. Tsoi DT, Lee K-H, Khokhar WA, Mir NU, Swalli JS, Gee KA, et al. Is facial emotion recognition impairment in schizophrenia identical for different emotions? A signal detection analysis. Schizophr Res. (2008) 99:263–9. doi: 10.1016/j.schres.2007.11.006
42. Schneider F, Gur RC, Koch K, Backes V, Amunts K, Shah NJ, et al. Impairment in the specificity of emotion processing in schizophrenia. Am J Psychiatry. (2006) 163:442–7. doi: 10.1176/appi.ajp.163.3.442
43. Edwards J, Pattison PE, Jackson HJ, Wales RJ. Facial affect and affective prosody recognition in first-episode schizophrenia. Schizophr Res. (2001) 48:235–53. doi: 10.1016/S0920-9964(00)00099-2
44. Kohler CG, Turner TH, Bilker WB, Brensinger CM, Siegel SJ, Kanes SJ, et al. Facial emotion recognition in schizophrenia: intensity effects and error pattern. Am J Psychiatry. (2003) 160:1768–74. doi: 10.1176/appi.ajp.160.10.1768
45. Martin D, Croft J, Pitt A, Strelchuk D, Sullivan S, Zammit S. Systematic review and meta-analysis of the relationship between genetic risk for schizophrenia and facial emotion recognition. Schizophr Res. (2020) 218:7–13. doi: 10.1016/j.schres.2019.12.031
46. Campanella S, Belin P. Integrating face and voice in person perception. Trends Cogn Sci. (2007) 11:535–43. doi: 10.1016/j.tics.2007.10.001
47. Ethofer T, Anders S, Erb M, Droll C, Royen L, Saur R, et al. Impact of voice on emotional judgment of faces: an event-related fMRI study. Hum Brain Mapp. (2006) 27:707–14. doi: 10.1002/hbm.20212
48. Focker J, Roder B. Event-related potentials reveal evidence for late integration of emotional prosody and facial expression in dynamic stimuli: an ERP study. Multisens Res. (2019) 32:473–97. doi: 10.1163/22134808-20191332
49. De Gelder B, Vroomen J. The perception of emotions by ear and by eye. Cogn Emot. (2000) 14:289–311. doi: 10.1080/026999300378824
50. Paulmann S, Jessen S, Kotz SA. It's special the way you say it: an ERP investigation on the temporal dynamics of two types of prosody. Neuropsychologia. (2012) 50:1609–20. doi: 10.1016/j.neuropsychologia.2012.03.014
51. Rigoulot S, Pell MD. Seeing emotion with your ears: emotional prosody implicitly guides visual attention to faces. PLoS One. (2012) 7:e30740. doi: 10.1371/journal.pone.0030740
52. Palama A, Malsert J, Grandjean D, Sander D, Gentaz E. The cross-modal transfer of emotional information from voices to faces in 5-, 8- and 10-year-old children and adults: an eye-tracking study. Emotion. (2020) 50:1609–20. doi: 10.1037/emo0000758
Keywords: schizophrenia, emotion perception, multimodal integration, audiovisual, training
Citation: Jeong JW, Kim HT, Lee S-H and Lee H (2021) Effects of an Audiovisual Emotion Perception Training for Schizophrenia: A Preliminary Study. Front. Psychiatry 12:522094. doi: 10.3389/fpsyt.2021.522094
Received: 20 December 2019; Accepted: 18 March 2021;
Published: 05 May 2021.
Edited by:
Wolfgang Wölwer, Heinrich Heine University, GermanyReviewed by:
Monika Mak, Pomeranian Medical University, PolandSebastian Walther, University of Bern, Switzerland
Copyright © 2021 Jeong, Kim, Lee and Lee. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Hyejeen Lee, aHllamVlbmxlZSYjeDAwMDQwO2dtYWlsLmNvbQ==