- 1Department of Psychology and Educational Counseling, The Center for Psychobiological Research, Max Stern Yezreel Valley College, Emek Yezreel, Israel
- 2Department of Brain Sciences, Faculty of Medicine, Imperial College London, London, United Kingdom
- 3Educational Psychology Department, Center for Psychobiological Research, Max Stern Yezreel Valley College, Emek Yezreel, Israel
This study evaluated the potential of ChatGPT, a large language model, to generate mentalizing-like abilities that are tailored to a specific personality structure and/or psychopathology. Mentalization is the ability to understand and interpret one’s own and others’ mental states, including thoughts, feelings, and intentions. Borderline Personality Disorder (BPD) and Schizoid Personality Disorder (SPD) are characterized by distinct patterns of emotional regulation. Individuals with BPD tend to experience intense and unstable emotions, while individuals with SPD tend to experience flattened or detached emotions. We used ChatGPT’s free version 23.3 and assessed the extent to which its responses akin to emotional awareness (EA) were customized to the distinctive personality structure-character characterized by Borderline Personality Disorder (BPD) and Schizoid Personality Disorder (SPD), employing the Levels of Emotional Awareness Scale (LEAS). ChatGPT was able to accurately describe the emotional reactions of individuals with BPD as more intense, complex, and rich than those with SPD. This finding suggests that ChatGPT can generate mentalizing-like responses consistent with a range of psychopathologies in line with clinical and theoretical knowledge. However, the study also raises concerns regarding the potential for stigmas or biases related to mental diagnoses to impact the validity and usefulness of chatbot-based clinical interventions. We emphasize the need for the responsible development and deployment of chatbot-based interventions in mental health, which considers diverse theoretical frameworks.
1. Introduction
Artificial intelligence (AI) has become an important tool in mental health research and treatment (1). ChatGPT is one such AI model that has shown great potential in multiple fields. Based on the GPT-3.5 architecture and trained on vast amounts of data, ChatGPT receives natural language input and generates human-like responses (2). The pervasive dissemination of ChatGPT, coupled with its non-targeted design for mental health applications, underscores the scholarly imperative to investigate its theoretical and clinical potential in the domain of mental health.
In a recent study exploring the use of AI in the realm of mental health, findings highlighted its capacity to enhance efficiency by managing technical tasks, aiding in diagnosis, and incorporating biological feedback (1, 3, 4). Additionally, A review of various studies investigating the influence of chatbots indicated their potential to provide therapeutic support and alleviate mental distress (5). However, it was observed that certain interpersonal skills exhibited by these chatbots, such as empathy and emotional awareness (EA), are still in the early stages of development. A previous study (6) comparing ChatGPT’s capacity to produce responses with EA with the general population’s norms revealed that ChatGPT’s EA-like abilities are significantly superior to those of humans. This result also demonstrates the potential of this new AI technology in “soft” psychological skills, such as empathy and EA. The present study seeks to evaluate whether ChatGPT exhibits promising capabilities in accurately generating mentalizing-like abilities that are tailored to a specific personality structure or psychopathology.
Mentalizing is the ability to understand and interpret the mental states of oneself and others, including thoughts, feelings, and intentions (7). This term encompasses a range of related concepts, such as theory of mind, social cognition, perspective taking, EA, and empathy (8). Mentalizing widely recognized as a crucial psychological skill that has significant implications in various domains (9). For example, in the realm of child development, a parent’s mentalizing capacity is deemed vital for promoting healthy cognitive and emotional growth in their offspring (10). On the other hand, mentalization has been identified as a potential transdiagnostic factor in psychopathology (9), as numerous studies have found impairments in mentalizing across a variety of psychiatric and neurological disorders, including psychosis (11), personality disorders (such as Borderline Personality Disorder) (12), depression (13), anxiety (14), trauma-related disorders (15), addictions (16), eating disorders (17), and Machado-Joseph disease (18). Lastly, mentalizing is regarded as a fundamental aspect of psychotherapy (9). Many psychological therapies aim to enhance patients’ mentalizing abilities (19) in order to promote self-acceptance, awareness of their illness, and a more accurate understanding of their thoughts, emotions, and behaviors. Given that various psychopathologies exhibit distinct impairments in mentalizing capacities (9), it is imperative for the therapist to comprehend the patient’s inner realm together with their unique attributes, personality structure, and current mental condition in order to offer more personalized treatment. Two cases that represent opposite ends of the spectrum of mental experience resulting from diverse personality structures are Borderline Personality Disorder (BPD) and Schizoid Personality Disorder (SPD).
BPD includes several typical symptoms (20): (1) An intense fear of being abandoned, resulting in extreme measures to avoid real or perceived rejection or separation; (2) An erratic pattern of intense relationships, characterized by idealization followed by sudden beliefs that the other person is cruel or does not care enough; (3) Rapid fluctuations in self-identity and self-image, including changes in values, goals, and viewing oneself as either “bad” or nonexistent; (4) Brief periods of paranoia and detachment from reality caused by stress, lasting anywhere from a few minutes to a few hours; (5) Impulsive and dangerous actions, such as gambling, reckless driving, unsafe sex, binge eating or drug abuse, or self-sabotaging behaviors like abruptly quitting a good job or ending a positive relationship; (6) Suicidal tendencies or self-harm, often linked to fear of separation or rejection; (7) Wide mood swings that can persist for a few hours or days, including intense happiness, irritability, shame, or anxiety; (8) Chronic feelings of emptiness; and (9) Inappropriate and intense anger, frequently leading to outbursts, sarcasm, bitterness, or physical altercations. In short, BPD is defined by a turbulent and intense emotional experience that encompasses a broad spectrum of emotions, which emerge primarily within the context of interpersonal relationships.
Schizoid Personality Disorder (SPD) includes several typical symptoms (21): (1) A preference for solitude and a proclivity to engage in activities that are solitary in nature; (2) A lack of interest in or enjoyment of close interpersonal relationships; (3) Minimal or nonexistent sexual desire; (4) Perception of an inability to experience pleasure or enjoyment; (5) Difficulty expressing emotions and responding in an appropriate manner to various situations; (6) A demeanor that may be perceived by others as humorless, indifferent, or emotionally cold; (7) The possibility of appearing unmotivated or lacking in clear goals; and (8) A lack of response to both praise and criticism from others. In short, people with SPD are characterized by relative emotional narrowing, detachment, and interpersonal distance.
The present study examines the potential of ChatGPT to display mentalizing-like performances that are tailored to a specific personality structure or psychopathology. The present study will concentrate on a particular facet of mentalizing: Emotional Awareness. Specifically, we examine whether ChatGPT’s emotional descriptions can differentiate between the emotional experiences of individuals with BPD and SPD. Our hypothesis was that ChatGPT will be able to describe the emotional reactions of individuals with BPD as richer and more intense than those with SPD. The selection of emotional intensity and richness as variables in mentalizing studies is strategic and quantifiable. Emotional richness is a recognized measure in the Levels of Emotional Awareness Scale (LEAS) and in the broader literature for evaluating EA (22). In this study, the intensity index was introduced as a novel measure to provide a quantitative assessment of the intensity of the emotional experience, not just the emotional richness. By quantifying these variables, we can objectively measure and compare mentalizing abilities across different contexts. The study examines the theoretical implications of the findings through two approaches: the critical social approach and the clinical approach. We selected ChatGPT 3.5 due to its widespread popularity and general-purpose design, which wasn’t specifically tailored for mental health or “soft skills” like EA. Furthermore, we utilized its free version to guarantee broad accessibility without financial constraints.
2. Methods
2.1. Artificial intelligence procedure
We used ChatGPT’s free version 23.3 (OpenAI San Francisco, OpenAI, n.d.) on April 19–20, 2023 to evaluate its EA performance using the Levels of Emotional Awareness Scale (LEAS; 22). The complete study protocol was approved by the Institutional Review Board (2023-40 YVC EMEK).
2.2. Input source
The LEAS (22) consists of 20 open-ended questions that described emotionally charged scenarios intended to elicit emotions, such as anger, fear, happiness, and sadness. The LEAS was found to have high reliability and validity. Cronbach’s alpha was between 0.7–0.91.
In the original version, participants are required to imagine themselves experiencing the situation and write down their own (“you”) emotions, as well as those of the other person described in the scenario. Since ChatGPT does not report on self-emotions, to address the research question, we presented the scenarios by replacing the “you” instruction with “Human with Borderline Personality Disorder” (BPD condition) or “Human with Schizoid Personality Disorder” (SPD condition) (6). The 20 items of the LEAS were presented once under the BPD condition and once under the SPD condition. Each transfer was made in a new tab. In each transfer, after the ChatGPT responses was received, a follow-up question was asked: “For each emotion you indicated its intensity from 0 – the weakest intensity to 10 – the strongest intensity.”
2.3. Scoring
ChatGPT’s performance was scored using the standard LEAS manual (22) and contained two sub-scales that evaluated the main character (MC) (in our study, individual with BPD/SPD) and the other character (OC) (in our study, a secondary character introduced without any mention of psychopathology). MC and OC scores were calculated separately (0–4 scores per scenario, range 0–80) with a high score indicating higher EA. In addition to the standard scoring, we added two indexes: the number of emotions per scenario for the MC and OC and emotions intensity per scenario for the MC and OC range from 0 (weakest) to 10 (strongest).
2.4. Statistical analysis
We used a T-test to compare ChatGPT’s responses to the 20 scenarios in the standard LEAS score and the number of emotions and the intensity of emotions between the MC in the first condition (human with BPD) to the second condition (human with SPD) and the OC in the first to the OC in the second condition. A paired T-test was used to compare ChatGPT’s responses to the 20 scenarios in the standard LEAS score and the number of emotions and the intensity of emotions between the MC to the OC in the first (BPD) and second (SPD) conditions. The statistical analyzes were performed using SPSS Statistics (IBM) version 28.
3. Results
3.1. Main character with borderline personality disorder vs. main character with Schizoid Personality Disorder
In this section, we report on the results of the LEAS analysis conducted on ChatGPT’s responses to MC with BPD and SPD in 20 scenarios (see Table 1 for an example of these responses in five different scenarios).
The LEAS scores calculated from ChatGPT’s responses were significantly higher for the MC with BPD than the MC with SPD (t (2,19) = 5.82, p < 0.001) (see Figure 1). The score for all 20 LEAS scenarios for the former was 80 (the maximum score of the LEAS) and 47 for the latter. Furthermore, the MC with BPD was rated by ChatGPT with significantly higher number of emotions and emotion intensity than the MC with SPD [number of emotions (t (2,38) = 7.3, p < 0.001); emotions intensity (t (2,25.74) = 5.82, p < 0.05)].
Figure 1. The standard LEAS score (mean ± SEM) of the Main Character (MC) with Borderline Personality Disorder (BPD), (MC) with Schizoid Personality Disorder (BPD), and Other Character (OC) with no psychopathology in the BPD and SPD conditions. ***p < 0.001.
3.2. Main characters vs. other characters
The LEAS scores calculated from ChatGPT’s responses were significantly higher for the MC with BPD than the OC (t (18) = 4.23, p < 0.001) (see Figure 1). The score for all 20 scenarios on the LEAS was 80 for the former and 66 for the latter. Furthermore, the MC with BPD had a significantly higher number of emotions and emotion intensity than the OC [number of emotions (t (19) = 11.13, p < 0.001); emotions intensity (t (19) = 5.82, p < 0.001)].
The LEAS scores calculated from ChatGPT’s responses were not significantly different for the MC with SPD compared to the OC (p > 0.05) (see Figure 1). Furthermore, the MC with SPD was not rated by ChatGPT as significantly different in the emotion intensity compared to the OC [emotions intensity (p > 0.05)] but was rated with a significantly higher number of emotions than the OC [number of emotions (t (19) = 3.4, p < 0.01)].
3.3. Other character vs. other character
The LEAS scores calculated from ChatGPT’s responses were significantly higher for the OC in the scenarios where the MC had BPD than the OC in the scenarios where the MC had SPD (t (2,38)= 2.62, p < 0.05) (see Figure 1). No significant differences were found between the OC in the scenarios where the MC had BPD and the OC in the scenarios where the MC had SPD regarding the number of emotions and emotion intensity (p > 0.05).
4. Discussion
This study explores the capacity of ChatGPT to exhibit mentalizing-like abilities that are customized to individual personality structures or psychopathologies, namely BPD versus SPD. In line with the central hypothesis, ChatGPT manifested a greater degree of emotional insight in all three measures (LEAS scoring, number of emotions, and emotion intensity) when comparing the experiences of MC with BPD and MC with SPD. This result indicates a certain level of plasticity in ChatGPT ability to generate mentalizing-like responses or to understand and interpret mental states in a manner consistent with various personality structures or psychopathologies.
As expected, the emotional experience of the MC with BPD was found to be richer and more intense compared to the OC in the same scenario. This finding suggests that the ChatGPT attributes a more turbulent emotional response to a person presented with BPD than to OC without any psychopathology. Interestingly, when the MC was portrayed with SPD (which is supposed to reflect lower emotional intensity and richness), there was no difference between their LEAS score and emotional intensity of the OC without psychopathology. Several reasons might explain the absence of these differences: It’s possible that the model perceives interacting emotionally with someone with SPD as also emotionally numbing for the OC. The model might prefer reflecting the emotions of the OC based on what the MC is experiencing, and when the MC is presented with a limited emotional experience, it affects the OC.
In general, this study findings have implications that can be analyzed through the theoretical lenses of critical psychology (23) and clinical applied psychology (20, 24).
According to the first approach of critical psychology (25) and of disability studies (26), we suggest that the generation of ChatGPT responses may be impacted by stigmas or biases related to mental diagnoses. The association of emotions experienced with mental disabilities can be viewed as an example of stereotypical thinking from a social psychology perspective (27, 28). This line of thinking assumes that a person’s mental diagnosis determines their entire being and mental state, potentially overlooking the complexity and diversity of their emotional experiences.
Moreover, the generation of ChatGPT responses is based on data that may contain stigmas or biases related to mental diagnoses (29). This issue is particularly relevant given the extensive literature on the negative effects of stigmatization on mental health (30, 31) and highlights the potential for stigma to impact the validity and usefulness of chatbots that based on large language models interventions. As such, future developments in this technology must balance the need to avoid stigmatization with the desire to provide personalized support and guidance based on relevant mental health information.
Understanding these implications and considering them within these theoretical frameworks is crucial for the responsible development and deployment of chatbot-based interventions in the field of mental health. For example, disability studies perspectives offer insights into how technology can be designed to respect the diversity of experiences associated with psychopathologies (30).
According to the second approach of applied clinical psychology (20, 24), our findings align with existing literature. Previous studies have shown that individuals with BPD tend to have a more tumultuous emotional experience (32). High emotional intensity and emotional dysregulation were also reported in multiple neurobiological (33–35), cognitive (36), and behavioral studies (37). Additionally, SPD has been described by the central diagnostic manuals of the International Classification of Diseases (ICD) (24) and the Diagnostic and Statistical Manual of Mental Disorders (DSM) (38) as a state of emotional reduction, emotional detachment, emotional coldness, and difficulty expressing emotions. Indeed, a study by Coolidge et al. (39) found that schizoid symptoms correlated with the Toronto Alexithymia Scale (TAS-20). It can, therefore, be suggested that the ability to personalize interpersonal responses to psychopathology is in line with the clinical and theoretical knowledge.
The advantages of using ChatGPT in this context are numerous. It can, for example, serve as a diagnostic tool, or as a support system in psychotherapy, providing objective information to clinicians that can help them better understand their patients’ emotional experiences. Additionally, ChatGPT can serve as a research tool, providing researchers with new ways of exploring and understanding the emotional experiences of individuals with various psychopathologies.
Conversely, while the DSM describes the emotional response in SPD as flat and detached (20), other clinical approaches fundamentally disagree. From a psychoanalytic standpoint, ChatGPT’s portrayal of the schizoid individual’s mental experience may be seen as shallow and external, only scratching the surface of their inner world. According to McWilliams (40), one of the editors of the Psychodynamic Diagnostic Manual (PDM) (41), the schizoid personality represents a defensive reaction and withdrawal from further interpersonal stimulation following a traumatic event. It is important to recognize that schizoid individuals may outwardly appear detached yet internally yearn for human connection and intimacy. They may exhibit an intense emotional experience despite their emotional darkness and be perceived by others as “gentle” while harboring fantasies of destruction. Hence, we assume that from McWilliams’ perspective (40) the practical utilization of ChatGPT in the realms of diagnostics, treatment, and professional training may overlook significant and crucial aspects of the psychological experience.
Prior studies on the utilization of AI in mental health have primarily focused on dedicated applications that serve specific purposes, such as Woebot (42), Replika (5), or TEO (43). However, a research review evaluating the performance of chatbot applications found that although they exhibit promising results in various domains, their proficiency in soft skills, including empathy and EA, remains undeveloped (5).
To enhance the efficacy of AI in the domain of mental health, it is vital to not only gather evidence pertaining to the attainment or lack thereof of therapeutic goals but also to elucidate the mechanisms underpinning the therapeutic process. Such comprehension should be grounded in the ramifications of research findings in line with diverse theoretical frameworks. To ensure that novel and promising advances in AI include references to these implications and the existing knowledge base in the field of mental health, it is incumbent upon researchers and therapists in the domain of mental health and people with mental disabilities to actively and substantively participate in the research, comprehension, and refinement of tools based on this technology (44). Moreover, the thoughtful and ethical amalgamation of AI into mental health care is paramount, which necessitates a balanced consideration of the possible advantages alongside potential issues related to data ethics, the absence of clear directions for AI application development and integration, possible abuse leading to health disparities, honoring patient autonomy, and algorithmic transparency (45).
Despite this study’s promising results, it is important to consider the limitations that might impact their accuracy. First, the study focus on the ChatGPT model, neglecting a comparative analysis with other large language models such as BLOOM, Bard, claude.ai, T5, which could potentially offer different insights into AI’s general EA capabilities. Second, the study focused only on one aspect of mentalization, EA measured by emotional intensity and richness, which is only one component of a broader concept. Future studies are recommended to evaluate other aspects of mentalization using a variety of measurement tools. Third, the study examined only two forms of psychopathology or personality structures, which are at the opposite ends. Further research is required to determine whether ChatGPT can distinguish smaller differences between types of psychopathologies or personality structures.
5. Conclusion
This study demonstrates that ChatGPT has the capacity to exhibit mentalizing-like abilities tailored to individual personality structures or psychopathologies. However, it is crucial to consider the implications from the approaches of critical psychology and clinical applied psychology to avoid reinforcing stereotypes and stigmatization related to mental diagnoses. Balancing personalized support with the avoidance of stigma is essential in the responsible development of AI mental health interventions. Our findings align with existing clinical literature supporting the use of ChatGPT as a diagnostic tool which can aid in treatment processes and provide new avenues for research in the field of mental health.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to ZG9yaXRoQHl2Yy5hYy5pbA==.
Author contributions
DH-S: conception and design of the study, acquisition and analysis of data, and drafting of a significant portion of the manuscript. ZE: conception and design of the study, acquisition and analysis of data, and drafting of a significant portion of the manuscript and tables. ML: conception and design of the study and drafting of a significant portion of the manuscript. All authors contributed to the article and approved the submitted version.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Lee, EE, Torous, J, De Choudhury, M, Depp, CA, Graham, SA, Kim, H, et al. Artificial intelligence for mental health care: clinical applications, barriers, facilitators, and artificial wisdom. Biol Psychiatry Cogn Neurosci Neuroimaging. (2021) 6:856–64. doi: 10.1016/j.bpsc.2021.02.001
2. Rudolph, J, Tan, S, and Tan, S. ChatGPT: bullshit spewer or the end of traditional assessments in higher education? J Appl Learn Teach. (2023) 6:1–22. doi: 10.37074/jalt.2023.6.1.9
3. Bzdok, D, and Meyer-Lindenberg, A. Review machine learning for precision psychiatry: opportunities and challenges. Biol Psychiatry Cogn Neurosci Neuroimaging. (2018) 3:223–30. doi: 10.1016/j.bpsc.2017.11.007
4. Topol, E. Deep medicine: how artificial intelligence can make healthcare human again. New York: Basic Books (2019).
5. Pham, KT, Nabizadeh, A, and Selek, S. Artificial intelligence and chatbots in psychiatry. Psychiatry Q. (2022) 93:249–53. doi: 10.1007/s11126-022-09973-8
6. Elyoseph, Z, Hadar Shoval, D, Asraf, K, and Lvovsky, M. ChatGPT outperforms humans in emotional awareness evaluations. Front Psychol. (2023) 14:1199058. doi: 10.3389/fpsyg.2023.1199058
7. Freeman, C. What is mentalizing? An overview. Br J Psychother. (2016) 32:189–201. doi: 10.1111/bjp.12220
8. Aival-Naveh, E, and Kurman, J. Keeping culture in mind: a systematic review and initial conceptualization of mentalizing from a cross-cultural perspective. Clin Psychol. (2019) 26:1–25. doi: 10.1111/cpsp.12300
9. Luyten, P, Campbell, C, Allison, E, and Fonagy, P. The mentalizing approach to psychopathology: state of the art and future directions. Annu Rev Clin Psychol. (2020) 16:297–325. doi: 10.1146/annurev-clinpsy-071919-015355
10. Luyten, P, Nijssens, L, Fonagy, P, and Mayes, LC. Parental reflective functioning: theory, research, and clinical applications. Psychoanal Study Child. (2017) 70:174–99. doi: 10.1080/00797308.2016.1277901
11. Baslet, G, Termini, L, and Herbener, E. Deficits in emotional awareness in schizophrenia and their relationship with other measures of functioning. J Nerv Ment Dis. (2009) 197:655–60. doi: 10.1097/NMD.0b013e3181b3b20f
12. Fonagy, P, and Luyten, P. A multilevel perspective on the development of borderline personality disorder In: D Cicchetti, editor. Developmental psychopathology: maladaptation and psychopathology. Hoboken: Wiley (2016). 726–92.
13. Luyten, P, and Fonagy, P. The stress–reward–mentalizing model of depression: an integrative developmental cascade approach to child and adolescent depressive disorder based on the research domain criteria (RDoC) approach. Clin Psychol Rev. (2018) 64:87–98. doi: 10.1016/j.cpr.2017.09.008
14. Nolte, T, Guiney, J, Fonagy, P, Mayes, LC, and Luyten, P. Interpersonal stress regulation and the development of anxiety disorders: an attachment-based developmental framework. Front Behav Neurosci. (2011) 5:1–21. doi: 10.3389/fnbeh.2011.00055
15. Frewen, P, Lane, RD, Neufeld, RW, Densmore, M, Stevens, T, and Lanius, R. Neural correlates of levels of emotional awareness during trauma script-imagery in posttraumatic stress disorder. Psychosom Med. (2008) 70:27–31. doi: 10.1097/PSY.0b013e31815f66d4
16. Carton, S, Bayard, S, Jouanne, C, and Lagrue, G. Emotional awareness and alexithymia in smokers seeking help for cessation: a clinical analysis. J Smok Cessat. (2008) 3:81–91. doi: 10.1375/jsc.3.2.81
17. Robinson, P, Skårderud, F, and Sommerfeldt, B. eds. Eating disorders and mentalizing In: Hunger: mentalization-based treatments for eating disorders. Sveits: Springer (2018). 35–49.
18. Elyoseph, Z, Geisinger, D, Nave-Aival, E, Zaltzman, R, and Gordon, CR. “I do not know how you feel and how i feel about that”: mentalizing impairments in Machado-Joseph disease. Cerebellum. (2023). doi: 10.1007/s12311-023-01536-2
19. Carla, S, Fonagy, P, and Goodyer, I eds. Social cognition and developmental psychopathology, international perspectives in philosophy & psychiatry. Oxford: Oxford University Press (2008).
20. Shedler, J, Beck, A, Fonagy, P, Gabbard, GO, Gunderson, J, Kernberg, O, et al. Personality disorders in DSM-5. Am J Psychiatry. (2010) 167:1026–8. doi: 10.1176/appi.ajp.2010.10050746
21. Fariba, K, Madhanagopal, N, and Gupta, V. Schizoid personality disorder In: StatPearls [Internet]. Treasure Island: StatPearls Publishing (2022) Available at: (Accessed June 9, 2022) https://www.ncbi.nlm.nih.gov/books/NBK559234/
22. Lane, R, Quinlan, D, Schwartz, G, and Walker, P. The levels of emotional awareness scale: a cognitive–developmental measure of emotion. J Pers Assess. (1990) 55:124–34. doi: 10.1080/00223891.1990.9674052
23. Burr, V, and Dick, P. Social constructionism In: B Gough, editor. The Palgrave handbook of critical social psychology. London: Palgrave Macmillan (2017).
24. Reed, GM. Progress in developing a classification of personality disorders for ICD-11. World Psychiatry. (2018) 17:227–9. doi: 10.1002/wps.20533
25. Gough, B ed. The Palgrave handbook of critical social psychology. London: Palgrave Macmillan (2017).
26. Watson, N, Roulstone, A, and Thomas, C eds. Routledge handbook of disability studies. London: Routledge (2019).
27. Arboleda-Flórez, J, and Stuart, H. From sin to science: fighting the stigmatization of mental illnesses. Can J Psychiatr. (2012) 57:45–63. doi: 10.1177/070674371205700803
28. Shedlosky-Shoemaker, R, Strassle, CG, Def, SA, and Engler, JN. Addressing stigmatization of psychological disorders in introductory psychology. Scholarsh Teach Learn Psychol. (2022). doi: 10.1037/stl0000329
29. Timmons, AC, Duong, JB, Simo Fiallo, N, Lee, T, Vo, HPQ, Ahle, MW, et al. A call to action on assessing and mitigating bias in artificial intelligence applications for mental health. Perspect Psychol Sci. (2022) 1–22. doi: 10.1177/17456916221134490
30. Alexander, D, Walsh, B, Louise, J, and Foster, H. A call to action: a critical review of mental health related anti-stigma campaigns. Front Public Health. (2021) 8:1–15. doi: 10.3389/fpubh.2020.569539
31. Eylem, O, de Wit, L, Van Straten, A, Steubl, L, Melissourgaki, Z, Danışman, GT, et al. Stigma for common mental disorders in racial minorities and majorities a systematic review and meta-analysis. BMC Public Health. (2020) 20:1–20. doi: 10.1186/s12889-020-08964-3
32. Köhling, J, Schauenburg, H, and Ehrenthal, JC. Change of emotional experience in major depression and borderline personality disorder during psychotherapy: associations with depression severity and personality functioning. J Personal Disord. (2019):35. doi: 10.1521/pedi
33. Hall, NT, and Hallquist, MN. Dissociation of basolateral and central amygdala effective connectivity predicts the stability of emotion-related impulsivity in adolescents and emerging adults with borderline personality symptoms: a resting-state fMRI study. Psychol Med. (2023) 58:3533–47. doi: 10.1101/2021.05.17.444525
34. De La Peña-Arteaga, V, Berruga-Sánchez, M, Steward, T, Martínez-Zalacaín, I, Goldberg, X, Wainsztein, A, et al. An fMRI study of cognitive reappraisal in major depressive disorder and borderline personality disorder. Eur Psychiatry. (2021) 64:1–8. doi: 10.1192/j.eurpsy.2021.2231
35. Van Zutphen, L, Siep, N, Jacob, GA, Domes, G, Sprenger, A, Willenborg, B, et al. Impulse control under emotion processing: an fMRI investigation in borderline personality disorder compared to non-patients and cluster-C personality disorder patients. Brain Imaging Behav. (2020) 14:2107–21. doi: 10.1007/s11682-019-00161-0
36. Daros, AR, and Williams, GE. Meta-analysis and systematic review of emotion-regulation strategies in borderline personality disorder. Harv Rev Psychiatry. (2019) 27:217–32. doi: 10.1097/HRP.0000000000000212
37. De Panfilis, C, Schito, G, Generali, I, Gozzi, LA, Ossola, P, Marchesi, C, et al. Emotions at the border: increased punishment behavior during fair interpersonal exchanges in borderline personality disorder. J Abnorm Psychol. (2019) 128:162–72. doi: 10.1037/abn0000404
38. American Psychiatric Association. Diagnostic and statistical manual of mental disorders, 5th edition (DSM-5) American Psychiatric Association (2013).
39. Coolidge, FL, Estey, AJ, Segal, DL, and Marle, PD. Are alexithymia and schizoid personality disorder synonymous diagnoses? Compr Psychiatry. (2013) 54:141–8. doi: 10.1016/j.comppsych.2012.07.005
40. McWilliams, N. Some thoughts about schizoid dynamics. Psychoanal Rev. (2006) 93:1–24. doi: 10.1521/prev.2006.93.1.1
41. Lingiardi, V, and McWilliams, N eds. Psychodynamic diagnostic manual: PDM-2. New York: Guilford Publications (2017).
42. Fitzpatrick, KK, Darcy, A, and Vierhile, M. Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial. JMIR Ment Health. (2017) 4:e19. doi: 10.2196/mental.7785
43. Danieli, M, Ciulli, T, Mousavi, SM, Silvestri, G, Barbato, S, Di Natale, L, et al. Assessing the impact of conversational artificial intelligence in the treatment of stress and anxiety in aging adults: randomized controlled trial. JMIR Ment Health. (2022) 9:e38067. doi: 10.2196/38067
44. Elyoseph, Z, and Levkovich, I. Beyond human expertise: the promise and limitations of ChatGPT in suicide risk assessment. Front Psychiatry. (2023) 14:1213141. doi: 10.3389/fpsyt.2023.1213141
Keywords: artificial intelligence, borderline personality disorder, emotional intelligence, empathy, emotional awareness, Schizoid Personality Disorder
Citation: Hadar-Shoval D, Elyoseph Z and Lvovsky M (2023) The plasticity of ChatGPT’s mentalizing abilities: personalization for personality structures. Front. Psychiatry. 14:1234397. doi: 10.3389/fpsyt.2023.1234397
Edited by:
Xuntao Yin, Guangzhou Medical University, ChinaReviewed by:
Amir Tal, Tel Aviv University, IsraelAvi Segal, Ben-Gurion University of the Negev, Israel
Copyright © 2023 Hadar-Shoval, Elyoseph and Lvovsky. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Dorit Hadar-Shoval, RG9yaXRoQHl2Yy5hYy5pbA==
†These authors have contributed equally to this work