Skip to main content

ORIGINAL RESEARCH article

Front. Psychiatry

Sec. Digital Mental Health

Volume 16 - 2025 | doi: 10.3389/fpsyt.2025.1473614

This article is part of the Research Topic Application of chatbot Natural Language Processing models to psychotherapy and behavioral mood health View all 4 articles

GPT-4 generated psychological reports in psychodynamic perspective: a pilot study on quality, risk of hallucination and client satisfaction

Provisionally accepted
Namwoo Kim Namwoo Kim 1*Jiseon Lee Jiseon Lee 2Sung-Hyun Park Sung-Hyun Park 3Yoonseo On Yoonseo On 4Jieun Lee Jieun Lee 5Musung Keum Musung Keum 6Sanghoon Oh Sanghoon Oh 2Yoojin Song Yoojin Song 7Junhee Lee Junhee Lee 8Geun Hui Won Geun Hui Won 9Joon Sung Shin Joon Sung Shin 10Silvia Kyungjin Lho Silvia Kyungjin Lho 11Yoon Jung Hwang Yoon Jung Hwang 8Tae-Suk Kim Tae-Suk Kim 8
  • 1 Department of Clinical Medical Sciences, Seoul National University Hospital, Seoul, Republic of Korea
  • 2 Department of Psychiatry, Uijeongbu Eulji Medical Center, Eulji University, Gyeonggi, Republic of Korea
  • 3 Department of Medical Informatics, College of Medicine, Catholic University of Korea, Seoul, Seoul, Republic of Korea
  • 4 College of Medicine, Ewha Womans University, Seoul, Seoul, Republic of Korea
  • 5 Department of Pediatrics, College of Medicine, Inje University, Seoul, Seoul, Republic of Korea
  • 6 Department of Neuropsychiatry, Hallym University Dongtan Sacred Heart Hospital, Hwaseong, Gyeonggi, Republic of Korea
  • 7 Department of Psychiatry, Kangwon National University Hospital, Chuncheon, Gangwon, Republic of Korea
  • 8 Department of Psychiatry, Seoul St. Mary's Hospital, The Catholic University of Korea, Seoul, Seoul, Republic of Korea
  • 9 Department of Psychiatry, Chungnam National University Sejong Hospital, Sejong, Republic of Korea
  • 10 Department of Neuropsychiatry, Seoul National University Hospital, Seoul, Republic of Korea
  • 11 Department of Psychiatry, Seoul Metropolitan Government - Seoul National University Boramae Medical Center, Seoul, Republic of Korea

The final, formatted version of the article will be published soon.

    Background: Recently, there have been active proposals on how to utilize large language models (LLMs) in the fields of psychiatry and counseling. It would be interesting to develop programs with LLMs that generate psychodynamic assessments to help individuals gain insights about themselves, and to evaluate the features of such services. However, studies on this subject are rare. This pilot study aims to evaluate quality, risk of hallucination (incorrect AI-generated information), and client satisfaction with psychodynamic psychological reports generated by GPT-4.Methods: The report comprised five components: psychodynamic formulation, psychopathology, parental influence, defense mechanisms, and client strengths. Participants were recruited from individuals distressed by repetitive interpersonal issues. The study was conducted in three steps: 1) Questions provided to participants, designed to create psychodynamic formulations: 14 questions were generated by GPT for inferring psychodynamic formulations, while 6 fixed questions focused on the participants’ relationship with their parents. A total of 20 questions were provided. 2) Seven professors of psychiatry from different university hospitals evaluated the quality and risk of hallucinations in the psychological reports by reading the reports only, without meeting the participants. This quality assessment compared the psychological reports generated by GPT-4 with those inferred by the experts. 3) Participants evaluated their satisfaction with the psychological reports. All assessments were conducted using self-report questionnaires based on a Likert scale developed for this studyResults: A total of 10 participants were recruited, and the average age was 32 years. The median response indicated that quality of all five components of the psychological report was similar to the level inferred by the experts. The risk of hallucination was assessed as ranging from unlikely to minor. According to the median response in the satisfaction evaluation, the participants agreed that the report is clearly understandable, insightful, credible, useful, satisfying, and recommendable.Conclusion: This study suggests the possibility that artificial intelligence could assist users by providing psychodynamic interpretations.

    Keywords: artificial intelligence, Large language models, gpt, psychodynamic, formulation, Hallucination, Psychopathology, defence mechanisms

    Received: 31 Jul 2024; Accepted: 25 Feb 2025.

    Copyright: © 2025 Kim, Lee, Park, On, Lee, Keum, Oh, Song, Lee, Won, Shin, Lho, Hwang and Kim. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

    * Correspondence: Namwoo Kim, Department of Clinical Medical Sciences, Seoul National University Hospital, Seoul, Republic of Korea

    Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

    Research integrity at Frontiers

    Man ultramarathon runner in the mountains he trains at sunset

    94% of researchers rate our articles as excellent or good

    Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


    Find out more