Skip to main content

ORIGINAL RESEARCH article

Front. Psychiatry, 24 May 2022
Sec. Mood Disorders
This article is part of the Research Topic Interventions based on New Technologies for the Management of Mood Disorders View all 6 articles

Detection of Depression and Suicide Risk Based on Text From Clinical Interviews Using Machine Learning: Possibility of a New Objective Diagnostic Marker

\r\nDaun Shin&#x;Daun Shin1†Kyungdo Kim&#x;Kyungdo Kim2†Seung-Bo LeeSeung-Bo Lee3Changwoo LeeChangwoo Lee4Ye Seul BaeYe Seul Bae4Won Ik ChoWon Ik Cho5Min Ji KimMin Ji Kim1C. Hyung Keun ParkC. Hyung Keun Park6Eui Kyu Chie,Eui Kyu Chie4,7Nam Soo KimNam Soo Kim5Yong Min Ahn,,*\r\nYong Min Ahn1,8,9*
  • 1Department of Neuropsychiatry, Seoul National University Hospital, Seoul, South Korea
  • 2Transdisciplinary Department of Medicine & Advanced Technology, Seoul National University Hospital, Seoul, South Korea
  • 3Department of Medical Information, Keimyung University School of Medicine, Daegu, South Korea
  • 4Office of Hospital Information, Seoul National University Hospital, Seoul, South Korea
  • 5Department of Electrical and Computer Engineering and INMC, Seoul National University College of Engineering, Seoul, South Korea
  • 6Department of Psychiatry, Asan Medical Center, Seoul, South Korea
  • 7Department of Radiation Oncology, Seoul National University College of Medicine, Seoul, South Korea
  • 8Department of Psychiatry, Seoul National University College of Medicine, Seoul, South Korea
  • 9Institute of Human Behavioral Medicine, Seoul National University Medical Research Center, Seoul, South Korea

Background: Depression and suicide are critical social problems worldwide, but tools to objectively diagnose them are lacking. Therefore, this study aimed to diagnose depression through machine learning and determine whether it is possible to identify groups at high risk of suicide through words spoken by the participants in a semi-structured interview.

Methods: A total of 83 healthy and 83 depressed patients were recruited. All participants were recorded during the Mini-International Neuropsychiatric Interview. Through the suicide risk assessment from the interview items, participants with depression were classified into high-suicide-risk (31 participants) and low-suicide-risk (52 participants) groups. The recording was transcribed into text after only the words uttered by the participant were extracted. In addition, all participants were evaluated for depression, anxiety, suicidal ideation, and impulsivity. The chi-square test and student’s T-test were used to compare clinical variables, and the Naive Bayes classifier was used for the machine learning text model.

Results: A total of 21,376 words were extracted from all participants and the model for diagnosing patients with depression based on this text confirmed an area under the curve (AUC) of 0.905, a sensitivity of 0.699, and a specificity of 0.964. In the model that distinguished the two groups using statistically significant demographic variables, the AUC was only 0.761. The DeLong test result (p-value 0.001) confirmed that the text-based classification was superior to the demographic model. When predicting the high-suicide-risk group, the demographics-based AUC was 0.499, while the text-based one was 0.632. However, the AUC of the ensemble model incorporating demographic variables was 0.800.

Conclusion: The possibility of diagnosing depression using interview text was confirmed; regarding suicide risk, the diagnosis accuracy increased when demographic variables were incorporated. Therefore, participants’ words during an interview show significant potential as an objective and diagnostic marker through machine learning.

Introduction

Depression and suicide are emerging as important problems worldwide. The lifetime prevalence of depression in the general population has been shown to range between 10 and 15% and has been rapidly increasing over recent decades (1, 2). In addition, unipolar depression is predicted to become the second leading cause of death by 2030 (3). The suicide rate has also increased worldwide by 6.7% over 26 years, and in many European, North American, and Asia-Pacific countries, suicide ranks among the top 10 leading causes of death (4). Moreover, it is known that the risk of death by suicide in people with depression is more than 20 times higher that of the general population, with approximately 15–20% of patients with depression ending their own lives (1, 5).

Early and accurate diagnosis is necessary to implement effective interventions for depression and suicide. However, at present, the only way to diagnose depression and suicidal tendencies is to rely on the patient’s subjective report of symptoms. The diagnosis of depression is made as per the Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition (DSM-V) or International Classification of Diseases and Related Health Problems, 10th Edition (ICD-10). Based on the DSM-V, a major depressive disorder can be diagnosed when five or more depressive symptoms occur and last for two weeks or longer, including (1) depressed mood or (2) loss of interest or pleasure. The ICD-10 diagnostic criteria are similar to those of the DSM and are based on the patients’ subjective report of symptoms (6, 7). Therefore, depression often leads to under-diagnosis in primary care settings, where it is difficult for patients to under-report or have an in-depth interview about symptoms (8, 9). Moreover, in the case of suicide, suicidal behavior disorder was recently defined by the DSM-V but the manual does not specify how to evaluate the actual intention to commit suicide (6, 7). In addition, various scales to assess suicide risk have been developed. Among them, the Columbia-Suicide Severity Rating Scale (C-SSRS) is considered a gold standard for the clinical field (10, 11). However, the gold standard for suicide research remains unclear (12).

By overcoming subjectivity, various attempts have been made to detect and address depression and suicidal ideation. Studies have been conducted to identify depression based on the fact that people with depression exhibit greater hesitation in their voice and a monotonous tone (13). Studies are also underway that aim to predict depression and suicide risk by clustering texts published on social media (14, 15). A study comparing the detection of depression through speech and text showed slightly more accurate results for text (16). Some studies attempt to detect suicide risk early through social media based on the patterns in the texts used by people who committed suicide (1719). Suicide victims tend to use more the word for a future point in time in their notes than those attempting suicide, and differences in texts were also confirmed, such as expressing positive emotions (20). However, existing studies have mostly been conducted based on written texts, such as social media posts. To the best of our knowledge, no research has been conducted on differences between the texts of patients with depression and those at high risk of suicide based on actual clinical interviews.

Therefore, based on the words used by participants during a semi-structured interview called the Mini-International Neuropsychiatric Interview (MINI), commonly used in the field of mental health, this study established an algorithm that can detect depression and high-suicide-risk groups and examined its accuracy. We hypothesize that there is a difference between the text of people with depression and those with a high suicide risk. The confirmation of our hypothesis can help diagnose depression and predict high-risk suicide cases through artificial intelligence. This study aims to confirm the possibility of using text as an objective marker that can accurately diagnose depression and high suicide risk.

Materials and Methods

Participants

Adults aged 19–65 years were recruited among patients with depressive symptoms attending the mood disorder clinic at Seoul National University Hospital from 10 January 2019 to 31 August 2021. A healthy control group was recruited through Internet promotion and notices on nearby campuses. Adults aged 19–65 years, who were healthy and had no depression symptoms, participated in the study. We excluded from our study those: with an impaired ability to independently read and comprehend questionnaires, diagnosed with borderline intellectual disability or dementia, with a history of intracranial surgery, with a history of psychosis, unable to voice themselves due to laryngeal surgery or disease, and cases with significant changes in voice. The participants’ psychiatric diagnosis was confirmed through the MINI. Additionally, in the healthy group, if a previous or present mental illness was found through the MINI, the participant was excluded from the study. Two participants who did not complete the self-report questionnaire were also excluded from this study.

After fully understanding the explanation of the study, all participants signed a consent form to participate in the study, in line with the Declaration of Helsinki. The research procedure was approved by the Institutional Review Committee of Seoul National University Hospital (1812-081-995).

Assessment

Demographics such as sex, age, height, weight, socioeconomic status (SES), and drugs taken due to non-psychiatric conditions were assessed. Body Mass Index (BMI) was calculated through the recorded height and weight. A precise psychiatric diagnosis was confirmed using the MINI version 7.0.2. Among the items of the MINI, participants with a 1-month suicide-risk assessment and high-degree risk were added to the “depression with high suicidal risk” (DHSR) group. Participants with low, moderate, or no risk of suicide within a month were included in the “depression with low suicidal risk (DLSR)” group.

Antipsychotic drugs taken by the participants can cause changes in speech, such as monotonizing the tone of the voice because of extrapyramidal symptom (21). Therefore, in this study, to consider the effect of antipsychotic drugs, the doses of all antipsychotic drugs being taken were substituted with the olanzapine equivalent dose and summed (22).

Other factors such as depression, impulsivity, and suicidal thoughts were also evaluated. The Patient Health Questionnaire-9 (PHQ-9) was used to assess participants’ subjective depression. The PHQ-9 was developed as a screening scale for depression and comprises nine items rated on a 4-point Likert scale ranging from 0 (not at all) to 3 (nearly every day). Scores of 10 points or higher indicate moderate to severe depression (23, 24). The Hamilton Depression Rating Scale (HDRS) was used to evaluate objective depression. The HDRS comprises 17 items related to depression severity, and each item was rated using a 5-point Likert scale ranging from 0 (not present) to 4 (severe). Anxiety was evaluated using the Beck Anxiety Inventory (BAI). The BAI comprises 21 items rated on a 4-point Likert scale ranging from 0 (not at all) to 3 (severely) (25). Based on the findings of a meta-analysis conducted in 2016, a determination of pathological anxiety was suggested for scores above 16 points (26). When the scores for each item were summed, a score of 17–23 indicated moderate depression, and 24 or higher indicated severe depression (27, 28). Suicide risk was assessed using Beck’s Suicidal Ideation Scale (BSS) (29). The BSS consists of 19 items focused on the intention to commit suicide, and although there is no set cutoff, the higher the score, the higher the suicide risk (17, 30). In addition, since suicide is associated with impulsivity, the latter was assessed using the Barratt Impulsivity Scale (BIS). The BIS-11 consists of 30 questions rated using a 5-point Likert scale ranging from 1 (never) to 4 (always) (31, 32).

Text Extraction

During the research, the entire duration of each MINI session was recorded audio; subsequently, only the participant’s words were extracted by tagging the time at which the participant’s speech started and ended. Thereafter, speech shorter than 3 s in length was removed under the assumption that there would be little content, and the remaining speech was segmented into 10-s sections. The segmented speech file was converted into text using a speech recognition toolkit, a Python library. The speech recognizer in the toolkit used Google API, and it was set to “ko-KR” for recognizing Korean speech. For each case, the text of all utterances was recorded independently, and any speech not recognized during this process was dropped. The most widely used Korean preprocessing library, the Konlpy package, was used to extract necessary text information from sentences. Part-of-speech tagging was performed based on a morphological analysis of sentences, and a dataset was constructed by removing stop words (33).

Consequently, the sequence of the words from which the stop word was removed matched to each participant. The model we used in this study used frequency information rather than words’ positional information. Therefore, we again matched all words with each participant after splitting all sentences uttered by the participant into word format. Our model learns the probability of a participant-specific label (depression or risk of suicide) based on the frequency of a word in the data. In the evaluation process, sentences entering the model were separated into words, part-of-speech tagging was performed, stop words were removed, and then matched to the participant again. The model outputs the probability of whether the participant has depression or is at high risk of suicide based on their uttered word sequence.

Data Analysis

Categorical variables were compared using the chi-square test, and continuous variables were analyzed using student’s t-test. Analyses were conducted using IBM SPSS Statistics for Windows (version 25.0; SPSS Inc., Chicago, IL, United States).

The experiment was conducted after building the data pipeline and model pipeline. First, in the data pipeline, fivefold cross-validation was employed. In numerous machine learning studies, K-fold cross-validation is mainly used to verify model rigor and data efficiency. As the interview length differed depending on the participant, the data were divided into the 80% training and 20% test sets based on the participant. First, interviews of the training participants were divided into word units; then, training was carried out according to the label (depression or risk of suicide). At this time, the model was designed based on the Naive Bayes algorithm.

The Naive Bayes classifier is a conditional probability-based machine learning algorithm that calculates the probability of data belonging to each class. In the text domain, the Naive Bayes classifier counts the frequency of words appearing in the entire sentence and then trains a statistical model based on it. In this study, the classifier learned to arrive at a distribution of word frequencies according to the control group and target group (depression and high suicidal risk group). Subsequently, based on the participant’s words, the probability of belonging to each group was predicted.

In particular, for the demographic ensemble model, demographic data, which are structured data, were converted into a probability density function. In this process, because the number of each group is more than 30, a Gaussian normal distribution was assumed (3437). After generating a Gaussian density distribution from a given demographic bin distribution, the probability of the demographic feature of the target participant belonging to each group was calculated. Subsequently, the ensemble model was built by implementing this probability in the text-based Naive Bayes classifier. The Naive Bayes classifier was used by the Natural Language Toolkit package, and the scikit-learn package was employed for density estimation and analyses (38). All experiments were conducted through fivefold cross-validation.

During training, our model learned word frequency–label relationships. Thereafter, based on what the test participant said in the interview, a prediction based on probability was made. After one training and evaluation in this way, the data were newly split and divided into novel training participant and novel test participant. After this, the model parameters were reset, and training and prediction were performed on the newly split participants. This process was repeated five times, while the data split was performed by setting the participant to be included as a whole in the validation group only once.

Results

Comparison Between Healthy Control and Participants With Current Depression

Eighty-three patients with depression were recruited after two patients, who did not submit a self-report questionnaire, were excluded from the analysis. In the healthy control group, 83 participants were evaluated after 22 participants, with a lifetime prevalence of mental illness in the MINI, were excluded from the analysis.

Regarding demographics, the healthy control (HC) group was older and had a lower SES on average, compared to the group with current depression (CD). We also found more people taking other drugs in the CD group compared to the HC group (Table 1). When comparing clinical variables, depression, anxiety, and risk of suicidal ideation were statistically higher in the CD group. However, there was no difference between the two groups in BIS (Table 2).

TABLE 1
www.frontiersin.org

Table 1. Demographic comparison of HC and CD groups.

TABLE 2
www.frontiersin.org

Table 2. Differences in clinical characteristics between HC and CD groups.

The total number of words spoken by the participants was 21,376, of which rare words, spoken only once, accounted for 4.49%. The maximum sentence length spoken by one participant was 1,504 characters, while the average sentence length was 33.91 characters.

The demographic model for distinguishing between HC and CD included age, SES, BMI, non-psychiatric medication use, and sex, all of which showed statistical differences between the two groups. As for the current depression diagnostic model, the area under the curve (AUC) of the model using demographic information was 0.761, while that of the model using text was 0.902. By contrast, there was no significant difference between the ensemble model trained using demographic and text information and the model trained using only text information (Table 3 and Figure 1). When comparing the ROC curve using demographic data and text through the DeLong test, a p-value of 0.001 was confirmed. The statistical significance of this result was confirmed when only demographic data were used and in the ensemble model. By checking the words that the model evaluated as important in the classification process, it was confirmed that the proportion of words with a negative connotation was higher in the CD group (see Figure 2).

TABLE 3
www.frontiersin.org

Table 3. Differences in classification results between HC and CD groups.

FIGURE 1
www.frontiersin.org

Figure 1. Differences in classification results between healthy control group and current depression group.

FIGURE 2
www.frontiersin.org

Figure 2. Feature importance for classifying current depression group.

In addition, when text mining was performed using words with high frequency, and when the relative importance of the words used by patients with depression was evaluated, the prominent words had a negative connotation, such as “hospitalization,” “mistake,” “negation,” and “floor,” and were found to be widely used (Figure 3).

FIGURE 3
www.frontiersin.org

Figure 3. Text mining of important features in current depression group.

Comparison Between Depression With Low Suicide Risk and Depression With High Suicide Risk

Among the 83 patients, 31 were classified as suffering from depression with high suicide risk (DHSR) based on the MINI, and the remaining 52 participants were classified as having depression with low suicide risk (DLSR).

The DHSR group was statistically significantly younger and was on a higher dose of antipsychotic drugs. There were no other differences related to BMI, sex, and SES. When looking at the diagnosis through MINI, the ratios of major depressive disorder and bipolar disorder were similar in both groups, while panic disorder and social anxiety disorder were more common in the DHSR group (Table 4.) As for clinical variables, all indicators of depression, anxiety, and suicidal ideation, except for BIS, were higher in the DHSR group (Table 5).

TABLE 4
www.frontiersin.org

Table 4. Differences in demographics between DLSR and DHSR groups.

TABLE 5
www.frontiersin.org

Table 5. Differences in clinical characteristics between DLSR and DHSR groups.

In the model using demographics as a predictor, the DLSR and DHSR groups were classified by including statistically significant differences between the groups related to age, antipsychotic drug dose, SES, BMI, and the BAI score. The AUC of the model trained solely with demographic information was 0.499, showing the lowest accuracy, while the AUC was approximately 0.632 even when DHSR was classified using text, distinguishing between DLSR and DHSR. However, in the case of the ensemble model training, combining the features used in the demographic and text models, the AUC was 0.800, being the highest among all models (Table 6 and Figure 4).

TABLE 6
www.frontiersin.org

Table 6. Differences in classification results between DLSR and DHSR groups.

FIGURE 4
www.frontiersin.org

Figure 4. Differences in classification results between depression with low suicidal risk group and depression with high suicidal risk group.

Discussion

This study aimed to diagnose depression based on the words spoken by participants in a semi-structured interview (MINI) and determine whether suicide risk among patients with depression could be predicted based on such textual analysis. We found that the accuracy of detecting depression using speech-converted-text was 83.1%, and the specificity was 96.4%, resulting in better predictions compared to the model using only demographics; the results were statistically significantly different in both groups. In detecting depression, the ensemble model trained with demographics showed similar performance to the model predicting solely through text. However, when predicting the risk of suicide in patients with depression, the sensitivity of the model that predicted high risk solely through text was 74.4%, with its specificity being 47.7%; this result cannot be considered a good performance. In predicting the groups’ suicide risk, when the ensemble model incorporated demographics, better performance was confirmed—a sensitivity of 81.6%, specificity of 64.7%, and AUC of 0.800. This ensemble model performed better than the model based only on demographics. Moreover, the words used in the model to classify depression are clinically noteworthy, such as “hospitalization,” being used by patients themselves. Hence, it is possible to extract important clinically noteworthy features based on machine learning, accumulated by clinicians through experience and assessed based on interviews.

This study is the first to determine whether it is possible to distinguish between healthy groups and those affected by depression, based on the text derived from participants’ speech in semi-structured interviews, and to identify high-suicide-risk groups. Moreover, this study confirmed the diagnosis result by text, developed an algorithm to predict the high-suicide-risk group by integrating text and demographics, and confirmed a surprising AUC value of 79.95%. Previous studies have been conducted based on emotion classification through natural language processing and have mostly performed analysis based on fixed texts written in electronic medical records or social media (39, 40). A previous study predicted depression through a natural language analysis of Twitter posts. The accuracy of predicting depression was 83%, and the F1 score was 0.8329 (41). However, our study showed that the AUC of diagnosing depression using interview text was over 0.9. This is because we employed text taken from an interview with relatively open questions, increasing the possibility of its accuracy. However, when only text was used to predict suicide risk, the AUC was 0.632. This may be due to data imbalance since the number of individuals at suicide risk is only 37.349% among depressed patients. If more data for the patients’ group can be collected, accuracy can be improved.

This study has many strengths, as it is the first to diagnose depression in patients using speech-converted text and to evaluate the risk of suicide. First, the diagnosis was made based on text used in real-time interviews. A toolkit was used to convert the interview into text, making it possible to apply this algorithm in the actual clinical field. In particular, the study can predict depression with high accuracy even without refined data because it uses text that reflects data loss or errors that may occur while transforming the recording file into a toolkit. Second, it builds an ensemble algorithm that can utilize text and demographic data. In this study, using only text data had limitations in predicting groups at high risk for suicide. Therefore, an ensemble model that can utilize various clinical data was constructed. However, to confirm the accuracy of the text, the variables used in the ensemble model were minimized. This ensemble model confirmed its potential as a new diagnostic tool for classifying the risk of suicide. This ensemble model can be proposed as a new objective indicator in psychiatric diagnosis, if elaborated by including more diverse text data and clinical variables. This depression and suicide risk diagnosis algorithm based on artificial intelligence can be seen as a clinical decision support system, and it could help clinicians diagnose depression and suicide risk in various clinical settings, such as when a general practitioner needs to diagnose depression or when visiting the emergency room after an incident related to suicide. An appropriate diagnosis, along with therapeutic intervention, can create more strong therapeutic alliance to better serve patients and caregivers.

There are several limitations to this study. First, the text content used is based on the speech collected from the patients answering the MINI questions, so there is a possibility that critical data were not included. However, it was confirmed that the primary words used by patients with depression were different compared to those used by patients that did not have depression, even if the corresponding question was the same for both groups. This study was conducted based on the text of a semi-structured interview; therefore, further studies based on the speech used in free interviews are required. If data are collected based on a free interview between a clinician and a patient, the range of words used by the patient is likely to be more expansive, and the accuracy of diagnosis through text may decrease. Hence, research is required based on more extensive data, and it is hoped that this study will encourage greater use of the text from interviews for analysis. Second, in the cause-and-effect relationship of the depression discrimination algorithm of this study, whether the text changed due to depression was unclear. Therefore, future studies must confirm the change in text usage patterns according to the change in depression through longitudinal data collection and analysis. Third, as most participants in this study were female (80%) and had experienced bipolar disorder (86.75%), selection bias cannot be ignored. Moreover, since this study was conducted with participants in tertiary medical institutions, words such as “hospitalization” will likely be used. Therefore, it is necessary to confirm whether the results can be replicated based on interview content involving various age groups and clinical sites. Fourth, the amount of text used in this study was small, and deep learning-based analysis could not be applied because the number of participants, especially in the high-risk group, was small. However, meaningful classification was performed only with answers to these limited questions, with the small data size creating a basis for conducting future research with more participants and diverse text content. Fifth, in textual analysis, emotional language classification through natural language processing has been studied extensively. However, such an analysis was not performed in this study due to the limitations in the emotion classification system for Koreans. Although many Korean-based natural language classification datasets have been evaluated, the meaning of Korean words is often reversed depending on the adverb or intonation that follows a word, and many words have not yet been included in the classification. Hence, it could not be used in this study. If the classification of emotional language is more straightforward and can be used for analysis in future studies, the text will be more valuable as a diagnostic tool. Finally, since the text used for analysis did not go through a manual pre-processing evaluation step, there is an error rate to the toolkit, and it may have been evaluated as being lower than the actual rate.

Although limited semi-structured interviews were used, and latest analysis techniques, such as emotional analysis, were not applied, this study confirmed the possibility that the text derived from participant interviews can be an important objective marker for diagnosing depression and detecting suicide risk.

In this study, based on the words spoken by the participants in the MINI interview, depression was detected through machine learning based on the Naive Bayes classifier technique, and the accuracy was confirmed by constructing an ensemble model that predicts the risk of suicide among patients with depression. Detecting depression using text only showed an AUC of 0.905 and predicting high-suicide-risk among such patients showed an AUC of 0.632. In diagnosing depression, speech-converted text showed potential as a good objective marker. In predicting suicide risk, text showed diagnostic utility with an AUC of 0.800 when used with demographics. Whether the results of this study can be replicated will require additional research based on various interviews with more diverse participants.

Data Availability Statement

The raw data supporting the conclusions of this article are provided upon request from the corresponding author after being reviewed for feasibility.

Ethics Statement

The study research procedure was approved by the Institutional Review Board of the Seoul National University Hospital (1812-081-995). The patients/participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author Contributions

DS, WC, CH, NK, and YA designed the study protocol. DS, MK, and YA recruited the participants. DS, KK, S-BL, CL, and WC analyzed the data and prepared the figures and tables. DS and KK contributed majorly to the manuscript writing. DS, YB, MK, EC, NK, and YA administrated the study. S-BL, YB, MK, CP, EC, NK, and YA edited and revised the manuscript. All authors reviewed the manuscript.

Funding

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) and was funded by the Ministry of Education (Grant No. NRF-2018R1D1A1A02086027). This work was also supported by the Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korean Government (MSIT) (No. 2021-0-00312, development of non-face-to-face patient infection activity prediction and protection management SW technology at home and community treatment centers for effective response to infectious disease). The funding source was not involved in the study design, data collection, analysis, or interpretation, including any decision to write and publish the thesis.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We thank everyone who participated in this study.

References

1. Lépine JP, Briley M. The increasing burden of depression. Neuropsychiatr Dis Treat. (2011) 7(Suppl. 1):3–7. doi: 10.2147/ndt.s19617

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Andersen I, Thielen K, Bech P, Nygaard E, Diderichsen F. Increasing prevalence of depression from 2000 to 2006. Scand J Public Health. (2011) 39:857–63. doi: 10.1177/1403494811424611

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Mathers CD, Loncar D. Updated Projections of Global Mortality and Burden of Disease, 2002–2030: Data Sources, Methods and Results. Geneva: World Health Organization (2005).

Google Scholar

4. Naghavi M. Global burden of disease self-harm collaborators. global, regional, and national burden of suicide mortality 1990 to 2016: systematic analysis for the global burden of disease study 2016. BMJ. (2019) 364:l94. doi: 10.1136/bmj.l94

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Goodwin FK, Jamison KR, Ghaemi SN. Manic-Depressive Illness: Bipolar Disorders and Recurrent Depression. 2nd ed. New York, NY: Oxford University Press (2007). p. 247–69.

Google Scholar

6. American Psychological Association. Diagnostic and Statistical Manual of Mental Disorders (DSM-5§). Arlington, TX: American Psychiatric Publishing (2013).

Google Scholar

7. World Health Organization. International Statistical Classification of Diseases and Related Health Problems, Tenth Revision. (Vol. 1). Geneva: World Health Organization (1992).

Google Scholar

8. Sheehan DV. Depression: underdiagnosed, undertreated, underappreciated. Manag Care. (2004) 13:6–8.

PubMed Abstract | Google Scholar

9. Davidson JR, Meltzer-Brody SE. The underrecognition and undertreatment of depression: what is the breadth and depth of the problem? J Clin Psychiatry. (1999) 60(Suppl. 7):4–9; discussion 10.

PubMed Abstract | Google Scholar

10. Watson D, Goldney R, Fisher L, Merritt M. The measurement of suicidal ideation. Crisis. (2001) 22:12–4.

Google Scholar

11. Giddens JM, Sheehan KH, Sheehan DV. The Columbia-suicide severity rating scale (C–SSRS): has the ‘gold standard’ become a liability? Innov Clin Neurosci. (2014) 11:66–80.

PubMed Abstract | Google Scholar

12. McCall WV, Porter B, Pate AR, Bolstad CJ, Drapeau CW, Krystal AD, et al. Examining suicide assessment measures for research use: using item response theory to optimize psychometric assessment for research on suicidal ideation in major depressive disorder. Suicide Life Threat Behav. (2021) 51:1086–94. doi: 10.1111/sltb.12791

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Shin D, Cho WI, Park CHK, Rhee SJ, Kim MJ, Lee H, et al. Detection of minor and major depression through voice as a biomarker using machine learning. J Clin Med. (2021) 10:3046. doi: 10.3390/jcm10143046

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Magami F, Digiampietri LA. Automatic detection of depression from text data: a systematic literature review. In: Proceedings of the XVI Braz Symposium on Information Systems. New York, NY (2020).

Google Scholar

15. Yang W, Mu L. GIS analysis of depression among twitter users. Appl Geogr. (2015) 60:217–23. doi: 10.1016/j.apgeog.2014.10.016

CrossRef Full Text | Google Scholar

16. Morales MR, Levitan R. Speech vs. text: a comparative analysis of features for depression detection systems. In: Proceedings of the 2016 IEEE Spoken Language Technology Workshop (SLT). San Diego, CA: IEEE Publications (2016).

Google Scholar

17. Lai K, Li D, Peng H, Zhao J, He L. Assessing suicide reporting in top newspaper social media accounts in China: content analysis study. JMIR Ment Health. (2021) 8:e26654. doi: 10.2196/26654

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Sawhney R, Joshi H, Gandhi S, Jin D, Shah RR. Robust suicide risk assessment on social media via deep adversarial learning. J Am Med Inform Assoc. (2021) 28:1497–506. doi: 10.1093/jamia/ocab031

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Parrott S, Britt BC, Hayes JL, Albright DL. Social media and suicide: a validation of terms to help identify suicide-related social media posts. J Evid Based Soc Work. (2020) 17:624–34. doi: 10.1080/26408066.2020.1788478

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Handelman LD, Lester D. The content of suicide notes from attempters and completers. Crisis. (2007) 28:102–4. doi: 10.1027/0227-5910.28.2.102

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Bouloudnine S, Azorin JM, Petitjean F, Parry-Pousse P, Sciorato F, Blin O, et al. [Perception of atypical antipsychotics’ side effects through speech analysis of schizophrenic patients. TALK Study]. Encéphale. (2011) 37(Suppl. 2):S143–50. doi: 10.1016/S0013-7006(11)70042-2

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Leucht S, Samara M, Heres S, Patel MX, Furukawa T, Cipriani A, et al. Dose equivalents for second-generation antipsychotic drugs: the classical mean dose method. Schizophr Bull. (2015) 41:1397–402. doi: 10.1093/schbul/sbv037

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Kroenke K, Spitzer RL, Williams JB. The PHQ-9: validity of a brief depression severity measure. J Gen Intern Med. (2001) 16:606–13. doi: 10.1046/j.1525-1497.2001.016009606.x

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Levis B, Benedetti A, Thombs BD. DEPRESsion screening data (DEPRESSD) collaboration. accuracy of patient health questionnaire-9 (PHQ-9) for screening to detect major depression: individual participant data meta-analysis. BMJ. (2019) 365:l1476. doi: 10.1136/bmj.l1476

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Beck AT, Epstein N, Brown G, Steer RA. An inventory for measuring clinical anxiety: psychometric properties. J Consult Clin Psychol. (1988) 56:893–7. doi: 10.1037/0022-006x.56.6.893

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Bardhoshi G, Duncan K, Erford BT. Psychometric meta−analysis of the English version of the beck anxiety inventory. J Couns Dev. (2016) 94:356–73. doi: 10.1097/MD.0000000000020416

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Zimmerman M, Martinez JH, Young D, Chelminski I, Dalrymple K. Severity classification on the Hamilton depression rating scale. J Affect Disord. (2013) 150:384–8. doi: 10.1016/j.jad.2013.04.028

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Williams JB. Standardizing the Hamilton depression rating scale: past, present, and future. Eur Arch Psychiatry Clin Neurosci. (2001) 251(Suppl. 2):II6–12. doi: 10.1007/BF03035120

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Beck AT, Kovacs M, Weissman A. Assessment of suicidal intention: the Scale for Suicide Ideation. J Consult Clin Psychol. (1979) 47:343–52. doi: 10.1037/0022-006x.47.2.343

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Range LM, Knott EC. Twenty suicide assessment instruments: evaluation and recommendations. Death Stud. (1997) 21:25–58. doi: 10.1080/074811897202128

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Cole AB, Littlefield AK, Gauthier JM, Bagge CL. Impulsivity facets and perceived likelihood of future suicide attempt among patients who recently attempted suicide. J Affect Disord. (2019) 257:195–9. doi: 10.1016/j.jad.2019.07.038

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Barratt ES. Anxiety and impulsiveness related to psychomotor efficiency. Percept Mot Skills. (1959) 9:191–8. doi: 10.2466/pms.1959.9.3.191

CrossRef Full Text | Google Scholar

33. Park EL, Cho S. KoNLPy: Korean natural language processing in Python. In: Proceedings of the 26th Annual Conference on Human & Cognitive Language Technology. Human and Language Technology. (Vol. 6), Chuncheon (2014). p. 133–6.

Google Scholar

34. Tiku M. A new statistic for testing for normality. Commun Stat Theor Methods. (1974) 3:223–32. doi: 10.1080/03610927408827123

CrossRef Full Text | Google Scholar

35. Bullpit CJ. Confidence intervals. Lancet. (1987) 1:494–7.

Google Scholar

36. Öztuna D, Elhan AH, Tüccar E. Investigation of four different normality tests in terms of type 1 error rate and power under different distributions. Turk J Med Sci. (2006) 36:171–6.

Google Scholar

37. Ghasemi A, Zahediasl S. Normality tests for statistical analysis: a guide for non-statisticians. Int J Endocrinol Metab. (2012) 10:486–9. doi: 10.5812/ijem.3505

PubMed Abstract | CrossRef Full Text | Google Scholar

38. Bird S, Klein E, Loper E. Natural Language Processing with Python: Analyzing Text With the Natural Language Toolkit. Sebastopol, CA: O’Reilly Media, Inc (2009).

Google Scholar

39. McCoy TH Jr., Castro VM, Roberson AM, Snapper LA, Perlis RH. Improving prediction of suicide and accidental death after discharge from general hospitals with natural language processing. JAMA Psychiatry. (2016) 73:1064–71. doi: 10.1001/jamapsychiatry.2016.2172

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Coppersmith G, Leary R, Crutchley P, Fine A. Natural language processing of social media as screening for suicide risk. Biomed Inform Insights. (2018) 10:1178222618792860. doi: 10.1177/1178222618792860

PubMed Abstract | CrossRef Full Text | Google Scholar

41. Deshpande M, Rao V. Depression detection using emotion artificial intelligence. In: Proceedings of the 2017 International Conference on Intelligent Sustainable Systems (ICISS). Palladam (2017). p. 858–62. doi: 10.1109/ISS1.2017.8389299

CrossRef Full Text | Google Scholar

Keywords: depression, suicide risk, machine learning, text analysis, objective marker

Citation: Shin D, Kim K, Lee S-B, Lee C, Bae YS, Cho WI, Kim MJ, Hyung Keun Park C, Chie EK, Kim NS and Ahn YM (2022) Detection of Depression and Suicide Risk Based on Text From Clinical Interviews Using Machine Learning: Possibility of a New Objective Diagnostic Marker. Front. Psychiatry 13:801301. doi: 10.3389/fpsyt.2022.801301

Received: 25 October 2021; Accepted: 04 April 2022;
Published: 24 May 2022.

Edited by:

Bénédicte Nobile, Hôpital Lapeyronie, France

Reviewed by:

Adrian Alacreu-Crespo, University of Valencia, Spain
Daniela Flores Mosri, Universidad Intercontinental, Mexico
Tamara Escrivá-Martínez, University of Valencia, Spain

Copyright © 2022 Shin, Kim, Lee, Lee, Bae, Cho, Kim, Hyung Keun Park, Chie, Kim and Ahn. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yong Min Ahn, YXltQHNudS5hYy5rcg==

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.