- 1Escuela de Psicología, Universidad Continental, Huancayo, Peru
- 2Facultad de Psicología, Universidad Científica del Sur, Lima, Peru
With the implementation of new dimensional models of personality disorder (PD) in the DSM-5 and ICD-11, several investigators have developed and evaluated the psychometric properties of measures of severity. The diagnostic accuracy of these measures, an important cross-cultural metric that falls between validity and clinical utility, remains unclear. This study aimed to analyze and synthesize the diagnostic performance of the measures designed for both models. For this purpose, searches were carried out using three databases: Scopus, PubMed, and Web of Science. Studies that presented sensitivity and specificity parameters for cut-off points were selected. There were no restrictions on the age and gender of the participants nor on the reference standard used or the settings. Study quality and synthesis were assessed using QUADAS-2 and MetaDTA software, respectively. Twelve studies were eligible covering self-reported and clinician-rated measures based on the ICD-11 and DSM-5 PD severity models. A total of 66.7% of the studies showed a risk of bias in more than 2 domains. The 10th and 12th studies provided additional metrics, resulting in a total of 21 studies for evidence synthesis. Adequate overall sensitivity and specificity (Se = 0.84, Sp = 0.69) of these measures were obtained; however, the cross-cultural performance of specific cut-off points could not be assessed due to the paucity of studies on the same measure. Evidence suggests that patient selection processes should mainly be improved (avoid case–control design), use adequate reference standards, and avoid only reporting metrics for the optimal cut-off point.
1. Introduction
PD is a common condition in the general population and is associated with negative outcomes for those who suffer from it and their families (1). The limited categorical conception of PD is changing towards a dimensional paradigm in current diagnostic systems (1, 2). A hybrid model is presented in the DSM-5 which combines specific categorical PD diagnoses with a dimensional Alternative Model of Personality Disorders (AMPD) to allow a smooth transition from its use to many practitioners who are accustomed to the earlier model. In the AMPD (section III of the DSM-5), criterion A is the first diagnostic step, since it allows the detection of PD (at the moderate level) and the assignment of the severity of its dysfunction from none, some, moderate, severe until extreme. Criterion B is then evaluated by assigning the maladaptive traits. In contrast, in ICD-11 the PD model is based mainly on a dimensional approach based on the severity of personality dysfunction and optionally on trait qualifiers and the borderline pattern.
Criterion A of the DSM-5 AMPD is operationalized by the Level of Personality Functioning Scale (LPFS; 3), an official measure rated by the physician to measure the patient’s personality dysfunction in four components and two domains self (identity and self-direction) and interpersonal (empathy and intimacy). Based on this measure, three semi-structured interviews have been developed: the Clinical Assessment of the Level of Personality Functioning Scale (CALF; 4), Structured Clinical Interview for the Level of Personality Functioning Scale (SCID-AMPD Module I; 5), and the Semi-Structured Interview for Personality Functioning DSM-5 (STiP 5.1; 6). Nine self-report measures have also been developed, such as the DSM-5 Levels of Personality Functioning Questionnaire (DLOPFQ; 7), and its short form (DLOPFQ-SF; 8), the Level of Personality Functioning Scale – Self-Report (LPFS-SR; 9), Level of Personality Functioning Scale – Brief Form (LPFS-BF; 10) and its second version (LPFS-BF 2.0; 11), Personality Functioning Scale (PFS; 12), Self and Interpersonal Functioning Scale (SIFS; 13), Levels of Personality Functioning Questionnaire for Adolescents from 12 to 18 Years (LoPF-Q 12–18; 14), and its short form (LoPF-Q 12–18 SF; 15).
ICD-11 severity has not been presented with an official measure, but several researchers have recently begun to develop them as CDDG guidelines for PD and related traits have been generated. The first measure developed was the Standardized Assessment of Severity of Personality Disorder (SASPD; 16) which was designed even before the final version of the guidelines was published. Other recent measures include the ICD-11 Personality Disorder Severity Scale (PDS-ICD-11; 17), Clark et al. scales (18), and PF scale of the Integrative Dimensional Personality Inventory for ICD-11 (IDPI-11; 19). Unlike criterion A of the DSM-5 AMPD, these measures have a unifactorial nature since self and interpersonal functioning are defined in a more interconnected way and linked to real-life consequences at moderate to severe levels, such as self-harm or harm to others and the reality test (20).
Diagnostic accuracy studies evaluate the performance of clinical tests (diagnostic tests), in terms of their ability to differentiate between individuals with and without the target condition, either with explanatory scientific objectives or with a pragmatic approach in clinical practice. This is done primarily through statistical analyses (e.g., sensitivity and specificity) that allow inferences to be drawn about the accuracy of clinical tests (21). Specifically, clinical tests are procedures for evaluating an individual’s current health status or predicting their future health status; and diagnostic accuracy studies provide evidence of tests for the diagnosis, staging, detection, monitoring, and surveillance of diseases (22). Improving the accuracy of the tests makes it possible for relevant referrals (or derivations) to be made, and given certain therapies to the correct patients. The clinical utility and validity of a model/measure are overlapping concepts (23) and diagnostic accuracy or precision is located differentially from the other metrics in this overlap.
Many validation studies of PD severity measures from the DSM-5 AMPD and ICD-11 models have included complementary diagnostic accuracy analyses. These studies have mainly focused on the internal structure and convergent validity of these measures, and the few studies that have made efforts to assess the precision of these measures have probably either performed them incorrectly or drawn imprecise inferences from limited methodology. Overcoming the arbitrary division into individuals with and without the disorder and exploiting the multiple gradations of severity – to improve the psychometric properties of measures of severity (1) – involves evaluating the sensitivity and specificity of each PD dysfunction threshold (target condition).
2. The current review
Reviews of studies on the accuracy of a diagnostic test aim to address the need for health decision makers to have access to relevant, up-to-date and high-quality information on the use of a diagnostic test as a tool for a specific setting (24). Several reviews have focused on analyzing the reliability, validity, and usefulness of PD severity measures based on the DSM-5 AMPD and ICD-11 models without delving into aspects of their diagnostic performance. Therefore, the current review aimed to determine the diagnostic accuracy of these measurements; since summarizing the literature published to date is necessary to make recommendations for clinical practice and to improve future research will be carried out. The research question was as follows: can the ICD-11 and DSM-5 severity measures be accurate for the detection of personality disorder in the general population?
We searched the literature systematically in three main databases Scopus, PubMed and Web of Science, without any language restriction by combining the following text strings: personality AND (disorder* OR patholog*) | dimension* | function* OR severi* | validity OR diagnos* OR assessment | ICD OR International Classification of Diseases | DSM-5 OR Diagnostic and Statistical Manual of Mental Disorders. The review was performed according to PRISMA-DTA (21, 25, 26) and Cochrane Handbook for Systematic Reviews of Diagnostic Test Accuracy-Version 2 (27). The search returned 531 results (2,625 in Scopus, 64 in Web of Science, and 91 in PubMed). There were no restrictions on the age and gender of the participants or for the reference standard used or the settings; because we assumed that the literature collected could be scarce. Only studies that presented sensitivity and specificity indices for one or more PD dysfunction thresholds in both models were included. The assessment of the risk of bias of the included studies was carried out using QUADAS-2 (28) and synthesis with MetaDTA v. 2.01 (29).
3. Results
3.1. Characteristics of included studies
Table 1 describes the 12 studies that represent evidence based on the subject over the last 10 years. The severity measures used in these studies include the PDS-ICD-11 and SASPD from the ICD-11 PD model; and the SIFS, LPFS-SR, LoPF-Q 12–18, LoPF-Q 12–18 SF, LPFS and algorithms of Criterion A from the PD model of the DSM-5 AMPD. These studies comprised measures administered in 12 countries (including 2 non-Western nations) and six languages. Eight of these studies used mixed samples – clinical and community – (14, 15, 30–35), and four studies used clinical samples (16, 36–38). Data from 8,390 participants were analyzed. On average, 55.9% were women; and the average age of adult and adolescent participants was 36.4 and 14.7, respectively. Study 4 (15) used data from study 7 (14); and study 12 (37), data from study 11 (36). For study 8 (34); although the target condition was initially ICD-11 PD severity, a measure designed to measure PD dysfunction according to the DSM-5 AMPD severity model was used; thus, we assigned the target condition to this last model.
Most studies that used mixed samples (case–control design) reported diagnostic accuracy metrics such as clinical utility statistics or discriminant or criterion validity. Only two studies reported these metrics as performance statistics and diagnostic accuracy (16, 38). Likewise, the third (16) and eleventh (36) studies reported sensitivity and specificity metrics for two or more PD dysfunction thresholds; on the other hand, study 8 (34) reported other dysfunction thresholds without these metrics. The fourth (15) and fifth (32) studies reported the optimal cut-off points and their diagnostic accuracy metrics according to the setting and reference standard, respectively. Only study 10 (38) reported additional sensitivity and specificity metrics for all cut-off points of the measure used as an index test. Finally, seven studies reported participant recruitment that reflected the dimensional spectrum of PD – e.g., students, outpatients, and hospitalized patients – (14, 16, 34, 36–38).
3.2. Results of the review
Due to the unanalyzed/unreported data in the reviewed studies, the diagnostic accuracy metrics provided in this section focus on the mild and moderate PD dysfunction thresholds of the ICD-11 and DSM-5 AMPD severity models, respectively. The sensitivity of the PDS-ICD-11 in the Spanish study was 0.80 and the specificity was 0.73. In the same way, a sensitivity between 0.75 and 0.85 and a specificity between 0.68 and 0.84 were found for the LoPF-Q 12–18. For the LPFS-SR, a sensitivity of 0.81 and a specificity of 0.74 were found. In addition, the sensitivity of the “any two” criteria A algorithm for the four areas of PD dysfunction ranged from 0.64 to 0.96 and its specificity from 0.29 to 0.85. Among the studies that highlighted specificity over sensitivity were those that evaluated the SASPD, SIFS, and LoPF-Q 12–18 SF. The sensitivity of SASPD ranged from 0.66 to 0.72 and its specificity ranged from 0.68 to 0.90. Similarly, the sensitivity of the SIFS was 0.79 and its specificity 0.86; likewise, the sensitivity of the LoPF-Q 12–18 SF was 0.88 and its specificity was 0.92.
3.3. Quality and synthesis of studies
A total of 66.7% of the studies showed a risk of bias in more than 2 domains (see Supplementary Table S1). Three studies showed bias in one domain; likewise, no study showed bias in two domains. Four studies showed bias in three domains, and in four studies we found bias in all four domains. The highest risk of bias occurred in the index test domain (91.7%), followed by the reference standard (66.7%), patient selection (58.3%), and flow and time (41.7%). To assign “risk” in each study we decided that two or more questions had to be answered affirmatively for the first two domains of QUADAS-2; while a single affirmative answer would imply an assignment of “not clear.” Five of the 12 studies showed a risk of bias in patient selection due to the case–control design used in their methodology and recruitment possibly for convenience (30–33, 35); which triggers spectrum and selection bias that could increase the sensitivity and specificity indices (39–42). The risk in this domain was not clear in two studies (14, 15), because they only used convenience sampling.
Eight of the 12 studies showed a risk of bias in the index test because there was no blinding of the results of the reference standard when applying the index test (14, 15, 30–35), generating a possible information bias that could overestimate the diagnostic performance metrics (39), and uniquely the optimal score was specified, which can also have the same effect (28). The risk in the index test was not clear for study 3 (16) due to the respective blinding, but only optimal cut-off points for mild and moderate levels of PD were reported. Eleven of the 12 studies showed bias in the reference standard because it did not correctly classify the target condition (14, 15, 30–38), causing misclassification bias or “copper standard” which can underestimate test accuracy scores (39–41). In this domain we decided to assign more weight to only one affirmative answer to assign high risk because several experts affirm that the reference standard should be the best available method to classify participants with and without the target condition (21).
In seven of the 12 studies, bias in flow and time was noted (14, 15, 30–34), since not all people received the same reference standard, generating partial verification bias that can increase sensitivity and reduce the specificity of the test (39–41). The risk in this domain was not clear for study 9 (35) because all participants had received the same reference standard (DSM-5 Section II PD semi-structured interviews) before but outside the study. In this domain we also decided to assign more weight to only one affirmative answer to assign high bias since the “multi reference standard” in the same analysis is a common negative practice in validation studies that has a significant effect on the interpretation of the results (43). There were no applicability concerns as the review question was open-ended with no exclusion criteria for patients, reference standard, index test, or recruitment settings. Studies 10 (38) and 12 (37) contributed to further analysis, generating a total of 21 studies for the synthesis of this review (Supplementary Table S2). As seen in Figure 1, the diagnostic accuracy metrics were individually appropriate for each study; which was also demonstrated in the HSROC plot. The statisticians. Se = 0.84, Sp = 0.69, FP rate = 0.31, logit(Se) = 1.6, logit(Sp) = 0.8 supports this assertion. Specific cut-off points could not be evaluated for each of the measures because of the insufficient number of studies. Supplementary Figure S2 shows the HSROC of studies with the QUADAS-2 domains.
Figure 1. Summary plots of the reviewed studies. Panel (A) shows the sensitivity forest plot, panel (B) shows the specificity forest plot and panel (C) shows the HSROC plot (of random effects) by index test.
4. Discussion
Many researchers and users of the DSM-5 and ICD-11 enthusiastically welcome the transition to a dimensional approach that is more valid, reliable, and useful for the evaluation and treatment of PD than the previous diagnostic systems (44, 45). Slight variations in the conceptualization of PD in the DSM-5 10 years ago have inspired a more radical change during the preliminary versions until the final version of the ICD-11 last year (45, 46). The severity of PD dysfunction is and will be the main requirement or decision tool in both models to define who will or will not receive treatment based on a known prognosis, how many professionals to hire, and how to manage health resources (47); at the same time, clinical and community actors are educated with a recuperative and preventive vision of PD instead of stigmatizing it (48). Therefore, it is important to precisely define whether the requirements in each of the thresholds of the PD (dys)functioning continuum are adequate for its diagnosis. This review is the first to delve into the diagnostic accuracy metrics reported by studies on PD severity measures of both diagnostic systems.
Much has been said about the good psychometric levels found in severity measures (1, 2, 49); however, in this review we have found fundamental errors in the methodology that impact the analyses of diagnostic accuracy. These errors include lack of blinding when applying the index test, uniquely reporting of optimal cut-off points, imperfect reference standards, case–control design, convenience sampling, and the application of multiple reference standards in the same analysis. This, in addition to the scarcity of studies, prevents us from providing cut-off points for each of the severity measures proposed for both models. Although we would have liked to find diagnostic accuracy literature of sufficient quality for this initial objective, the reviewed studies only allow us to offer a promising general mapping of the diagnostic performance of each of these DSM-5 AMPD severity measures and ICD-11. Consequently, this study corresponds to a scoping review, allowing us to warn that inappropriate practices in the design, methodology, analysis and reporting of results on the parameters of sensitivity and specificity are avoided.
Several of the reviewed studies only reported metrics to detect the presence or absence of PD – i.e., moderate and mild levels in DSM-5 and ICD-11, respectively – ; however, they did not explore the remaining spectrum of this condition or the subclinical threshold. We were also able to observe the confusion generated by the use of terms such as “criterion validity,” “discriminant validity,” “clinical utility” among others when diagnostic accuracy metrics were used. Therefore, we recommend that the sensitivity and specificity metrics are not used to assess the differential capacity of the measure with a case–control design. Often scientific hypotheses are valid for strengthening the concepts and statements – commonly applied in preclinical studies – (50). We better positioned diagnostic accuracy metrics as quantitative analyses of clinical utility (23, 46). This includes the use of large multicenter samples with suspected PD in a given setting who are administered the index test and the same ideal reference standard for the target condition in the same study. Only by following a rigorous methodology we can truly affirm that certain cut-off points are appropriate for decision-making in the care of patients with suspected PD. Perhaps these findings suggest considering more the use of projective tests such as the Rorschach or Thematic Apperception Test (TAT), which are currently underutilized in favor of easier-to-administer tools such as questionnaires.
5. Final observations
The diagnostic accuracy of a test includes a set of metrics that serve as a decision tool for healthcare professionals in assigning treatment to correct patients. Since the introduction of the dimensional approach to PD in current diagnostic systems, sensitivity and specificity indices have been reported for severity measures for this condition. In this paper we attempted to summarize these metrics through the reviewed studies; however, we found substantial deficiencies in their design that prevented us from achieving this objective. Despite these limitations, this study serves as a precedent to improve our methods if we want the PD severity measures of the DSM-5 AMPD and ICD-11 to really serve what they were created for.
Author contributions
All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.
Funding
The subsidy for the processing payment of this article was covered by Continental University (Company name: Universidad Continental S.A.C; RUC: 20319363221; Address: Av. San Carlos No. 1980, Huancayo, Peru) once accepted.
Acknowledgments
We thank the reviewers and handling editor for their comments regarding our paper for its improvement.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyt.2023.1209679/full#supplementary-material
References
1. Zimmermann, J, Hopwood, CJ, and Krueger, RF. The DSM-5 level of personality functioning scale In: RF Krueger and PH Blaney, editors. Oxford textbook of psychopathology. Oxford, UK: Oxford University Press (2023)
2. Birkhölzer, M, Schmeck, K, and Goth, K. Assessment of criterion a. Curr Opin Psychol. (2021) 37:98–103. doi: 10.1016/j.copsyc.2020.09.009
3. American Psychiatric Association. Diagnostic and statistical manual of mental disorders, fifth edition, text revision (DSM-5-TR™). Washington, DC: American Psychiatric Publishing (2022).
4. Thylstrup, B, Simonsen, S, Nemery, C, Simonsen, E, Noll, JF, Myatt, MW, et al. Assessment of personality-related levels of functioning: a pilot study of clinical assessment of the DSM-5 level of personality functioning based on a semi-structured interview. BMC Psychiatry. (2016) 16:1–8. doi: 10.1186/s12888-016-1011-6
5. Bender, DS, Skodol, AE, First, MB, and Oldham, J. Module I: structured clinical interview for the level of personality functioning scale In: MB First, AE Skodol, DS Bender, and JM Oldham, editors. User’s guide for the structured clinical interview for the DSM-5® alternative model for personality disorders. Arlington, VA: American Psychiatric Association (2018). SCID-5-AMPD.
6. Hutsebaut, J, Kamphuis, JH, Feenstra, DJ, Weekers, LC, and De, SH. Assessing DSM5-oriented level of personality functioning: development and psychometric evaluation of the semi-structured interview for personality functioning DSM5 (STiP-5.1). Personal Disord Theory Res Treat. (2017) 8:94–101. doi: 10.1037/PER0000197
7. Huprich, SK, Nelson, SM, Meehan, KB, Siefert, CJ, Haggerty, G, Sexton, J, et al. Introduction of the DSM-5 levels of personality functioning questionnaire. Personal Disord Theory Res Treat. (2018) 9:553–63. doi: 10.1037/per0000264
8. Siefert, CJ, Sexton, J, Meehan, K, Nelson, S, Haggerty, G, Dauphin, B, et al. Development of a short form for the DSM-5 levels of personality functioning questionnaire. J Pers Assess. (2019) 102:516–26. doi: 10.1080/00223891.2019.1594842
9. Morey, LC. Development and initial evaluation of a self-report form of the DSM-5 level of personality functioning scale. Psychol Assess. (2017) 29:1302–8. doi: 10.1037/pas0000450
10. Hutsebaut, J, Feenstra, DJ, and Kamphuis, JH. Development and preliminary psychometric evaluation of a brief self-report questionnaire for the assessment of the DSM-5 level of personality functioning scale: the LPFS brief form (LPFS-BF). Personal Disord Theory Res Treat. (2016) 7:192–7. doi: 10.1037/PER0000159
11. Weekers, LC, Hutsebaut, J, and Kamphuis, JH. The level of personality functioning scale-brief form 2.0: update of a brief instrument for assessing level of personality functioning. Personal Ment Health. (2019) 13:3–14. doi: 10.1002/pmh.1434
12. Stover, JB, Liporace, MF, and Solano, AC. Personality functioning scale: a scale to assess DSM-5’s criterion a personality disorders. Interpersona. (2020) 14:40–53. doi: 10.5964/ijpr.v14i1.3925
13. Gamache, D, Savard, C, Leclerc, P, and Côté, A. Introducing a short self-report for the assessment of DSM-5 level of personality functioning for personality disorders: the self and interpersonal functioning scale. Personal Disord Theory Res Treat. (2019) 10:438–47. doi: 10.1037/per0000335
14. Goth, K, Birkhölzer, M, and Schmeck, K. Assessment of personality functioning in adolescents with the LoPF-Q 12–18 self-report questionnaire. J Pers Assess. (2019) 100:680–90. doi: 10.1080/00223891.2018.1489258
15. Zimmermann, R, Steppan, M, Zimmermann, J, Oeltjen, L, Birkhölzer, M, Schmeck, K, et al. A DSM-5 AMPD and ICD-11 compatible measure for an early identification of personality disorders in adolescence – LoPF-Q 12–18 latent structure and short form. PLoS One. (2022) 17:e0269327. doi: 10.1371/journal.pone.0269327
16. Olajide, K, Munjiza, J, Moran, P, O’Connell, L, Newton-Howes, G, Bassett, P, et al. Development and psychometric properties of the standardized assessment of severity of personality disorder (SASPD). J Personal Disord. (2018) 32:44–56. doi: 10.1521/pedi_2017_31_285
17. Bach, B, Brown, TA, Mulder, RT, Newton-Howes, G, Simonsen, E, and Sellbom, M. Development and initial evaluation of the ICD-11 personality disorder severity scale: PDS-ICD-11. Personal Ment Health. (2021) 15:223–36. doi: 10.1002/PMH.1510
18. Clark, LA, Corona-Espinosa, A, Khoo, S, Kotelnikova, Y, Levin-Aspenson, HF, Serapio-García, G, et al. Preliminary scales for ICD-11 personality disorder: self and interpersonal dysfunction plus five personality disorder trait domains. Front Psychol. (2021) 12:2827. doi: 10.3389/fpsyg.2021.668724
19. Hualparuca-Olivera, L, Ramos, DN, Arauco, PA, and Coz, RM. Integrative dimensional personality inventory for ICD-11: development and evaluation in the Peruvian correctional setting. Liberabit. (2022) 28:e540–08. doi: 10.24265/liberabit.2022.v28n1.05
20. Bach, B, and Mulder, R. Clinical implications of ICD-11 for diagnosing and treating personality disorders. Curr Psychiatry Rep. (2022) 24:553–63. doi: 10.1007/s11920-022-01364-x
21. Salameh, JP, Bossuyt, PM, McGrath, TA, Thombs, BD, Hyde, CJ, Macaskill, P, et al. Preferred reporting items for systematic review and meta-analysis of diagnostic test accuracy studies (PRISMA-DTA): explanation, elaboration, and checklist. BMJ. (2020) 370:m2632. doi: 10.1136/BMJ.M2632
22. Deeks, JJ, and Bossuyt, PMM. Chapter 3: evaluating diagnostic tests. Draft version In: JJ Deeks, PMM Bossuyt, MMG Leeflang, and Y Takwoingi, editors. Cochrane handbook for systematic reviews of diagnostic test accuracy version 2. London, UK: Cochrane (2021)
23. Reed, GM, Keeley, JW, Rebello, TJ, First, MB, Gureje, O, Ayuso-Mateos, JL, et al. Clinical utility of ICD-11 diagnostic guidelines for high-burden mental disorders: results from mental health settings in 13 countries. World Psychiatry. (2018) 17:306–15. doi: 10.1002/WPS.20581
24. Flemyng, E, Cumpston, M, Arevalo-Rodriguez, I, Chandler, J, and Deeks, JJ. Chapter 2: planning a Cochrane review of diagnostic test accuracy. Draft version In: JJ Deeks, PMM Bossuyt, MMG Leeflang, and Y Takwoingi, editors. Handbook for systematic reviews of diagnostic test accuracy version 2. London, UK: Cochrane (2021). 2–13.
25. McInnes MDF, MI, Moher, D, Thombs, BD, TA, MG, and Bossuyt, PM, the PRISMA-DTA Group. Preferred reporting items for a systematic review and meta-analysis of diagnostic test accuracy studies: the PRISMA-DTA statement. JAMA. (2018) 319:388–96. doi: 10.1001/jama.2017.19163
26. Korevaar, DA, Bossuyt, PM, McInnes, MDF, and Cohen, JF. PRISMA-DTA for abstracts: a new addition to the toolbox for test accuracy research. Diagnostic Progn Res. (2021) 5:1–5. doi: 10.1186/S41512-021-00097-4
27. Deeks, J, Bossuyt, P, Leeflang, M, and Takwoingi, Y. Cochrane handbook for systematic reviews of diagnostic test accuracy. 2nd ed. London, UK: Cochrane (2022).
28. Whiting, PF, Rutjes, AW, Westwood, ME, Mallett, S, Deeks, JJ, Reitsma, JB, et al. QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Ann Intern Med. (2011) 155:529–36. doi: 10.7326/0003-4819-155-8-201110180-00009
29. National Institute for Health Research. MetaDTA: diagnostic test accuracy meta-analysis v2.01. Complex Rev Support Unit (2021) Available at: https://crsu.shinyapps.io/dta_ma/ (Accessed April 17, 2023)
30. Gutiérrez, F, Aluja, A, Rodríguez, C, Gárriz, M, Peri, JM, Gallart, S, et al. Severity in the ICD-11 personality disorder model: evaluation in a Spanish mixed sample. Front Psych. (2023) 13:1015489. doi: 10.3389/fpsyt.2022.1015489
31. Gutiérrez, F, Aluja, A, Ruiz, J, García, LF, Gárriz, M, Gutiérrez-Zotes, A, et al. Personality disorders in the ICD-11: Spanish validation of the PiCD and the SASPD in a mixed community and clinical sample. Assessment. (2021) 28:759–72. doi: 10.1177/1073191120936357
32. Kerr, S, McLaren, V, Cano, K, Vanwoerden, S, Goth, K, and Sharp, C. Levels of personality functioning questionnaire 12–18 (LoPF-Q 12–18): factor structure, validity, and clinical cut-offs. Assessment. (2022):10731911221124340. doi: 10.1177/10731911221124340
33. Cosgun, S, Goth, K, and Cakiroglu, S. Levels of personality functioning questionnaire (LoPF-Q) 12–18 Turkish version: reliability, validity, Factor structure and relationship with comorbid psychopathology in a Turkish adolescent sample. J Psychopathol Behav Assess. (2021) 43:620–31. doi: 10.1007/s10862-021-09867-2
34. Gamache, D, Savard, C, Leclerc, P, Payant, M, Berthelot, N, Côté, A, et al. A proposed classification of ICD-11 severity degrees of personality pathology using the self and interpersonal functioning scale. Front Psych. (2021) 12:292. doi: 10.3389/fpsyt.2021.628057
35. Hemmati, A, Morey, LC, McCredie, MN, Rezaei, F, Nazari, A, and Rahmani, F. Validation of the Persian translation of the level of personality functioning scale—self-report (LPFS-SR): comparison of college students and patients with personality disorders. J Psychopathol Behav Assess. (2020) 42:546–59. doi: 10.1007/s10862-019-09775-6
36. Morey, LC, Bender, DS, and Skodol, AE. Validating the proposed diagnostic and statistical manual of mental disorders, 5th edition, severity indicator for personality disorder. J Nerv Ment Dis. (2013) 201:729–35. doi: 10.1097/NMD.0b013e3182a20ea8
37. Morey, LC, and Skodol, AE. Convergence between DSM-IV-TR and DSM-5 diagnostic models for personality disorder: evaluation of strategies for establishing diagnostic thresholds. J Psychiatr Pract. (2013) 19:179–93. doi: 10.1097/01.pra.0000430502.78833.06
38. Christensen, TB, Hummelen, B, Paap, MCS, Eikenaes, I, Selvik, SG, Kvarstein, E, et al. Evaluation of diagnostic thresholds for criterion a in the alternative DSM-5 model for personality disorders. J Personal Disord. (2020) 34:40–61. doi: 10.1521/PEDI_2019_33_455
39. Roever, L. Types of bias in studies of diagnostic test accuracy. Evid Med Pract. (2016) 2:1000e113. doi: 10.4172/2471-9919.1000e113
40. Kohn, MA, Carpenter, CR, and Newman, TB. Understanding the direction of bias in studies of diagnostic test accuracy. Acad Emerg Med. (2013) 20:1194–206. doi: 10.1111/acem.12255
41. Schmidt, RL, and Factor, RE. Understanding sources of bias in diagnostic accuracy studies. Arch Pathol Lab Med. (2013) 137:558–65. doi: 10.5858/ARPA.2012-0198-RA
42. Hall, MK, Kea, B, and Wang, R. Recognising bias in studies of diagnostic tests part 1: patient selection. Emerg Med J. (2019) 36:431–4. doi: 10.1136/emermed-2019-208446
43. Whiting, PF, Rutjes, AWS, Westwood, ME, and Mallett, S. A systematic review classifies sources of bias and variation in diagnostic test accuracy studies. J Clin Epidemiol. (2013) 66:1093–104. doi: 10.1016/j.jclinepi.2013.05.014
44. Bach, B, Somma, A, and Keeley, JW. Editorial: entering the brave New World of ICD-11 personality disorder diagnosis. Front Psych. (2021) 12:2030. doi: 10.3389/FPSYT.2021.793133
45. Ayinde, OO, and Gureje, O. Cross-cultural applicability of ICD-11 and DSM-5 personality disorder. Curr Opin Psychiatry. (2021) 34:70–5. doi: 10.1097/YCO.0000000000000659
46. Tracy, M, Tiliopoulos, N, Sharpe, L, and Bach, B. The clinical utility of the ICD-11 classification of personality disorders and related traits: a preliminary scoping review. Aust&New Zeal. J Psychiatry. (2021) 55:849–62. doi: 10.1177/00048674211025607
47. Bach, B, and Simonsen, S. How does level of personality functioning inform clinical management and treatment? Implications for ICD-11 classification of personality disorder severity. Curr Opin Psychiatry. (2021) 34:54–63. doi: 10.1097/YCO.0000000000000658
48. Bach, B, Kramer, U, Doering, S, di Giacomo, E, Hutsebaut, J, Kaera, A, et al. The ICD-11 classification of personality disorders: a European perspective on challenges and opportunities. Borderline Personal Disord Emot Dysregulation. (2022) 9:12. doi: 10.1186/S40479-022-00182-0
49. Zimmermann, J, Kerber, A, Rek, K, Hopwood, CJ, and Krueger, RF. A brief but comprehensive review of research on the alternative DSM-5 model for personality disorders. Curr Psychiatry Rep. (2019) 21:92. doi: 10.1007/s11920-019-1079-z
Keywords: ICD-11, DSM-5, personality disorder, dimensional models, severity, diagnostic test accuracy
Citation: Hualparuca-Olivera L and Caycho-Rodríguez T (2023) Diagnostic accuracy of severity measures of ICD-11 and DSM-5 personality disorder: clarifying the clinical landscape with the most up-to-date evidence. Front. Psychiatry 14:1209679. doi: 10.3389/fpsyt.2023.1209679
Edited by:
Massimiliano Beghi, Azienda Unità Sanitaria Locale (AUSL) della Romagna, ItalyReviewed by:
Ilaria Casolaro, ASST Ovest Milanese, ItalyCopyright © 2023 Hualparuca-Olivera and Caycho-Rodríguez. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Luis Hualparuca-Olivera, lhualparuca@continental.edu.pe
†These authors have contributed equally to this work