Skip to main content

SYSTEMATIC REVIEW article

Front. Aging Neurosci., 12 April 2023
Sec. Neurocognitive Aging and Behavior

Prediction model for cognitive frailty in older adults: A systematic review and critical appraisal

\r\nJundan Huang&#x;Jundan Huang1†Xianmei Zeng&#x;Xianmei Zeng1†Mingyue HuMingyue Hu1Hongting NingHongting Ning1Shuang WuShuang Wu1Ruotong PengRuotong Peng1Hui Feng,,*Hui Feng1,2,3*
  • 1Xiangya School of Nursing, Central South University, Changsha, China
  • 2Oceanwide Health Management Institute, Central South University, Changsha, China
  • 3National Clinical Research Center for Geriatric Disorders, Xiangya Hospital, Central South University, Changsha, China

Background: Several prediction models for cognitive frailty (CF) in older adults have been developed. However, the existing models have varied in predictors and performances, and the methodological quality still needs to be determined.

Objectives: We aimed to summarize and critically appraise the reported multivariable prediction models in older adults with CF.

Methods: PubMed, Embase, Cochrane Library, Web of Science, Scopus, PsycINFO, CINAHL, China National Knowledge Infrastructure, and Wanfang Databases were searched from the inception to March 1, 2022. Included models were descriptively summarized and critically appraised by the Prediction Model Risk of Bias Assessment Tool (PROBAST).

Results: A total of 1,535 articles were screened, of which seven were included in the review, describing the development of eight models. Most models were developed in China (n = 4, 50.0%). The most common predictors were age (n = 8, 100%) and depression (n = 4, 50.0%). Seven models reported discrimination by the C-index or area under the receiver operating curve (AUC) ranging from 0.71 to 0.97, and four models reported the calibration using the Hosmer–Lemeshow test and calibration plot. All models were rated as high risk of bias. Two models were validated externally.

Conclusion: There are a few prediction models for CF. As a result of methodological shortcomings, incomplete presentation, and lack of external validation, the models’ usefulness still needs to be determined. In the future, models with better prediction performance and methodological quality should be developed and validated externally.

Systematic review registration: www.crd.york.ac.uk/prospero, identifier CRD42022323591.

Introduction

According to the World Health Organization, by 2030, older adults will increase to 1.4 billion (World Health Organization [WHO], 2021). Aging contributes to many chronic conditions, such as frailty and cognitive impairment (Robertson et al., 2013). In 2013, the International Consensus Group from the International Academy of Nutrition and Aging (IANA) and the International Association of Gerontology and Geriatrics (IAGG) reached a consensus on the concept of cognitive frailty (CF). CF is the presence of physical frailty and cognitive impairment (clinical dementia score of 0.5) with the absence of dementia (Kelaiditi et al., 2013). Due to the difference in assessment tools and criteria, the prevalence of CF ranged from 1.2 to 22.0% in community-dwelling older adults (Arai et al., 2018). One study focused on community-dwelling older adults indicated that the pooled prevalence of CF was 9.0% (Qiu et al., 2022). CF may contribute to a range of adverse outcomes (Feng et al., 2017; Zhang et al., 2021), for instance, dementia, disability, low quality of life, depression, and mortality.

CF includes two subtypes: reversible and potentially reversible CF, depending on the degree of cognitive impairment. The cognitive impairment of potentially reversible CF is mild cognitive impairment, while reversible CF is usually subjective cognitive decline. Studies have shown that if an early intervention was implemented, the CF could be reversed (Ruan et al., 2015; Merchant et al., 2021). However, there are no unified diagnostic tools for CF. Most previous studies have been conducted by trained clinicians and used comprehensive but time-consuming tools to assess cognitive function (Sugimoto et al., 2018). It is recommended to develop prediction models to identify high-risk individuals (Chen et al., 2022).

There are some prediction models on CF. Yang and Zhang (2021) developed a prediction model for CF in community-dwelling older adults with chronic diseases. Predictors included age, living alone, physical exercise, nutrition, and depression. The C-index of the model was as high as 0.97. However, this model did not validate externally. Tseng et al. (2019) also developed a prediction model for CF among community-dwelling older adults. Predictors included age, gender, waist circumference, calf circumference, memory deficits, and diabetes mellitus. However, the definition of CF varied in the development and validation datasets, which could cause the risk of bias (ROB). The existing models varied in predictors and prediction performances, and the methodological quality of these models still needs to be determined.

Therefore, this systematic review aims to summarize the prediction models for CF, assess their methodological quality, and provide some insight for future research.

Methods

This systematic review was designed according to the Preferred Reporting of Items in Systematic Reviews and Meta-Analyzes (PRISMA) guidance (Page et al., 2021). A protocol for this study had been registered in the PROSPERO (CRD42022323591).

Eligibility criteria

Inclusion criteria: (1) Studies’ participants were older adults (60 years older and over); (2) studies reported developing or validating at least one multivariable model for predicting reversible or potentially reversible CF or both; (3) studies had the outcome of interest as CF; (4) studies published in English or Chinese; and (5) Studies were conducted in institutionalized, hospitalized, or community-dwelling settings.

Exclusion criteria: (1) studies were reviews or case reports; (2) studies provided insufficient data for pooling despite several attempts to contact authors; and (3) studies were univariate prediction models.

Literature search

Through preliminary literature search and expert consultation, we systematically searched PubMed, Embase, Cochrane Library, Web of Science, Scopus, PsycINFO, CINAHL, China National Knowledge Infrastructure (CNKI), and Wanfang Databases from the inception to March 1, 2022. The following MeSH terms and free words were used: “aging,” “older adults,” “cognitive frailty,” “cognitive impairments,” “frailty,” “clinical decision rules,” “prediction model,” “risk model,” and so on. The detailed search strategy and results are provided in the Supplementary appendix. Additionally, we manually searched citations for potential studies.

Literature selection

Two reviewers (JH and XZ) independently assessed the titles and abstracts of studies for inclusion. Both authors compared selected studies for inclusion. Discrepancies were resolved through discussion. If there are still discrepancies, a third senior researcher (MH) will join in to discuss and make a consensus. After this initial screening, the full-text articles were retrieved for all records. Full-text articles were also screened independently by two reviewers (JH and XZ).

Data extraction

One reviewer (JH) independently extracted the data, while another (XZ) checked the extracted data. The two reviewers resolved discrepancies through discussion. If there are still discrepancies, a third senior researcher (MH) will join in to discuss and make a consensus.

Data were extracted using a form based on the Checklist for Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies (CHARMS) (Moons et al., 2014). Extracted information on each study included: (1) basic information (e.g., first author, year of publication); (2) outcome measurement methods; (3) source of data; (4) participants (age, gender, country); (5) outcomes to be predicted; (6) predictors; (7) event per variable (EPV); (8) missing data and handling methods; (9) modeling methods; (10) model performances (discrimination, calibration, and classification); (11) model evaluation, and (12) model presentation. An EPV was calculated as overfitting for model development. Studies for model development with EPVs less than 10, while for model validation with participants less than 100, are likely to be overfitting (Wolff et al., 2019).

Model performances were assessed by discrimination, calibration, and classification. Discrimination, measured by the C-index and area under the receiver operating characteristic curve (AUC), is the extent to distinguish those at higher or lower risk of having an event. Calibration refers to the degree of agreement between predicted and observed risks, measured by calibration plot, calibration slope, and Hosmer–Lemeshow test (Alba et al., 2017). The AUC or C-index of 0.7–0.8 is acceptable, 0.8–0.9 is excellent, and more than 0.9 is outstanding (Mandrekar, 2010).

Model validation includes internal and external validation. Internal validation aims to test the reproducibility of the model based on the development cohort data. While external validation is to evaluate the model prediction performance in the new data, focusing on model transportability and generalizability (Royston and Altman, 2013).

When a study included multiple models for the same population and outcome, the model with the best performance was extracted for data extraction. When a study developed multiple models, separate data extraction was performed for each model.

Risk of bias and application assessment

Two reviews (JH and XZ) independently assessed the ROB and application of included studies using the Prediction Model Risk of Bias Assessment Tool (PROBAST) (Moons et al., 2019; Wolff et al., 2019). Discrepancies were resolved through discussion. If there are still discrepancies, a third senior researcher (MH) will join in to discuss and make a consensus. PROBAST consists of 4 domains and 20 signaling questions for ROB assessment. Each signaling question is answered with YES (Y), PROBABLY YES (PY), NO INFORMATION (NI), PROBABLY NO (PN), or NO (N). In a domain, this area can be rated as a low risk only when the answers to all questions are Y or PY. A domain containing at least one question rated as N or PN will be considered high risk. Otherwise, it will be deemed to be at unclear risk. The overall ROB is graded as low risk when all domains are regarded as low risk, and the overall ROB is considered high risk when at least one of the domains is deemed high risk. Otherwise, it will be deemed to be at unclear risk. The application assessment is similar to the ROB assessment but includes only three domains.

Statistical analysis

Due to the differences in the definition criteria of CF, types of predictors, modeling methods, and characteristics of participants, we just calculated and reported descriptive statistics to summarize the characteristics of the models without any quantitative synthesis.

Results

Study selection

This review searched 1535 records, and only seven studies (Tseng et al., 2019; Navarro-Pardo et al., 2020; Rivan et al., 2020; Sargent et al., 2020; Wen et al., 2021; Yang and Zhang, 2021; Chen et al., 2022) were eligible (Figure 1), including eight prediction models. Among these models, seven were diagnostic, and only one was prognostic.

FIGURE 1
www.frontiersin.org

Figure 1. PRISMA 2020 flow diagram for literature screening and selection.

Study designs and study populations

Included study characteristics are summarized in Tables 1, 2. Most of the models were developed using data from cross-sectional studies (n = 7, 87.5%), and most of the participants were from China (n = 4, 50.0%) (Tseng et al., 2019; Wen et al., 2021; Yang and Zhang, 2021; Chen et al., 2022), while the others were from Spain (Navarro-Pardo et al., 2020) (n = 1, 12.5%), Malaysia (Rivan et al., 2020) (n = 1, 12.5%) and Italy (Sargent et al., 2020) (n = 1, 12.5%).

TABLE 1
www.frontiersin.org

Table 1. Characteristics of included studies.

TABLE 2
www.frontiersin.org

Table 2. Validation methods and predictive performance of included studies.

All models were developed for older adults aged 60–91 years. Furthermore, there was no specific gender.

Outcome prevalence or incidence

Only one model measured outcomes of potentially reversible and reversible CF (Chen et al., 2022), and the remainings measured potentially reversible CF. The prevalence of potentially reversible and reversible CF was 37.5%, while the prevalence of potentially reversible CF ranged from 11.91 to 36.0%. One study suggested that the incidence of CF was 35.5% (Rivan et al., 2020).

Predictors

All prediction models reported their predictors. The number of predictors ranged from four to seven. Only two (Sargent et al., 2020) had more than 50 predictors because of the inclusion of single nucleotide polymorphisms (SNPs), protein biomarkers, and other covariates. Most of the models included similar predictors, for example, age (n = 8, 100%), depression (n = 4, 50.0%), physical exercise (n = 3, 37.5%), education (n = 4, 50.0%), and chronic disease (n = 3, 37.5%).

Sample size

All models reported the number of participants included in the final prediction model. The number of participants for model development ranged from 282 to 1155, and EPV ranged from 0.87 to 25.11. Among the eight prediction models, one prediction model (Yang and Zhang, 2021) (12.5%) had an EPV of more than 20, two prediction models (Tseng et al., 2019; Wen et al., 2021) (20.0%) had an EPV between 10 and 20, and the remaining five prediction models (Navarro-Pardo et al., 2020; Rivan et al., 2020; Sargent et al., 2020; Chen et al., 2022) (62.5%) had an EPV less than 10.

Modeling method

Logistic regression (n = 6, 75.0%) was the most common modeling method (Tseng et al., 2019; Navarro-Pardo et al., 2020; Rivan et al., 2020; Wen et al., 2021; Yang and Zhang, 2021; Chen et al., 2022) among the included models, while the other two prediction models (Sargent et al., 2020) were developed by machine learning (n = 2, 25.0%).

Model performance

Seven prediction models reported AUC or C-index, ranging from 0.71 to 0.970. Specifically, two models showed outstanding (Yang and Zhang, 2021; Chen et al., 2022), three showed excellent, and one showed acceptable discrimination (Tseng et al., 2019; Rivan et al., 2020; Sargent et al., 2020). Five prediction models reported calibration using the Hosmer–Lemeshow test (Tseng et al., 2019; Navarro-Pardo et al., 2020; Wen et al., 2021; Yang and Zhang, 2021) (n = 4, 50.0%) and the calibration plot and Brier score (Chen et al., 2022) (n = 1, 12.5%), while the remaining three prediction models did not report calibration.

The classification includes sensitivity, specificity, and accuracy. Three models reported the sensitivity ranging from 70.0 to 81.1%, and the specificity ranging from 60 to 89.1%. One model only reported the accuracy of 76.6% (Navarro-Pardo et al., 2020).

Model evaluation

Four prediction models were internally validated using bootstrapping (Tseng et al., 2019; Wen et al., 2021; Yang and Zhang, 2021; Chen et al., 2022) (50.0%), and the other two (Sargent et al., 2020) were validated by the random split method and bootstrapping. Only two (Navarro-Pardo et al., 2020; Rivan et al., 2020) prediction models did not report whether internal validation was performed (25.0%). Notably, only two models were externally validated (25.0%).

Model presentation

Only four models presented the final models using nomogram (Wen et al., 2021; Yang and Zhang, 2021; Chen et al., 2022) (n = 3, 37.5%) and risk score (Tseng et al., 2019) (n = 1, 12.5%), while the remaining four (Navarro-Pardo et al., 2020; Rivan et al., 2020; Sargent et al., 2020) (50.0%) did not report the presentation.

Risk of bias and application assessment

All models were evaluated as high ROB. In the outcome domain, one model (Rivan et al., 2020) was judged as high ROB because predictors were not excluded from the outcome definition. In the analysis domain, all models were at high ROB, mainly due to the small sample size, failure to handle missing data, selecting predictors by univariable analysis, and inappropriate ways to categorize continuous predictors.

In terms of application, two models (Sargent et al., 2020) had a high concern in the predictor domain. They were developed using SNPs and protein biomarkers to understand the underlying biological mechanisms for the relationship between physical frailty and cognitive impairment, so they were probably not suitable for screening CF in older adults.

Details of the ROB and applicability of all studies were presented in Table 3 and Supplementary Table 1.

TABLE 3
www.frontiersin.org

Table 3. ROB and application assessment of cognitive frailty prediction models.

Model comparison

We compared all models to assess their performances and predictors. More details were shown in Table 4.

TABLE 4
www.frontiersin.org

Table 4. Predictors, advantages, and disadvantages of included studies.

Discussion

This systematic review summarized and critically appraised eight prediction models on CF described in seven studies. Overall, most of them showed excellent or outstanding discrimination. However, there was still room for improvement. For example, the calibration and external validation could be conducted better.

It is worth noting that some models only included a few easy-to-obtain predictors with excellent performance (Yang and Zhang, 2021; Chen et al., 2022). Age, education, depression, and chronic diseases were the robust predictors consistent with previous findings (Yuan et al., 2022). Physical exercise was also a robust predictor. Studies have proven that physical exercise improves older adults’ cognitive function and physical status (Li et al., 2022). Liu et al. (2018) conducted a randomized controlled trial and confirmed that physical activity reduced the odds of worsening CF by 21%. Future studies are suggested to pay attention to an exercise intervention in older adults to prevent CF.

Modeling methods in these studies included logistic regression and machine learning. In general, machine learning models have advantages of flexibility, scalability, and better performance than logistic regression models (DeGregory et al., 2018; Ngiam and Khor, 2019), since logistic regression methods need specific data requirements or assumptions (Wu et al., 2021). However, in this study, the performance of machine learning models was worse than logistic regression models. The reason may be that the model’s predictive performance is influenced not only by the modeling approach but also by the methodological quality, which may lead to an inaccurate estimation of the prediction performance (Moons et al., 2019). However, the sample size needs to be much bigger while machine learning is performed (Moons et al., 2019). Therefore, the machine learning approaches are better suited for large sample sizes with numerous variables.

In this study, we found that most models were calibrated using the Hosmer–Lemeshow test. However, the p-value obtained by the test cannot be used to quantify calibration, so this test is not recommended (Steyerberg and Vergouwe, 2014). There is no best method to calibrate, and all methods have advantages and limitations (Huang et al., 2020). A calibration plot is the most commonly used method for visual displaying.

Models were rarely validated externally. Since only internal validation will lead to high prediction performance, external validation is required to quantify the prediction performance in other study populations for portability and generalization (Moons et al., 2012). Therefore, the external validation of the CF prediction model needs to be strengthened in the future.

All models were judged as high ROB, mainly due to the insufficient sample size, failure to handle missing data, selecting predictors by univariable analysis, and inappropriate ways to categorize continuous predictors, which may lead to an inaccurate estimation of prediction performances of the final model (Moons et al., 2019). Handling missing data by direct elimination will reduce the adequate sample size and increase the ROB of the prediction model. Multiple imputations are considered the best way to handle missing data because it allows users to explicitly incorporate the uncertainty about the actual value of imputed variables (Austin et al., 2021). Predictors selection based on univariable analysis is to independently analyze whether this factor has statistical significance on the occurrence of the outcome, usually measured by p-value. It may miss some essential variables and increase ROB because there may be some correlation between predictors and only after adjusting for other predictors to reach statistical significance (Moons et al., 2019). Although transforming continuous into categorical variables helps explain the results, it will lead to the loss of information and reduce the prediction performance (Collins et al., 2016). An EPV of more than 20 is less likely to have overfitting for model development, and a higher EPV is required (often > 200) when using the machine-learning technique (Moons et al., 2019).

We found some limitations in the geographical location of the models. Most models were developed among Asians. The study has suggested that CF in America and Europe are also prevalent, but rare models have been developed in these regions (Qiu et al., 2022). Since race disparities in the prevalence of cognitive impairment or frailty, tailored prediction models or external validation for American and European older adults are significant (Pandit et al., 2020; Wright et al., 2021).

Strengths and limitations

In this study, we conducted a systematic literature search, extracted data in detail, applied PROBAST to evaluate the model, and provided advice for CF models’ future improvement. However, this study also has some limitations. Firstly, the results were not quantitatively synthesized because all models were judged as high ROB, and the assessment criteria of CF, types of predictors, modeling methods, and characteristics of participants varied. Secondly, we only included Chinese and English literature and did not retrieve the gray literature. These may result in an incomplete inclusion of CF prediction models. Lastly, geographical limitations existed because the included models were mainly developed based on Asians.

Recommendations and implications

There are several suggestions for future research. Firstly, more prediction models should be developed based on a large sample size to make the results more accurate. Secondly, future studies should strictly follow the Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD) statement to standardize and transparent the studies of prediction models (Collins et al., 2015). Thirdly, more attention should be paid to improving the prediction performance and methodological quality, such as choosing appropriate predictors selection methods and modeling methods, using the method of multiple imputations to handle missing data, and validating internally and externally. Finally, it is necessary to evaluate the heterogeneity of prediction models for CF in different subpopulations by the individual participant data (IPD) meta-analysis so that prediction models can be customized for different subgroups (Damen et al., 2016).

Conclusion

There are a few prediction models for CF. As a result of methodological shortcomings, incomplete presentation, and lack of external validation, the models’ usefulness still needs to be determined. In the future, models with better prediction performance and methodological quality should be developed and validated externally.

Data availability statement

The original contributions presented in this study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

HF designed the study question. JH completed the literature search and wrote the draft. JH, XZ, and MH selected the articles, extracted the data, and assessed the quality. JH and XZ prepared the tables and figures. XZ, MH, HN, SW, RP, and HF revised the article, and provided critical recommendations on structure and presentation. All authors approved the final version of the manuscript.

Funding

This study was supported by the National Key R&D Program of China (Grant numbers: 2020YFC2008602 and 2020YFC2008603) and the Special Funding for the Construction of Innovative Provinces in Hunan (2020SK2055).

Acknowledgments

We are grateful to the authors of the various studies reviewed.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnagi.2023.1119194/full#supplementary-material

Abbreviations

AUC, area under the receiver operating curve; CF, cognitive frailty; CHARMS, Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies; EPV, event per variable; IAGG, International Association of Gerontology and Geriatrics; IANA, International Academy of Nutrition and Aging; PRISMA, Preferred Reporting of Items in Systematic Reviews and Meta-Analyses; PROBAST, Prediction Model Risk of Bias Assessment Tool; ROB, risk of bias.

References

Alba, A. C., Agoritsas, T., Walsh, M., Hanna, S., Iorio, A., Devereaux, P. J., et al. (2017). Discrimination and calibration of clinical prediction models: Users’ guides to the medical literature. JAMA 318, 1377–1384. doi: 10.1001/jama.2017.12126

PubMed Abstract | CrossRef Full Text | Google Scholar

Arai, H., Satake, S., and Kozaki, K. (2018). Cognitive Frailty in Geriatrics. Clin. Geriatr. Med. 34, 667–675. doi: 10.1016/j.cger.2018.06.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Austin, P. C., White, I. R., Lee, D. S., and van Buuren, S. (2021). Missing data in clinical research: A tutorial on multiple imputation. Can. J. Cardiol. 37, 1322–1331. doi: 10.1016/j.cjca.2020.11.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, Y., Zhang, Z., Zuo, Q., Liang, J., and Gao, Y. (2022). Construction and validation of a prediction model for the risk of cognitive frailty among the elderly in a community (Chinese). Chin. J. Nurs. 57, 197–203. doi: 10.3761/j.issn.0254-1769.2022.02.012

CrossRef Full Text | Google Scholar

Collins, G. S., Ogundimu, E. O., Cook, J. A., Manach, Y. L., and Altman, D. G. (2016). Quantifying the impact of different approaches for handling continuous predictors on the performance of a prognostic model. Stat. Med. 35, 4124–4135. doi: 10.1002/sim.6986

PubMed Abstract | CrossRef Full Text | Google Scholar

Collins, G. S., Reitsma, J. B., Altman, D. G., and Moons, K. G. (2015). Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): The TRIPOD statement. BMJ 350:g7594. doi: 10.1136/bmj.g7594

PubMed Abstract | CrossRef Full Text | Google Scholar

Damen, J. A., Hooft, L., Schuit, E., Debray, T. P., Collins, G. S., Tzoulaki, I., et al. (2016). Prediction models for cardiovascular disease risk in the general population: Systematic review. BMJ 353:i2416. doi: 10.1136/bmj.i2416

PubMed Abstract | CrossRef Full Text | Google Scholar

DeGregory, K. W., Kuiper, P., DeSilvio, T., Pleuss, J. D., Miller, R., Roginski, J. W., et al. (2018). A review of machine learning in obesity. Obes. Rev. 19, 668–685. doi: 10.1111/obr.12667

PubMed Abstract | CrossRef Full Text | Google Scholar

Feng, L., Zin Nyunt, M. S., Gao, Q., Feng, L., Yap, K. B., and Ng, T. P. (2017). Cognitive frailty and adverse health outcomes: Findings from the singapore longitudinal ageing studies (SLAS). J. Am. Med. Dir. Assoc. 18, 252–258. doi: 10.1016/j.jamda.2016.09.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Huang, Y., Li, W., Macheret, F., Gabriel, R. A., and Ohno-Machado, L. (2020). A tutorial on calibration measurements and calibration models for clinical prediction models. J. Am. Med. Inform. Assoc. 27, 621–633. doi: 10.1093/jamia/ocz228

PubMed Abstract | CrossRef Full Text | Google Scholar

Kelaiditi, E., Cesari, M., Canevelli, M., van Kan, G. A., Ousset, P. J., Gillette-Guyonnet, S., et al. (2013). Cognitive frailty: Rational and definition from an (I.A.N.A./I.A.G.G.) international consensus group. J. Nutr. Health Aging 17, 726–734. doi: 10.1007/s12603-013-0367-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, X., Zhang, Y., Tian, Y., Cheng, Q., Gao, Y., and Gao, M. (2022). Exercise interventions for older people with cognitive frailty-a scoping review. BMC Geriatr. 22:721. doi: 10.1186/s12877-022-03370-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, Z., Hsu, F. C., Trombetti, A., King, A. C., Liu, C. K., Manini, T. M., et al. (2018). Effect of 24-month physical activity on cognitive frailty and the role of inflammation: The LIFE randomized clinical trial. BMC Med. 16:185. doi: 10.1186/s12916-018-1174-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Mandrekar, J. N. (2010). Receiver operating characteristic curve in diagnostic test assessment. J. Thorac. Oncol. 5, 1315–1316. doi: 10.1097/JTO.0b013e3181ec173d

PubMed Abstract | CrossRef Full Text | Google Scholar

Merchant, R. A., Chan, Y. H., Hui, R. J. Y., Tsoi, C. T., Kwek, S. C., Tan, W. M., et al. (2021). Motoric cognitive risk syndrome, physio-cognitive decline syndrome, cognitive frailty and reversibility with dual-task exercise. Exp. Gerontol. 150:111362. doi: 10.1016/j.exger.2021.111362

PubMed Abstract | CrossRef Full Text | Google Scholar

Moons, K. G. M., Wolff, R. F., Riley, R. D., Whiting, P. F., Westwood, M., Collins, G. S., et al. (2019). PROBAST: A tool to assess risk of bias and applicability of prediction model studies: Explanation and elaboration. Ann. Intern. Med. 170, W1–W33. doi: 10.7326/m18-1377

PubMed Abstract | CrossRef Full Text | Google Scholar

Moons, K. G., de Groot, J. A., Bouwmeester, W., Vergouwe, Y., Mallett, S., Altman, D. G., et al. (2014). Critical appraisal and data extraction for systematic reviews of prediction modelling studies: The CHARMS checklist. PLoS Med. 11:e1001744. doi: 10.1371/journal.pmed.1001744

PubMed Abstract | CrossRef Full Text | Google Scholar

Moons, K. G., Kengne, A. P., Grobbee, D. E., Royston, P., Vergouwe, Y., Altman, D. G., et al. (2012). Risk prediction models: II. External validation, model updating, and impact assessment. Heart 98, 691–698. doi: 10.1136/heartjnl-2011-301247

PubMed Abstract | CrossRef Full Text | Google Scholar

Navarro-Pardo, E., Facal, D., Campos-Magdaleno, M., Pereiro, A. X., and Juncos-Rabadán, O. (2020). Prevalence of cognitive frailty, do psychosocial-related factors matter? Brain Sci. 10:968. doi: 10.3390/brainsci10120968

PubMed Abstract | CrossRef Full Text | Google Scholar

Ngiam, K. Y., and Khor, I. W. (2019). Big data and machine learning algorithms for health-care delivery. Lancet Oncol. 20, e262–e273. doi: 10.1016/s1470-2045(19)30149-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Page, M. J., McKenzie, J. E., Bossuyt, P. M., Boutron, I., Hoffmann, T. C., Mulrow, C. D., et al. (2021). The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 372:n71. doi: 10.1136/bmj.n71

PubMed Abstract | CrossRef Full Text | Google Scholar

Pandit, V., Nelson, P., Kempe, K., Gage, K., Zeeshan, M., Kim, H., et al. (2020). Racial and ethnic disparities in lower extremity amputation: Assessing the role of frailty in older adults. Surgery 168, 1075–1078. doi: 10.1016/j.surg.2020.07.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Qiu, Y., Li, G., Wang, X., Zheng, L., Wang, C., Wang, C., et al. (2022). Prevalence of cognitive frailty among community-dwelling older adults: A systematic review and meta-analysis. Int. J. Nurs. Stud. 125:104112. doi: 10.1016/j.ijnurstu.2021.104112

PubMed Abstract | CrossRef Full Text | Google Scholar

Rivan, N. F. M., Shahar, S., Rajab, N. F., Singh, D. K. A., Che Din, N., Mahadzir, H., et al. (2020). Incidence and predictors of cognitive frailty among older adults: A community-based longitudinal study. Int. J. Environ. Res. Public Health 17:1547. doi: 10.3390/ijerph17051547

PubMed Abstract | CrossRef Full Text | Google Scholar

Robertson, D. A., Savva, G. M., and Kenny, R. A. (2013). Frailty and cognitive impairment–a review of the evidence and causal mechanisms. Ageing Res. Rev. 12, 840–851. doi: 10.1016/j.arr.2013.06.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Royston, P., and Altman, D. G. (2013). External validation of a Cox prognostic model: Principles and methods. BMC Med. Res. Methodol. 13:33. doi: 10.1186/1471-2288-13-33

PubMed Abstract | CrossRef Full Text | Google Scholar

Ruan, Q., Yu, Z., Chen, M., Bao, Z., Li, J., and He, W. (2015). Cognitive frailty, a novel target for the prevention of elderly dependency. Ageing Res. Rev. 20, 1–10. doi: 10.1016/j.arr.2014.12.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Sargent, L., Nalls, M., Amella, E. J., Slattum, P. W., Mueller, M., Bandinelli, S., et al. (2020). Shared mechanisms for cognitive impairment and physical frailty: A model for complex systems. Alzheimers Dement. 6:e12027. doi: 10.1002/trc2.12027

PubMed Abstract | CrossRef Full Text | Google Scholar

Steyerberg, E. W., and Vergouwe, Y. (2014). Towards better clinical prediction models: Seven steps for development and an ABCD for validation. Eur. Heart J. 35, 1925–1931. doi: 10.1093/eurheartj/ehu207

PubMed Abstract | CrossRef Full Text | Google Scholar

Sugimoto, T., Sakurai, T., Ono, R., Kimura, A., Saji, N., Niida, S., et al. (2018). Epidemiological and clinical significance of cognitive frailty: A mini review. Ageing Res. Rev. 44, 1–7. doi: 10.1016/j.arr.2018.03.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Tseng, S. H., Liu, L. K., Peng, L. N., Wang, P. N., Loh, C. H., and Chen, L. K. (2019). Development and validation of a tool to screen for cognitive frailty among community-dwelling elders. J. Nutr. Health Aging 23, 904–909. doi: 10.1007/s12603-019-1235-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Wen, F., Chen, M., Zhao, C., Guo, X., and Xi, X. (2021). Development of a cognitive frailty prediction model for elderly patients with stable coronary artery disease (Chinese). J. Nurs. Sci. 36, 21–26. doi: 10.3870/j.issn.1001-4152.2021.10.021

CrossRef Full Text | Google Scholar

Wolff, R. F., Moons, K. G. M., Riley, R. D., Whiting, P. F., Westwood, M., Collins, G. S., et al. (2019). PROBAST: A tool to assess the risk of bias and applicability of prediction model studies. Ann. Intern. Med. 170, 51–58. doi: 10.7326/m18-1376

PubMed Abstract | CrossRef Full Text | Google Scholar

World Health Organization [WHO] (2021). Ageing and health [Online]. Geneva: World Health Organization.

Google Scholar

Wright, C. B., DeRosa, J. T., Moon, M. P., Strobino, K., DeCarli, C., Cheung, Y. K., et al. (2021). Race/ethnic disparities in mild cognitive impairment and dementia: The northern manhattan study. J. Alzheimers Dis. 80, 1129–1138. doi: 10.3233/jad-201370

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, W. T., Li, Y. J., Feng, A. Z., Li, L., Huang, T., Xu, A. D., et al. (2021). Data mining in clinical big data: The frequently used databases, steps, and methodological models. Mil. Med. Res. 8:44. doi: 10.1186/s40779-021-00338-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, Z., and Zhang, H. (2021). A nomogram for predicting the risk of cognitive frailty in community-dwelling elderly people with chronic diseases (Chinese). J. Nurs. Sci. 36, 86–89. doi: 10.3870/j.issn.1001-4125.2021.12.086

CrossRef Full Text | Google Scholar

Yu, R., Wong, M., Chong, K. C., Chang, B., Lum, C. M., Auyeung, T. W., et al. (2018). Trajectories of frailty among Chinese older people in Hong Kong between 2001 and 2012: An age-period-cohort analysis. Age Ageing 47, 254–261. doi: 10.1093/ageing/afx170

PubMed Abstract | CrossRef Full Text | Google Scholar

Yuan, M., Xu, C., and Fang, Y. (2022). The transitions and predictors of cognitive frailty with multi-state Markov model: A cohort study. BMC Geriatr. 22:550. doi: 10.1186/s12877-022-03220-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, T., Ren, Y., Shen, P., Jiang, S., Yang, Y., Wang, Y., et al. (2021). Prevalence and associated risk factors of cognitive frailty: A systematic review and meta-analysis. Front. Aging Neurosci. 13:755926. doi: 10.3389/fnagi.2021.755926

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: cognitive frailty, prediction models, older adults, systematic review, critical appraisal

Citation: Huang J, Zeng X, Hu M, Ning H, Wu S, Peng R and Feng H (2023) Prediction model for cognitive frailty in older adults: A systematic review and critical appraisal. Front. Aging Neurosci. 15:1119194. doi: 10.3389/fnagi.2023.1119194

Received: 08 December 2022; Accepted: 28 March 2023;
Published: 12 April 2023.

Edited by:

Stephanie Marie-Rose Duguez, Ulster University, United Kingdom

Reviewed by:

Lucy Beishon, University of Leicester, United Kingdom
Pedro Miguel Gaspar, University Institute of Maia (ISMAI), Portugal

Copyright © 2023 Huang, Zeng, Hu, Ning, Wu, Peng and Feng. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Hui Feng, feng.hui@csu.edu.cn

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.