- 1The Department of Geriatric Psychiatry, Shanghai Mental Health Center, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- 2School of Computer and Information Engineering and Henan Engineering Research Center of Intelligent Technology and Application, Henan University, Kaifeng, China
Alzheimer's disease (AD) is the most common cause of dementia. Accurate prediction and diagnosis of AD and its prodromal stage, i.e., mild cognitive impairment (MCI), is essential for the possible delay and early treatment for the disease. In this paper, we adopt the data from the China Longitudinal Aging Study (CLAS), which was launched in 2011, and includes a joint effort of 15 institutions all over the country. Four thousand four hundred and eleven people who are at least 60 years old participated in the project, where 3,514 people completed the baseline survey. The survey collected data including demographic information, daily lifestyle, medical history, and routine physical examination. In particular, we employ ensemble learning and feature selection methods to develop an explainable prediction model for AD and MCI. Five feature selection methods and nine machine learning classifiers are applied for comparison to find the most dominant features on AD/MCI prediction. The resulting model achieves accuracy of 89.2%, sensitivity of 87.7%, and specificity of 90.7% for MCI prediction, and accuracy of 99.2%, sensitivity of 99.7%, and specificity of 98.7% for AD prediction. We further utilize the SHapley Additive exPlanations (SHAP) algorithm to visualize the specific contribution of each feature to AD/MCI prediction at both global and individual levels. Consequently, our model not only provides the prediction outcome, but also helps to understand the relationship between lifestyle/physical disease history and cognitive function, and enables clinicians to make appropriate recommendations for the elderly. Therefore, our approach provides a new perspective for the design of a computer-aided diagnosis system for AD and MCI, and has potential high clinical application value.
1. Introduction
Alzheimer's disease (AD) is the most common dementia in the elderly, which is a slow and lengthy progressive neurodegenerative disorder and accounts for 60–80% of dementia cases. The population of AD is projected to reach 106.8 million by 2050 (Brookmeyer et al., 2007). Although numerous therapies have been investigated, there has been no successful trial that can modify the course of the disease. On the other hand, according to the data from epidemiologic studies and clinical trials, it has indicated that early intervention may delay the AD progression (Brookmeyer et al., 1998; Norton et al., 2014; Ngandu et al., 2015). The prodromal stage of AD, termed as mild cognitive impairment (MCI), can lead to cognitive decline and has a high risk to develop AD. Thus, accurate prediction and diagnosis of AD and MCI are very critical for the prevention and therapy of the disease.
Previous studies reported that some clinical and demographic features are considered to have strong predictive abilities (Livingston et al., 2017). However, none of them is strong enough to differentiate AD/MCI among the community elderly independently. It is more likely that clinical/demographic features may have complex relationships and, as a whole, jointly predict AD progression. Hence, the artificial intelligence (AI) approach may be a suitable way to combine these data to solve the problem.
Recently, many researchers have applied AI techniques for AD prediction. Zhang et al. (2019) propose a deep learning approach based on two convolutional neural networks (CNN) and multimodal medical images. Then correlation analysis is applied to judge the consistency of the output of the two CNN. Salvatore et al. (2015) extract features from MRI data using principal component analysis, and apply a machine learning algorithm to predict whether MCI patients will convert to AD.
Loddo et al. (2022) presents a deep learning approach for Alzheimer's disease diagnosis using brain images. It compares different deep learning models and proposes a fully automated deep-ensemble approach for dementia-level classification. Discusses the challenges in detecting Alzheimer's disease (AD) in its early stages and reviews the current research on machine learning techniques for its detection and classification, with a focus on neuroimaging. The review suggests that deep learning techniques hold promise for AD diagnosis, and new algorithms have yet to be explored for AD diagnosis. These studies above apply various machine learning and deep learning methods to predict AD based on data from different modalities. But they only focus on the models' performance while neglecting the interpretation of the output of these models.
The following studies not only design a new model, but also analyze the output of the model. El-Sappagh et al. (2021) develop a two-layer model with random forest (RF), and use the SHapley Additive exPlanations (SHAP) framework to make overall and individual explanations for the result of each layer. Additionally, 22 explainers are developed based on decision tree and fuzzy rule-based systems to provide supplementary justifications for every RF decision in each layer. Danso et al. (2021) develop a framework that integrates transfer learning and ensemble learning algorithms to develop explainable personalized risk prediction models for dementia, SHAP is used to visualize the risk factors responsible for the prediction.
In this paper, we adopt the data from China Longitudinal Aging Study (CLAS; Haibo et al., 2013; Xiao, 2013; Xiao et al., 2016), which is a community-based cohort study launched in 2011. The project was conducted jointly by 15 institutions located in eastern, middle, and western parts of China. A total of 4,411 people at least 60 years old participated in the project, where 3,514 people completed the baseline survey. The survey collected data including demographic information, daily lifestyle, medical history, and routine physical examination. In addition, a variety of psychological and psychosocial measures were assessed by psychologists. A normal diagnostic method was adopted to classify the cognitive condition of all subjects, i.e., normal control (NC), MCI, or AD.
Based on this data, we aim to propose a joint detection of interpretable machine learning models and predictive indicators for predicting AD and MCI as follows:
1) we have processed the missing and default values in the data through a unified arrangement and data cleaning steps as part of data preprocessing. At the same time, we compared five feature selection methods to reduce the dimensionality of the data and the complexity of the model calculation. Finally, we used nine classifiers of general interpretable machine learning for classification comparison.
2) based on previous research, our dataset includes more comprehensive information, including lifestyle, physical diseases, and medical check-up results. To our knowledge, this is the first work that aims to predict cognitive status using large-scale and multi-faceted information, especially detailed lifestyle and clinical information.
2. Materials and methods
This section describes the details of our proposed system. As shown in Figure 1, it has four stages, which are data preprocessing, feature engineering, classification, and explanation, respectively. The framework also displays the methods adopted in these stages. The detailed introduction of every stage can be found in the following subsections.
2.1. Study participants and data collection
The population of our study is a community-based cohort study, named CLAS. The Chinese Longitudinal Aging Study (CLAS) was designed to provide information about the cognitive, mental, and psychosocial health of older people in China (Xiao, 2013). This survey was a joint effort of 15 institutions located in the eastern, middle, and western parts of China. The sample was randomly selected from all permanent residents aged over 60 in the 2010 national census (Xiao, 2013).
As reported in the above protocol (Haibo et al., 2013; Xiao, 2013; Xiao et al., 2016). These clinical diagnoses were made according to accepted criteria and with consideration of comorbid conditions. MCI was classified using the Petersen criteria (Petersen et al., 2001) and AD dementia were diagnosed according to the DSM-IV criteria (American Psychiatric Association, 2000), both of which were clinically diagnosis.
Out of the 3,514 participants who completed the survey, a total of 2,658 people had cognitive condition results, which includes 98 individuals (3.69%) with AD, 556 individuals (20.92%) with MCI, and 2,004 individuals with NC.
The dataset has 53 features, including demographic information, daily lifestyle, medical history, and routine physical examination. Tables 1, 2 shows the summary results of the standard deviation, mean, and interquartile range (IQR) of every feature for three classes.
2.2. Data preprocessing
2.2.1. Missing value
In the dataset, most features have missing values, yet the missing rate is low (<7%). For the features with missing values, we first treat a feature with missing value as a new tag, and the remaining features and original tags form new input values. Then we apply a random forest algorithm to predict the missing values in the new tag (Liaw and Wiener, 2002). All features are filled up in turn following the steps above.
2.2.2. Data augmentation
In our dataset, the number of AD (98 samples) and MCI (556 samples) are far less than that of NC (2,004 samples). The data imbalance may seriously degrade the performance of the machine learning algorithm. For example, overfitting may occur due to the imbalanced training data. We use the adaptive synthetic sampling approach (ADASYN) to handle the issue (He et al., 2008). ADASYN can adaptively generate samples for the minority class based on its distribution.
2.2.3. Data normalization
In the dataset, every feature has a different value range. This may lead to unreasonable results, since the feature with larger values will have higher weights on the learned model. Thus, it is necessary to use data normalization to mitigate this effect. Max-min normalization is applied to each feature, which can be expressed by
where is the standard deviation of X, given by
2.3. Feature selection
There are a total of 53 features in the dataset, including demographics, daily lifestyle, medical history, and routine physical examination as shown in Tables 1, 2. Dimensionality reduction is a fundamental requirement for achieving simplicity and assessing the complexity of the model. The curse of dimensionality can adversely impact the model in terms of runtime and exhaustion of storage resources, particularly for non-scalable classifiers. For these reasons, we need to use the feature selection methods.
Feature selection is the preprocessing step before applying the classifier, which aims to eliminate unrelated and redundant features while preserving the key information of the original dataset by selecting the representative features.
Five feature selection methods are tested in our experiment, which are ReliefF (Kononenko, 1994), Gini index (Gini, 1971), Information gain (IG; Alhaj et al., 2016), Least absolute shrinkage and selection operator (LASSO; Tibshirani, 1996), Sparse group LASSO (SGL; Friedman et al., 2010).
2.4. Machine learning algorithm
Several machine learning models are compared to select the best classifier for AD/MCI prediction, which include three basic classifiers and six ensemble classifiers.
Three basic classifiers are K-Nearest Neighbors (KNN; Altman, 1992), Decision Tree (DT; Loh, 2011), and Support Vector Machine (SVM; Cortes and Vapnik, 1995), respectively.
Ensemble learning is an integrated approach that can combine multiple base learners to achieve better performance, where many machine learning algorithms can be applied as the base learners, such as DT, neural network, etc. The base learners can be generated by two styles, i.e., the parallel style and the sequential style. Then all base learners will be combined to form a better learner, where the most common combination schemes are the majority for voting and weighted averaging for regression. To find out the best classifier, six ensemble learning methods are applied for comparison, which are Adaptive Boosting (AdaBoost; Freund and Schapire, 1996), eXtreme Gradient Boosting (XGBoost; Chen and Guestrin, 2016), Light Gradient Boosting Machine (LightGBM; Ke et al., 2017), Bootstrap Aggregation (Bagging; Breiman, 1996), Random Forest (RF; Breiman, 1996), and Extra Tree (ET; Geurts et al., 2006), respectively.
2.5. Model explainer
SHapley Additive exPlanation (SHAP) is a game-theoretic approach to explain the output of any machine learning model, which is proposed by Lundberg and Lee (2017). The goal of SHAP is to explain the prediction of a sample xi by computing the influence score of each feature to the prediction. The prediction yi can be expressed as follows:
where ybase is the average of predictions of all samples, and f(xij) is the SHAP value of xij which is the contribution of j-th feature to the prediction of xi. When f(xij)>0, the j-th feature can boost the prediction, otherwise, it has a negative effect. Compared with traditional measurement of feature importance, the strength of SHAP is that it can reflect the specific contribution of each feature to the model's output.
2.6. Performance metrics
To evaluate the model's performance, we use five performance criteria : Accuracy, Sensitivity, Specificity, G-mean, and Area Under Curve (AUC). Accuracy is the ratio between the correctly classified samples and all samples, which is defined as
where TP represents the true positive, TN represents the true negative, FP represents the false positive, and FN represents the false negative. Since the data imbalance exists among different classes, accuracy is not enough for evaluating performance. It may cause misleading if the model only predicts the majority class correctly while neglecting the minority class. It is necessary to use more metrics to evaluate the performance of each class. Sensitivity is a measure of how well a model can predict for positive samples, and specificity is a metric of how well a model can predict for negative samples. The definition of the above metrics are shown as follows:
G-mean is a reliable metric in the situation of overfitting the negative class and underfitting the positive class. As shown in Equation (7), it combines the sensitivity and specificity into a single score to balance both concerns. A model has a high G-mean, meaning that a classifier is not biased toward any of the classes (Kotsiantis et al., 2006).
AUC is another helpful metric to evaluate how effective the classifier is in separating different classes. The receiver operating characteristic curve (ROC) plots the Sensitivity against the 1−Specificity at various threshold settings, where the area of 1 indicates the model is excellent, and the area of 0.5 denotes it is a worthless model.
3. Experiments and results
In this section, we commence by revisiting hyperparameter optimization techniques and other pertinent works, followed by a detailed exposition of our specific experimental setup.
3.1. Hyperparameter optimization
In facing a plethora of algorithms mentioned within, each bearing distinct types of hyperparameters, the impact on model performance may manifest in varying manners. Take the Random Forest algorithm as an instance, wherein the number of estimators and the depth of trees serve as hyperparameters, casting a profound influence upon the model's performance. Currently, hyperparameter tuning can predominantly be categorized into the following methods: Grid Search: A classic technique diligently employed by examining all plausible parameter combinations. Grid Search contemplates the entire parameter space, partitioning it into a grid-like structure, where each point within the grid is evaluated as a hyperparameter (Bergstra and Bengio, 2012; Shekhar et al., 2021). This near-exhaustive optimization approach is apt for low-dimensional hyperparameter spaces, albeit our classifier algorithms necessitate multi-dimensional space optimization. Random Search: Randomized selection of hyperparameters marks the hallmark of this method, offering simplicity in implementation, yet challenging the adjustment of model hyperparameters based on commonality (Bergstra and Bengio, 2012). Bayesian Hyper-parameter Optimization: Adopting a Bayesian rules, this method fine-tunes the evaluation function through posterior distribution, markedly reducing the search process within the parameter space (Dewancker et al., 2016). In our experimental setup, we employed the HyperOpt library (Bergstra et al., 2013) for hyperparameter optimization. HyperOpt operates on the paradigm of Sequential Model-Based Optimization (SMBO; Hutter et al., 2011), with the Tree of Parzen Estimators (TPE) orchestrating the search within the current space.
The parameter search space across different classifiers is delineated in Table 3, showcasing the breadth and scope entailed in tuning the hyperparameters to adeptly tailor the model to our dataset.
3.2. System setup and implementation
We developed our framework using the Python 3.6 environment. Essential libraries employed in our study included scipy, matplotlib, pandas, sklearn, Hyperopt, and numpy. The computational experiments were conducted on a laptop equipped with an Intel Core i5-10310U CPU and 16 GB of RAM. The simulations demanded ~10 h to produce the outcomes.
The experiments are carried out based on the K-fold cross-validation technique with K=10. The dataset is divided into K subsets, where each subset is treated as a testing set in turn, while the rest of the data is used to train the model. Then the final result is the average of these K results. This method guarantees the training and testing processes are both applied to the whole dataset. During the generation of each fold, stratified sampling is also applied to ensure that the proportion of samples of each class in the training and testing sets is the same as that in the original dataset, which is important to have more representative samples and reduce sampling errors.
3.3. Performance analysis of all classifiers with oversampling
In order to select the optimal classifier for the classification task, we apply nine classifiers for comparison. As the disparity of the sample number between different classes is too large, we conduct the experiments repeatedly using different oversampling ratios. Every classifier's final results are obtained with the optimal oversampling ratio.
Furthermore, we also compare the results of the classifiers on the original data and the oversampling data.In the experiment, since the number of MCI and NC are 556 and 2,004, respectively, we use ADASYN to oversample for MCI at different ratios, starting from 100 to 300%, and the experiment is repeated three times for each classifier. Table 4 shows the results of all classifiers on original data and oversampling data with 10-fold cross validation. Compared with the results obtained from the original data, the specificity of all classifiers from the oversampling data decrease, which means the prediction ability of all classifiers for NC declines. On the contrary, the sensitivity of all models increases significantly, which indicates that the prediction ability of MCI has been greatly improved by using the oversampling method. Moreover, according to the increment in accuracy, G-mean, and AUC among most of the classifiers, more samples generated by oversampling technique make the whole performance improve. Although SVM achieves the best with respect to the accuracy and the G-mean, its overall performance is not as good as AdaBoost after applying the feature selection methods, which will be discussed in the next subsection.
Table 4. Performance under different classifiers for MCI/NC prediction (OR, Original data; OS, Oversampling data).
Table 5 shows the classification results of AD and NC with 10-fold cross validation. Since the number of AD is only 98, the oversampling ratio is from 100 to 2000%. In the experiment, except for the specificity, the other four metrics of all classifiers have been greatly improved by applying oversampling, especially the sensitivity. The AdaBoost achieves accuracy of 0.996, sensitivity of 0.999, specificity of 0.993, G-mean of 0.996, and AUC of 1. Although the sensitivity of AdaBoost is slightly worse than that of SVM, its other metrics are the best among all classifiers. In summary, we select AdaBoost as the classifier.
Table 5. Performance under different classifiers for AD/NC prediction (OR, Original data; OS, Oversampling data).
3.4. Performance analysis of feature selection methods under AdaBoost and oversampling
The aim of this experiment is to use the feature selection method to decrease the dimension of the dataset and computational complexity. Five feature selection methods are applied in the experiment, which are Gini index, IG, reliefF, LASSO, and SGL. They are used to reduce the dimension from 53 to 15 for the oversampling dataset, which is selected from the prior experiments with an optimal oversampling ratio when AdaBoost achieves the best performance. Then we train the AdaBoost model on these datasets. Tables 6, 7 show the results for different classification tasks with 10-fold cross validation under different feature selection methods. In the classification task of MCI/NC, reliefF achieves the best performance with respect to accuracy, sensitivity, G-mean, and AUC, which are 0.892, 0.877, 0.892, and 0.957, respectively. It also achieves the optimal values for the four metrics in the AD/NC classification task, where accuracy is 0.992, sensitivity is 0.997, G-mean is 0.992 and AUC is 1. Therefore, the reliefF is selected as the final feature selection method.
Table 6. Performance for different feature selection methods for MCI/NC classification using ADASYN and AdaBoost.
Table 7. Performance for different feature selection methods for AD/NC classification using ADASYN and AdaBoost.
We also compare Adaboost with SVM to classify MCI/NC on the dataset processed by reliefF. The results indicate that Adaboost indeed outperforms SVM, as shown in Table 8.
Table 8. The performance comparison between AdaBoost and SVM in MCI/NC classification task using ADASYN and ReliefF.
The runtime of the experiment with feature selection is 111.6 seconds in this experiment, which is much smaller than that without feature selection, 393.3 seconds. Although the performance of classifiers slightly reduces after feature selection, it decreases the computational complexity of the model and screens out the most important features, which lays the foundation for further analysis of these features.
3.5. Model explainability
The learned model is further analyzed by using the SHAP. The 15 features selected by reliefF in the MCI/NC classification task are Education, Sleeping time (elder), Heart rate, Height, Memory decline, Age, Eat fish, Diastolic blood pressure, Sport, Systolic blood pressure, Tea, Hypertension, Frequency of nap (elder), Smoke and Family history of physical disease. Regarding the classification of AD/NC, the corresponding features are Daily life function decline, Age, Sport, Diastolic blood pressure, Education, Frequency of nap (elder), Sleeping time (elder), Eat fish, Frequency of nap (middle-aged), Hypertension, Eating habits, Surgical history, Family history of physical disease and Professional nature. Figure 2 shows the SHAP summary plots for MCI/AD. The y-axis of the plot is the feature value, where the features are sorted by the mean of the absolute value of SHAP values in all samples. The x-axis is the SHAP value, which represents the contribution of the feature to the output. Each dot represents the impact on a particular class of a particular feature for a given sample. The color represents the feature value (red = high, blue = low).
We notice that Education is the most important feature for the MCI class, where high value of Education has a negative impact on predicting MCI class, meaning that Education is a factor that decreases MCI risk. Some features [e.g., Sleeping time (elder), Heart rate, Height, etc.] are globally less critical than Education, but they have more impacts in some cases. For instance, the largest SHAP value for Sleeping time (elder) is 0.0497, which is greater than the maximum SHAP value of Education, 0.0275. Similarly, for AD, the top feature is the Daily life function decline. The feature Age is less critical, however, when its value is very small, it has more negative impact than the Daily life function decline on the model for predicting AD. In addition, there are nine features that are identical for MCI and AD, but the importance of these features in MCI class is not as high as theirs in AD class.
We also analyze the impact of a single feature on the prediction. The Figure 3 shows the SHAP dependency plots for MCI class, where the x-axis represents the value distribution of each feature in all samples, and the y-axis represents the SHAP value. As seen in Figures 3A, C, D, E, G, with the increasing values of these features, the overall trend of the SHAP values is downward, indicating these features have negative effects on predicting MCI class. On the contrary, as the value of Age and Systolic blood pressure increase, their SHAP values also increase as shown in Figures 3B, F.
Figure 3. SHAP dependency plots for MCI class. (A) Education, (B) age, (C) eat fish, (D) sleeping time (elder), (E) tea, (F) systolic blood pressure, (G) diastolic blood pressure.
In the initial assessment targeting the prediction of Mild Cognitive Impairment (MCI) and Alzheimer's Disease (AD), as portrayed in Figures 4A, B respectively, the features manifesting the most substantial influence were “Memory Decline” and “Daily Life Function Decline.” This observation harmonizes with the discernments encountered in clinical diagnostic realms. Aiming for a more meticulous evaluation of our model's stability, we pivoted our attention toward the interplay of daily life features with MCI and AD prognostics. Specifically, we excised crucial features conventionally harnessed for clinical recognition: “Memory Decline,” “Daily Life Function Decline,” and “Disability in Work and Study.” The subsequent SHAP plots, delineated in Figure 5, vindicate that notwithstanding the exclusion, daily life features retain a pivotal role in rendering credible predictive values in our model's framework. This meticulous endeavor not only underscores the robustness of our model but also illumines the nuanced daily life attributes that contribute significantly to the predictive landscape of MCI and AD.
Figure 6 displays the SHAP dependency plots for AD class. We notice that the SHAP values of the Daily life function decline and Age increase as their feature values increase. In addition, the downward trend is observed in Figures 6C–E, which is the same as their trend for MCI. We also discover that the Eating habits (1 = Vegetarian-based diet, 2 = Meat-based diet, 3 = Meat, and vegetables) have a negative impact on classifier when its value is 1 or 3, which indicates that eating vegetables is helpful to prevent AD. Meanwhile, as shown in Figure 6G, the SHAP value is below 0 if the value of the Frequency of nap (elder) is 4 (0 = None, 1 = Sometimes, 2 = 1–3 days a week, 3 = 4–6 days a week, 4 = Every day), which means regular naps may help to reduce the risk of AD.
Figure 6. SHAP dependency plots for AD class. (A) Daily life function decline, (B) age, (C) eat fish, (D) education, (E) diastolic blood pressure, (F) eating habits, (G) Frequency of nap (elder).
SHAP can also conduct an analysis of a single sample. Figure 4 shows the contribution of each feature value to the classifier's judgement of MCI and AD instances. Each feature value is a force that either increases or decreases the prediction. As shown in Figure 4A, the sample is classified as MCI with a probability of 62%. The top four features are Memory decline, Diastolic blood pressure, Eat fish, and Education. And these values of features increase the probability that the classifier will judge the sample as MCI. Figure 4B shows the same thing for AD class. The model is 81% confident that the sample is AD. The Daily life function decline, Diastolic blood pressure, Education, and Sleeping time (elder) play important roles to push the prediction decision toward AD class.
4. Discussion
This study develops an explainable machine-learning framework to predict AD/MCI based on clinical data obtained from CLAS. The performance of the framework has been improved by oversampling. We also apply multiple classification methods and feature selection methods for comparison, so that the best methods for prediction are selected. The resulting model achieves accuracy of 89.2%, sensitivity of 87.7%, specificity of 90.7%, G-mean of 89.2%, and AUC of 0.957 for MCI/NC prediction, while it achieves accuracy of 99.2%, sensitivity of 99.7%, specificity of 98.7%, G-mean of 99.2%, and AUC of 1 for AD/NC prediction. Then we make a detailed analysis by visualizing the specific contributions of the features to the classifier's output. To the best of our knowledge, this is the first attempt to employ ensemble learning with feature selection to develop models for AD prediction based on the large-size lifestyle and medical information. The strengths of this study include an unprecedentedly large-size dataset, an advanced machine learning-based algorithm that jointly considers the associations among the clinical and lifestyle features toward an effective feature set, as well as an explainable prediction model.
Our results are compatible with previous intuitions and scientific knowledge. Tables 9, 10 summarize the existing studies about the relationship between some features and AD/MCI. In the two tables, 13 features and 10 features are shown to be associated with MCI and AD, respectively, which overlaps with our selected features. This validates that our feature selection method is fairly reasonable. Hebert et al. (1995) and Petersen et al. (2018) discover that the risk of MCI and AD increases with the increasing age. High education may reduce the risk of MCI and AD claimed by Sattler et al. (2012). Marshall et al. (2012) believe that daily life function decline aggravates the severity of dementia. The trends of the Diastolic blood pressure and Systolic blood pressure in Figures 3F, G, 6E also verify the conclusion in Ou et al. (2020).
Regarding the lifestyle, eating fish is beneficial for preventing MCI and AD, discussed by Barberger-Gateau et al. (2007) and Sinn et al. (2012). Barberger-Gateau et al. (2007) also find that frequent consumption of vegetables may decrease the risk of AD. Tea intake may reduce the risk for dementia discussed by Kakutani et al. (2019). Shi et al. (2018) and Brachem et al. (2020) find that the poor sleep quality can enhance the risk of MCI and AD. Cross et al. (2015) discover that the relationship between nap and the risk of dementia exists. The conclusions of these papers above are reflected accordingly in Figures 3, 6, which indicates that our model is fairly reasonable. In addition to these supportive research, our results further demonstrate that AD is a complicated disease that is affected by multiple factors, including daily lifestyle and physical disease. With an advanced feature selection and a unified framework of machine learning, we are able to detect the combination of such contributive features.
5. Conclusion
We develop an explainable machine-learning based model with oversampling and feature selection methods. The oversampling method is used to generate new samples for the minority class to solve the data imbalance issue. The feature selection method is applied to reduce the data dimension, so as to lower the computational complexity of the model, and to find out the most important features. We adopt the ensemble learning method to implement the prediction. Our model not only realizes the prediction, but also provides the specific contribution of each feature to the prediction classifiers by building an explainer. Experimental results demonstrate that the model achieves excellent performance, which coincides with other prior research. In sum, our model not only provides the prediction outcome, but also helps to understand the relationship between lifestyle/physical disease and cognitive function, and enables clinicians to make appropriate recommendations for the elderly. Therefore, our approach provides a new perspective for the design of a computer-aided diagnosis system for AD, and has potential high clinical application value.
6. Future work
The study has several limitations. Firstly, the cross-sectional study is not able to examine causal relationships between life style and individual cognitive decline. Follow-ups are needed to make the final outcome for these population. Furthermore, we did not include FDG-PET, Aβmarkers, and APOE genotype in this work, so the true extent of AD pathology remains unknown.Additionally, the exploration of multimodal data encompassing neuropsychological tests, structural, and functional neuroimaging data, genetic information, and other relevant biological indicators will be undertaken in future research to provide a multifaceted understanding of the pathophysiology of MCI and AD. The analysis of multimodal data through advanced machine learning and artificial intelligence techniques will be employed to unveil hidden patterns and relationships, aiding in the better understanding of cognitive decline risk factors and pathophysiological mechanisms. Moreover, these technologies will be harnessed to develop predictive models for the early identification of MCI and AD risks.
Data availability statement
All the code written to process and analyze the data can be made available upon request to the corresponding author. The CLAS data are not publicly available due to privacy restrictions.
Ethics statement
The studies involving humans were approved by the Ethics Committee of Shanghai Mental Health Center. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributions
LY: Writing—original draft. W-gC: Visualization, Writing—original draft. S-cL: Writing—original draft. S-bC: Conceptualization, Data curation, Writing—review & editing. S-fX: Conceptualization, Data curation, Writing—review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This research was funded by the China Ministry of Science and Technology (2009BAI77B03, STI2030-Major Projects-2022ZD0213100), the National Natural Science Foundation of China (82001123).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Alhaj, T. A., Siraj, M. M., Zainal, A., Elshoush, H. T., and Elhaj, F. (2016). Feature selection using information gain for improved structural-based alert correlation. PLoS ONE 11:e0166017. doi: 10.1371/journal.pone.0166017
Altman, N. S. (1992). An introduction to kernel and nearest-neighbor nonparametric regression. Am. Stat. 46, 175–185. doi: 10.1080/00031305.1992.10475879
American Psychiatric Association (2000). Quick Reference to the Diagnostic Criteria From DSM-IV-TR. Washington, DC: American Psychiatric Association.
Anstey, K. J., von Sanden, C., Salim, A., and O'Kearney, R. (2007). Smoking as a risk factor for dementia and cognitive decline: a meta-analysis of prospective studies. Am. J. Epidemiol. 166, 367–378. doi: 10.1093/aje/kwm116
Barberger-Gateau, P., Raffaitin, C., Letenneur, L., Berr, C., Tzourio, C., Dartigues, J.-F., et al. (2007). Dietary patterns and risk of dementia: the three-city cohort study. Neurology 69, 1921–1930. doi: 10.1212/01.wnl.0000278116.37320.52
Bergstra, J., and Bengio, Y. (2012). Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13, 281–305. doi: 10.5555/2188385.2188395
Bergstra, J., Yamins, D., and Cox, D. D. (2013). “Hyperopt: a python library for optimizing the hyperparameters of machine learning algorithms,” in Proceedings of the 12th Python in Science Conference (SciPy 2013), Vol. 13 (Austin, TX). Available online at: https://conference.scipy.org/proceedings/scipy2013/
Brachem, C., Winkler, A., Tebrügge, S., Weimar, C., Erbel, R., Joeckel, K.-H., et al. (2020). Associations between self-reported sleep characteristics and incident mild cognitive impairment: the Heinz Nixdorf recall cohort study. Sci. Rep. 10, 1–12. doi: 10.1038/s41598-020-63511-9
Brookmeyer, R., Gray, S., and Kawas, C. (1998). Projections of Alzheimer's disease in the United States and the public health impact of delaying disease onset. Am. J. Publ. Health 88, 1337–1342. doi: 10.2105/AJPH.88.9.1337
Brookmeyer, R., Johnson, E., Ziegler-Graham, K., and Arrighi, H. M. (2007). Forecasting the global burden of Alzheimer's disease. Alzheimer's Dement. 3, 186–191. doi: 10.1016/j.jalz.2007.04.381
Cai, H., Su, N., Li, W., Li, X., Xiao, S., and Sun, L. (2021). Relationship between afternoon napping and cognitive function in the ageing Chinese population. Gen. Psychiatry 34:e100361. doi: 10.1136/gpsych-2020-100361
Chen, T., and Guestrin, C. (2016). “Xgboost: a scalable tree boosting system,” in KDD '16: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (San Francisco, CA), 785–794.
Collie, A., Maruff, P., Shafiq-Antonacci, R., Smith, M., Hallup, M., Schofield, P., et al. (2001). Memory decline in healthy older people: implications for identifying mild cognitive impairment. Neurology 56, 1533–1538. doi: 10.1212/WNL.56.11.1533
Cortes, C., and Vapnik, V. (1995). Support-vector networks. Mach. Learn. 20, 273–297. doi: 10.1007/BF00994018
Cross, N., Terpening, Z., Rogers, N. L., Duffy, S. L., Hickie, I. B., Lewis, S. J., et al. (2015). Napping in older people 'at risk' of dementia: relationships with depression, cognition, medical burden and sleep quality. J. Sleep Res. 24, 494–502. doi: 10.1111/jsr.12313
Danso, S. O., Zeng, Z., Muniz-Terrera, G., and Ritchie, C. W. (2021). Developing an explainable machine learning-based personalised dementia risk prediction model: a transfer learning approach with ensemble learning algorithms. Front. Big Data 4:21. doi: 10.3389/fdata.2021.613047
Dewancker, I., McCourt, M., and Clark, S. (2016). Bayesian optimization for machine learning: a practical guidebook. arXiv preprint arXiv:1612.04858. doi: 10.48550/arXiv.1612.04858
El-Sappagh, S., Alonso, J. M., Islam, S. R., Sultan, A. M., and Kwak, K. S. (2021). A multilayer multimodal detection and prediction model based on explainable artificial intelligence for Alzheimer's disease. Sci. Rep. 11, 1–26. doi: 10.1038/s41598-021-82098-3
Falck, R. S., Landry, G. J., Best, J. R., Davis, J. C., Chiu, B. K., and Liu-Ambrose, T. (2017). Cross-sectional relationships of physical activity and sedentary behavior with cognitive function in older adults with probable mild cognitive impairment. Phys. Ther. 97, 975–984. doi: 10.1093/ptj/pzx074
Farina, N., Rusted, J., and Tabet, N. (2014). The effect of exercise interventions on cognitive outcome in Alzheimer's disease: a systematic review. Int. Psychogeriatr. 26, 9–18. doi: 10.1017/S1041610213001385
Freund, Y., and Schapire, R. E. (1996). “Experiments with a new boosting algorithm,” in ICML'96: Proceedings of the Thirteenth International Conference on International Conference on Machine Learning (San Francisco, CA: Morgan Kaufmann Publishers Inc.), 148–156.
Friedman, J., Hastie, T., and Tibshirani, R. (2010). A note on the group lasso and a sparse group lasso. arXiv preprint arXiv:1001.0736. doi: 10.48550/arXiv.1001.0736
Geurts, P., Ernst, D., and Wehenkel, L. (2006). Extremely randomized trees. Mach. Learn. 63, 3–42. doi: 10.1007/s10994-006-6226-1
Gini, C. W. (1971). Variability and mutability, contribution to the study of statistical distributions and relations. studi cconomico-giuridici della r. Universita de cagliari (1912). Reviewed in: Light, rj, margolin, bh: an analysis of variance for categorical data. J. Am. Stat. Assoc. 66, 534–544. doi: 10.1080/01621459.1971.10482297
Haibo, X., Shifu, X., Pin, N. T., Chao, C., Guorong, M., Xuejue, L., et al. (2013). Prevalence and severity of behavioral and psychological symptoms of dementia (BPSD) in community dwelling Chinese: findings from the shanghai three districts study. Aging Ment. Health 17, 748–752. doi: 10.1080/13607863.2013.781116
He, H., Bai, Y., Garcia, E. A., and Li, S. (2008). “Adasyn: adaptive synthetic sampling approach for imbalanced learning,” in 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence) (Hong Kong), 1322–1328. Available online at: https://ieeexplore.ieee.org/document/4633969
Hebert, L. E., Scherr, P. A., Beckett, L. A., Albert, M. S., Pilgrim, D. M., Chown, M. J., et al. (1995). Age-specific incidence of Alzheimer's disease in a community population. JAMA 273, 1354–1359. doi: 10.1001/jama.1995.03520410048025
Hutter, F., Hoos, H. H., and Leyton-Brown, K. (2011). “Sequential model-based optimization for general algorithm configuration,” in Learning and Intelligent Optimization: 5th International Conference, LION 5 (Rome: Springer), 507–523. doi: 10.1007/978-3-642-25566-3_40
Kakutani, S., Watanabe, H., and Murayama, N. (2019). Green tea intake and risks for dementia, Alzheimer's disease, mild cognitive impairment, and cognitive impairment: a systematic review. Nutrients 11:1165. doi: 10.3390/nu11051165
Ke, G., Meng, Q., Finley, T., Wang, T., Chen, W., Ma, W., et al. (2017). LightGBM: a highly efficient gradient boosting decision tree. Adv. Neural Inform. Process. Syst. 30, 3146–3154. Available online at: https://proceedings.neurips.cc/paper/2017/hash/6449f44a102fde848669bdd9eb6b76fa-Abstract.html
Kononenko, I. (1994). “Estimating attributes: analysis and extensions of relief,” in European Conference on Machine Learning (Berlin; Heidelberg: Springer), 171–182.
Kotsiantis, S., Kanellopoulos, D., and Pintelas, P. (2006). Handling imbalanced datasets: a review. GESTS Int. Trans. Comput. Sci. Eng. 30, 25–36. Available online at: https://www.semanticscholar.org/paper/Handling-imbalanced-datasets%3A-A-review-Kotsiantis-Kanellopoulos/95dfdc02010b9c390878729f459893c2a5c0898f
Liaw, A., and Wiener, M. (2002). Classification and regression by randomforest. R News 2, 18–22. Available online at: https://journal.r-project.org/articles/RN-2002-022/
Livingston, G., Sommerlad, A., Orgeta, V., Costafreda, S. G., Huntley, J., Ames, D., et al. (2017). Dementia prevention, intervention, and care. Lancet 390, 2673–2734. doi: 10.1016/S0140-6736(17)31363-6
Loddo, A., Buttau, S., and Di Ruberto, C. (2022). Deep learning based pipelines for Alzheimer's disease diagnosis: a comparative study and a novel deep-ensemble method. Comput. Biol. Med. 141:105032. doi: 10.1016/j.compbiomed.2021.105032
Loh, W.-Y. (2011). Classification and regression trees. Wiley Interdisc. Rev. 1, 14–23. doi: 10.1002/widm.8
Lundberg, S., and Lee, S.-I. (2017). A unified approach to interpreting model predictions. arXiv preprint arXiv:1705.07874. Available online at: https://proceedings.neurips.cc/paper/2017/hash/8a20a8621978632d76c43dfd28b67767-Abstract.html
Marshall, G. A., Amariglio, R. E., Sperling, R. A., and Rentz, D. M. (2012). Activities of daily living: where do they fit in the diagnosis of Alzheimer's disease? Neurodegener. Dis. Manage. 2, 483–491. doi: 10.2217/nmt.12.55
Ngandu, T., Lehtisalo, J., Solomon, A., Levälahti, E., Ahtiluoto, S., Antikainen, R., et al. (2015). A 2 year multidomain intervention of diet, exercise, cognitive training, and vascular risk monitoring versus control to prevent cognitive decline in at-risk elderly people (finger): a randomised controlled trial. Lancet 385, 2255–2263. doi: 10.1016/S0140-6736(15)60461-5
Norton, S., Matthews, F. E., Barnes, D. E., Yaffe, K., and Brayne, C. (2014). Potential for primary prevention of Alzheimer's disease: an analysis of population-based data. Lancet Neurol. 13, 788–794. doi: 10.1016/S1474-4422(14)70136-X
Ou, Y.-N., Tan, C.-C., Shen, X.-N., Xu, W., Hou, X.-H., Dong, Q., et al. (2020). Blood pressure and risks of cognitive impairment and dementia: a systematic review and meta-analysis of 209 prospective studies. Hypertension 76, 217–225. doi: 10.1161/HYPERTENSIONAHA.120.14993
Pereira, V. H., Costa, P. S., Santos, N. C., Cunha, P. G., Correia-Neves, M., Palha, J. A., et al. (2016). Adult body height is a good predictor of different dimensions of cognitive function in aged individuals: a cross-sectional study. Front. Aging Neurosci. 8:217. doi: 10.3389/fnagi.2016.00217
Petersen, R. C., Doody, R., Kurz, A., Mohs, R. C., Morris, J. C., Rabins, P. V., et al. (2001). Current concepts in mild cognitive impairment. Arch. Neurol. 58, 1985–1992. doi: 10.1001/archneur.58.12.1985
Petersen, R. C., Lopez, O., Armstrong, M. J., Getchius, T. S., Ganguli, M., Gloss, D., et al. (2018). Practice guideline update summary: Mild cognitive impairment: Report of the guideline development, dissemination, and implementation subcommittee of the American academy of neurology. Neurology 90, 126–135. doi: 10.1212/WNL.0000000000004826
Sabia, S., Elbaz, A., Dugravot, A., Head, J., Shipley, M., Hagger-Johnson, G., et al. (2012). Impact of smoking on cognitive decline in early old age: the Whitehall ii cohort study. Arch. Gen. Psychiatry 69, 627–635. doi: 10.1001/archgenpsychiatry.2011.2016
Salvatore, C., Cerasa, A., Battista, P., Gilardi, M. C., Quattrone, A., and Castiglioni, I. (2015). Magnetic resonance imaging biomarkers for the early diagnosis of Alzheimer's disease: a machine learning approach. Front. Neurosci. 9:307. doi: 10.3389/fnins.2015.00307
Sattler, C., Toro, P., Schönknecht, P., and Schröder, J. (2012). Cognitive activity, education and socioeconomic status as preventive factors for mild cognitive impairment and Alzheimer's disease. Psychiatry Res. 196, 90–95. doi: 10.1016/j.psychres.2011.11.012
Shekhar, S., Bansode, A., and Salim, A. (2021). “A comparative study of hyper-parameter optimization tools,” in 2021 IEEE Asia-Pacific Conference on Computer Science and Data Engineering (CSDE) (Brisbane, OLD), 1–6. Available online at: https://ieeexplore.ieee.org/abstract/document/9718485
Shi, L., Chen, S.-J., Ma, M.-Y., Bao, Y.-P., Han, Y., Wang, Y.-M., et al. (2018). Sleep disturbances increase the risk of dementia: a systematic review and meta-analysis. Sleep Med. Rev. 40, 4–16. doi: 10.1016/j.smrv.2017.06.010
Sinn, N., Milte, C. M., Street, S. J., Buckley, J. D., Coates, A. M., Petkov, J., et al. (2012). Effects of n-3 fatty acids, EPA v. DHA, on depressive symptoms, quality of life, memory and executive function in older adults with mild cognitive impairment: a 6-month randomised controlled trial. Br. J. Nutr. 107, 1682–1693. doi: 10.1017/S0007114511004788
Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B 58, 267–288. doi: 10.1111/j.2517-6161.1996.tb02080.x
Xiao, S. (2013). Methodology of china's national study on the evaluation, early recognition, and treatment of psychological problems in the elderly: China longitudinal aging study (CLAS). Shanghai Arch. Psychiatry 25, 91–98. doi: 10.3969/j.issn.1002-0829.2013.02.005
Xiao, S., Lewis, M., Mellor, D., McCabe, M., Byrne, L., Wang, T., et al. (2016). The china longitudinal ageing study: overview of the demographic, psychosocial and cognitive data of the shanghai sample. J. Ment. Health 25, 131–136. doi: 10.3109/09638237.2015.1124385
Keywords: Alzheimer's disease, mild cognitive impairment, ensemble learning, feature selection, explainable AI
Citation: Yue L, Chen W-g, Liu S-c, Chen S-b and Xiao S-f (2023) An explainable machine learning based prediction model for Alzheimer's disease in China longitudinal aging study. Front. Aging Neurosci. 15:1267020. doi: 10.3389/fnagi.2023.1267020
Received: 25 July 2023; Accepted: 09 October 2023;
Published: 03 November 2023.
Edited by:
Ines Moreno-Gonzalez, University of Malaga, SpainReviewed by:
Lorenzo Carnevale, Mediterranean Neurological Institute Neuromed (IRCCS), ItalyValentina Bessi, University of Florence, Italy
Copyright © 2023 Yue, Chen, Liu, Chen and Xiao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Sheng-bo Chen, Y2NiMDJraW5nZG9tJiN4MDAwNDA7Z21haWwuY29t; Shi-fu Xiao, eGlhb3NoaWZ1JiN4MDAwNDA7bXNuLmNvbQ==
†These authors have contributed equally to this work