Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 06 September 2024
Sec. Translational Neuroscience

Development of a robust parallel and multi-composite machine learning model for improved diagnosis of Alzheimer's disease: correlation with dementia-associated drug usage and AT(N) protein biomarkers

  • 1Department of Computer Application, Faculty of Engineering & IT, Integral University, Lucknow, India
  • 2Department of Computer Science, Faculty of Science, Aligarh Muslim University, Aligarh, India
  • 3Department of Computer Science, College of Engineering and Computer Science, Jazan University, Jazan, Saudi Arabia

Introduction: Machine learning (ML) algorithms and statistical modeling offer a potential solution to offset the challenge of diagnosing early Alzheimer's disease (AD) by leveraging multiple data sources and combining information on neuropsychological, genetic, and biomarker indicators. Among others, statistical models are a promising tool to enhance the clinical detection of early AD. In the present study, early AD was diagnosed by taking into account characteristics related to whether or not a patient was taking specific drugs and a significant protein as a predictor of Amyloid-Beta (Aβ), tau, and ptau [AT(N)] levels among participants.

Methods: In this study, the optimization of predictive models for the diagnosis of AD pathologies was carried out using a set of baseline features. The model performance was improved by incorporating additional variables associated with patient drugs and protein biomarkers into the model. The diagnostic group consisted of five categories (cognitively normal, significant subjective memory concern, early mildly cognitively impaired, late mildly cognitively impaired, and AD), resulting in a multinomial classification challenge. In particular, we examined the relationship between AD diagnosis and the use of various drugs (calcium and vitamin D supplements, blood-thinning drugs, cholesterol-lowering drugs, and cognitive drugs). We propose a hybrid-clinical model that runs multiple ML models in parallel and then takes the majority's votes, enhancing the accuracy. We also assessed the significance of three cerebrospinal fluid biomarkers, Aβ, tau, and ptau in the diagnosis of AD. We proposed that a hybrid-clinical model be used to simulate the MRI-based data, with five diagnostic groups of individuals, with further refinement that includes preclinical characteristics of the disorder. The proposed design builds a Meta-Model for four different sets of criteria. The set criteria are as follows: to diagnose from baseline features, baseline and drug features, baseline and protein features, and baseline, drug and protein features.

Results: We were able to attain a maximum accuracy of 97.60% for baseline and protein data. We observed that the constructed model functioned effectively when all five drugs were included and when any single drug was used to diagnose the response variable. Interestingly, the constructed Meta-Model worked well when all three protein biomarkers were included, as well as when a single protein biomarker was utilized to diagnose the response variable.

Discussion: It is noteworthy that we aimed to construct a pipeline design that incorporates comprehensive methodologies to detect Alzheimer's over wide-ranging input values and variables in the current study. Thus, the model that we developed could be used by clinicians and medical experts to advance Alzheimer's diagnosis and as a starting point for future research into AD and other neurodegenerative syndromes.

1 Introduction

Neurodegenerative diseases pose a significant challenge in contemporary medicine, presenting a substantial burden on healthcare systems worldwide and affecting the quality of life for millions globally (Whiteford et al., 2015). These disorders, exemplified by the gradual degeneration of the nervous system's structure and function, manifest in a myriad of ways, affecting cognition, motor skills, and overall neurological wellbeing. Neurological disorders affect roughly 15% of the global population at present (Feigin et al., 2020). Over the last three decades, the actual number of affected individuals has substantially increased.

After conducting a comprehensive analysis of various neurological conditions such as Alzheimer's Disease (AD), Huntington's Disease, Parkinson's Disease, and amyotrophic lateral sclerosis, dementia becomes evident as a significant outcome of neurological deterioration, with AD being the most prominent (Ritchie and Ritchie, 2012; Ciurea et al., 2023; Khan et al., 2023). Neurodegenerative disorders are multifaceted, and thus it is complex to diagnose since genetic, environmental, and age-related factors cause them.

Cognitive decline, a characteristic of AD and related dementias, encompasses a spectrum of cognitive impairments ranging from subtle changes in memory and thinking abilities to severe cognitive dysfunction affecting daily functioning (Whiteford et al., 2015; Feigin et al., 2020). Different etiologies contribute to cognitive decline, including neurodegenerative processes such as AD, vascular pathology, Lewy body disease, and other less common causes. Syndromic diagnosis, which includes subjective cognitive decline, mild cognitive impairment, and dementia stages of AD, plays a crucial role in characterizing the progression of cognitive decline (Ritchie and Ritchie, 2012; Ciurea et al., 2023).

Dementia currently has a staggering societal cost, accounting for 1.01% of global GDP (Mattap et al., 2022). This issue is expected to worsen in the coming years, with an estimated 85% increase in global societal costs by 2030, assuming no changes in potential underlying causes (e.g., macroeconomic aspects, dementia incidence and prevalence, treatment availability, and efficacy). As stated in the World Alzheimer Report 2023, the World Health Organization (WHO) warns of a rising global prevalence of dementia, which is anticipated to rise from 55 million in 2019 to 139 million by the year 2050. As societies continue to age, the related expenses of dementia are estimated to double, from $1.3 trillion in 2019 to $2.8 trillion by 2030 (Better, 2023).

AD poses a significant challenge to the global health landscape and is characterized by its relentless progression as a neurodegenerative disorder. The global rise in AD cases is directly linked to the aging population, with a rising number of people surviving above the age of 65 (the key age group prone to AD) (Jaul and Barron, 2017). The growing prevalence of AD in an aging population underscores the pressing need for a comprehensive understanding of the condition and the development of innovative diagnostic techniques (Saleem et al., 2022; Alqahtani et al., 2023). The defining characteristic of AD is the gradual deterioration of cognitive abilities, which ultimately compromises the quality of life for those affected. This disorder not only has a detrimental effect on memory but also disrupts various cognitive, behavioral, and daily functioning (Khan et al., 2023). The implications of this condition extend beyond those affected, impacting their families and caregivers and imposing a burden on healthcare systems globally.

The traditional diagnostic methods, however valuable, often fail to deliver timely and precise diagnoses. AD detection needs a more refined and advanced technique due to its complicated nature. Such an approach should not only involve identifying symptoms but also discerning underlying pathological changes in the brain. The current diagnostic framework for AD encompasses a comprehensive approach that combines clinical assessments, neuropsychological tests, neuroimaging methods, and biomarker analysis (Martí-Juan et al., 2020; Khan and Zubair, 2022a,b). Although these methodologies have provided valuable insights, there are still persistent problems, such as the need for early detection and the development of more accurate and reliable diagnostic tools. Thus, it is within this context that the role of protein biomarkers becomes crucial. Protein biomarkers, including Amyloid-Beta (Aβ), tau, and ptau [AT(N)] have emerged as a potential indicator in identifying the complex molecular and cellular alterations linked with AD (Martí-Juan et al., 2020; Khan et al., 2023). These biomarkers aid in the early detection of AD pathology, facilitating diagnosis at the MCI stage when interventions may be most effective. The incorporation of these indicators into diagnostic frameworks is consistent with the continued research of new strategies to address the challenges encountered by AD. However, diagnosis at preclinical stages, characterized by the absence of clinical symptoms, is not recommended in clinical practice and is primarily reserved for research purposes.

Machine learning (ML), a subfield of artificial intelligence, has emerged as a transformative technology in the healthcare industry, with the potential to change how neurodegenerative diseases are diagnosed and managed (Alowais et al., 2023). ML algorithms, equipped with the capacity to analyze vast datasets, recognize patterns, and derive meaningful insights, offer a paradigm shift in identifying subtle changes in neurological parameters that precede overt symptoms (Bhatia et al., 2022; Javaid et al., 2022). By assimilating information from multiple sources, for instance, neuroimaging, clinical data, and genetic profiling, ML algorithms contribute to the development of predictive models that aid in early detection and personalized treatment strategies (Jiang et al., 2017; Ahmed et al., 2020; Hossain and Assiri, 2020; Khan and Zubair, 2022a,b; Arafah et al., 2023; Assiri and Hossain, 2023).

In the present study, the optimization of predictive models for the diagnosis of AD pathologies was carried out using a set of baseline features, and the model performance was improved by incorporating additional variables associated with patient drugs and protein biomarkers into the model. Early AD was diagnosed by considering two key criteria: firstly, whether a patient was taking specific medications, and secondly, the presence of a significant protein serving as a predictor of Aβ, tau, and ptau levels among participants. In particular, we examined the relationship between AD diagnosis and the use of various medications (calcium and vitamin D supplements, blood-thinning medications, cholesterol-lowering drugs, and cognitive drugs). We also assessed the significance of three cerebrospinal fluid (CSF) biomarkers, tau, ptau, and Aβ in the diagnosis of AD. The relative importance of these biomarkers in diagnosing AD is still a topic of discussion in the academic community (Brookmeyer et al., 2007; Gauthier et al., 2021).

The adoption of a hybrid-clinical model, incorporating the simultaneous operation of multiple ML models in parallel, emerges as a viable strategy for enhancing predictive accuracy. Given the heterogeneous nature of the dataset under consideration, employing multiple ML models in parallel allows for a comprehensive classification approach. Subsequent to the individual classification outputs generated by each classifier, a majority voting mechanism is employed to aggregate predictions. This collective decision-making process, leveraging the consensus among classifiers, serves to enhance overall predictive accuracy. Notably, the incorporation of parallelization principles within our model framework not only contributes to improved performance but also facilitates efficiency gains by optimizing computational resources.

The proposed model is used to simulate the MRI-based data, with five diagnostic groups of individuals (cognitively normal, significant subjective memory concern, early mildly cognitively impaired, late mildly cognitively impaired, and AD), with a further refinement which includes preclinical characteristics of the disorder. It is noteworthy that we aimed to construct a pipeline design employing ML that incorporates comprehensive methodologies to detect Alzheimer's over a wide- ranging input values and variables in the current study. The proposed design builds a meta-model based on four distinct sets of criteria, which include diagnosing from baseline features, baseline and medication features, baseline and protein features, and baseline, medication, and protein features. The meta-model incorporated a 4-step data preprocessing strategy, followed by feature wrapping using the step-forward technique. Furthermore, twelve efficient ML algorithms served as base classifiers. During the construction of the hybrid model, both stacking and voting techniques were employed. Preceding this, cross-validation with 5 and 10 folds was implemented alongside hyperparameter optimization. Subsequently, performance evaluation and comparison were conducted based on various metrics.

Thus, this research seeks to contribute to the broader effort of improving diagnostic approaches for Alzheimer's. This study aims to develop a robust multi-composite machine learning model that improves diagnostic accuracy by studying the intricate relationship between protein biomarkers, drugs, and AD. The model that we have developed offers a tool for healthcare practitioners to advance Alzheimer's diagnosis while also laying the groundwork for further investigation into AD and other neurodegenerative conditions.

2 Methods

2.1 Proposed design

In this section, we present the proposed meta-model, followed by an outline of the key steps involved in our method. The purpose of creating a pipeline environment is to streamline the entire process and ensure that the procedure is successful, i.e., to facilitate internal verification and produce outcomes that are reproducible externally. A schematic flow of our end-to-end approach is illustrated in Figures 1, 2. In the subsequent sections, a detailed description of the proposed approach and the various steps undertaken in this study are presented.

Figure 1
www.frontiersin.org

Figure 1. Hybrid-clinical model architecture.

Figure 2
www.frontiersin.org

Figure 2. Workflow design for data cleansing framework.

Every machine learning model possesses unique characteristics that enable it to achieve success with specific types of data. As this task involves various kinds and groups of data, we suggest utilizing multiple machine learning models in parallel to classify the data. Once we obtain the output of each classifier, which is a specific class, we aggregate the majority vote of the classifiers to improve overall accuracy. Our model employs the principle of parallelism to increase system accuracy and expedite processing time.

2.2 Study design, participants and dataset collection

Alzheimer's Disease Neuroimaging Initiative (ADNI) is a comprehensive repository that was established in 2004 and is headed by Michael W. Weiner as the principal investigator.1 This repository contains data on clinical, biochemical, genetic, and imaging biomarkers for detecting AD and MCI at an early stage, monitoring their progression, and tracking their development over time through a series of longitudinal, multicenter studies.

The baseline statistical model that we designed was based on the ADNIMERGE dataset from ADNI, which contained selected factors relevant to individuals' clinical, genetic, neuropsychological, and imaging results. The ADNIMERGE dataset consists of four distinct studies, namely ADNI-1, ADNI-2, ADNI-3, and ADNI-GO, which were collected at varying stages of the research project and represent different time periods. Each dataset includes new patients who were enrolled during the study period, as well as previous patients who were continuously monitored. The ADNIMERGE dataset includes 2,175 individuals, ranging in age from 54 to 92 years, and contains 14,036 input values for 113 features. These values were collected over a period of 8 years (2004–2021), with the initial measurement taken when the patient first arrived, followed by a 6-month follow-up visit every year for 8 years. We selected only patients who participated in the ADNI-1 phase of the initiative, which comprised 818 individuals and a total of 5,013 input values across 113 variables. These participants were characterized by demographic information, neuropsychological, genetic, MRI, Diffusion-tensor imaging (DTI), electroencephalography, and positron emission tomography (PET) biomarkers. This was done in order to maintain uniformity across studies and data handling, as well as to ensure that we could successfully select only a single observation for each subject.

This study included and integrated three varied datasets i.e., baseline data, drug data and protein biomarker data. Separate data files containing drug and protein biomarker data were stored on the ADNI repository. They were examined to extract the pertinent ones and were added to the baseline data.

2.2.1 Baseline dataset

The baseline dataset comprised 818 participants from the ADNI-1 dataset. given our objective of using predictors that have been consistently evaluated and effectively integrated into clinical settings, while also being non-intrusive to patients, we have chosen to exclusively utilize variables from the ADNI-1 baseline dataset. these variables pertain to diagnostic subtypes, demographic details, and scores from clinical and neuropsychological tests. only diagnoses that were confirmed from screening up until the baseline visit were considered, whereas any patient data with more than 20% missing information were eliminated.

Based on the diagnosis at the follow-up visit, the patients who were diagnosed at baseline were classified into 5 distinct categories: Cognitively Normal, Early Mild Cognitive Impairment, Late Mild Cognitive Impairment, Significant Memory Complaints, and Alzheimer's Disease, abbreviated as CN, EMCI, LMCI, SMC, and AD. The diagnostic classes EMCI and LMCI were unified into a single diagnostic class MCI. An additional advantage of doing so was that it made it easier to extend our model in the future for analyses that might contain additional ADNI data. Because SMC patients met the criteria for being cognitively normal, the CN and SMC classes were combined into a single category. The diagnosis features, which served as a response variable, were thus divided into 3 categories: AD, MCI, and CN. Among the 818 patients involved in the study, 193 individuals were identified with AD, 396 with MCI, and 229 with CN. The demographic information of the study participants, categorized by their baseline diagnosis, is presented in Table 1. Age, years of education, gender, marital status, ADAS11, ADAS13, ADASQ4, CDRSB, DIGITSCOR, FAQ, LDETOTAL, MMSE, mPACCdigit, mPACCtrailsB, TRABSCOR, RAVLT-I, RAVLT-L, RAVLT-F, and RAVLT-PF, are the final variables considered in the present study. The statistics and descriptions of these variables are provided in Table 2.

Table 1
www.frontiersin.org

Table 1. Subject demographics.

Table 2
www.frontiersin.org

Table 2. Variable description and related statistics.

2.2.2 Drug dataset

We created five new variables from components inside the drug dataset that fall into one of our five analytic categories: blood thinners, calcium doses, cholesterol-lowering medicines, cognitive drugs, and vitamin d medicines. Table 3 contains a comprehensive list of individual drug and supplement names. it is critical to note that the use of any of these drugs did not preclude a patient from participating in the ADNI cohort (see footnote1).

Table 3
www.frontiersin.org

Table 3. Drug list.

2.2.3 Protein biomarker dataset

Using the ADNI dataset, we generated three new variables from components within the CSF biomarker data to determine the amounts of Aβ, tau, and ptau.

Therefore, altogether three unique datasets (baseline data, drug data, and protein data) were gathered, extracted and handled, and then were passed to the next step i.e., data cleansing framework for data pre-processing (Figure 2).

2.3 Data cleansing framework

A flowchart depicting the process of complete data cleansing is portrayed in Figure 2. Initially, the controllers were configured with specific data values, including C = 1 for baseline data, C = 2 for baseline and drug data, C = 3 for baseline and protein data, and C = 4 for baseline, drug, and protein data, respectively. Table 4 shows a complete description of the procedure for implementing the four sets of controllers. After the execution of the data cleansing process, 24 individual clean files were generated.

Table 4
www.frontiersin.org

Table 4. Execution steps for data cleansing framework.

3-step data pre-processing strategy:

The dataset acquired was processed with a three-step ITR approach i.e., Imputation (I), Transformation (T), and Reduction (R).

2.3.1 Imputation

In this study, the dataset encountered issues of noise, incompleteness, and inconsistency, which typically hinder the mining process. In general, inaccurate or dirty data often pose challenges for mining techniques, which can obstruct the extraction of valuable insights. The proportion of missing values for the extracted variables (at baseline) is depicted in Table 2. To overcome the challenge of missing data, there exists various techniques. Specifically in this study, three different approaches were employed to address this issue. The simplest method involved removing all instances of missing data, which was implemented as the first strategy, referred to as “without imputation.”

There are various methods for handling missing data, including weighting, case-based, and imputation-based techniques (Tartaglia et al., 2011). The latter technique was used in the present study. Imputation involves predicting missing data values and then filling them with suitable approximations, such as the mean, median or mode. Subsequently, standard complete-data techniques are then applied to the filled-in data to decrease the biases due to missing values and improve the efficiency of the model.2 The term “mean imputation” refers to replacing missing values with suitable approximations, such as the mean, and subsequently applying standard complete-data procedures to the filled-in data (Khan and Zubair, 2019). This was the second approach employed in the study. The “model imputation” method, on the other hand, involves replacing missing values with appropriate approximations, such as a linear regression model, and then using a standard complete-data process to the filled-in data (Khan and Zubair, 2019). This was the third approach used in the study.

2.3.2 Transformation

In this study, data transformation involved two main techniques: normalization and smoothing. These techniques aim to improve the quality and interpretability of the data (Pires et al., 2020; Maharana et al., 2022). In this study, normalization was applied to the ADNI dataset to standardize the scale of numerical values, which varied in range for different variables. The values were adjusted and transformed in a way that they fall within a specified range, often between 0 and 1 (Pires et al., 2020; Maharana et al., 2022). This adjustment ensured that each variable had equal importance in the analysis and prevented any one variable from dominating due to its scale. Smoothing was then performed to remove any noise or irregularities in the ADNI data, making it easier to identify underlying patterns. This was particularly useful for our research, as the data contained random variations and anomalies that could have masked meaningful trends and patterns.

2.3.3 Reduction

Data reduction methodologies play a significant role in the analysis of reduced datasets while maintaining the integrity of the original data (Khan and Zubair, 2022a). This approach is often used to enhance efficiency, streamline analysis, and effectively manage large datasets (Maharana et al., 2022). There are several techniques for implementing data reduction i.e., dimension reduction, sampling, aggregation, and binning. Each method is applied based on the type of dataset and variables present. In this study, we employed a dimension-reduction strategy. This method facilitated in identifying and eliminating variables and dimensions that were insignificant, poorly correlated, or redundant.

2.4 Load clean data

The subsequent step involved loading the clean data (as depicted in Figure 1). Initially, the clean data files generated when the controller was set to 1 were loaded and the entire pipeline was executed. Similarly, this process was repeated for the remaining three controllers, 2, 3, and 4, resulting in the creation of twelve different ML meta-models. Later, comparisons were performed to determine which model performed optimally across a range of applied methods.

2.5 Machine learning modeling

Machine learning techniques were utilized to develop a classifier capable of identifying potential instances of AD and MCI. A hybrid-clinical classification model was constructed, incorporating variables selected during the feature selection process. This model development process was repeated for each of the 4 controllers separately. The clean data was passed to the base ML classifiers. The model was trained on the training set. The feature selection process was conducted to determine the optimal set of features. The performance was assessed, and it was made to run iteratively until the best feature set was identified. Subsequently, we applied 5-fold repeated stratified cross-validation and hyperparameter optimization techniques to obtain an optimized set of algorithms. The optimized classifier, trained on the complete training set, was then applied to the independent test set following 10 iterations of 5-fold repeated stratified cross-validation on the training set.

The entire modeling process is explained in the following sections and can be seen in Figures 1, 3, 4 pictorially.

Figure 3
www.frontiersin.org

Figure 3. Five-step process: step forward feature selection.

Figure 4
www.frontiersin.org

Figure 4. Ten-fold repeated stratified cross-validation.

2.5.1 Base ML models

We used a varied set of ML algorithms and techniques in our operations. A good implementation of the algorithms in question was known while selecting these tools for this study. In this study, twelve efficient ML models were used, namely, Multinomial Logistic Regression, K-Nearest Neighbors, Linear Discriminant Analysis, Quadratic Discriminant Analysis, Decision Tree, Random Forest, AdaBoost, Principal Component Analysis with Logistic Regression, Support Vector Machine - Radial Basis Function, Perceptron, MultiLayer Perceptron, and Elastic Nets. Earlier, fifteen supervised learning classifiers were executed to study the impact of the built model. However, we selected twelve classifiers out of these fifteen classifiers. Gaussian Process, LinearSVC and Stochastic Gradient Descent are the three ML algorithms that could not fit on the ADNI dataset efficiently and hence resulted in reduced performance. These 12 models were chosen for their ability to produce high performance during the model development step. These models acted as the base ML models for the creation of a hybrid meta-model.

The decision to use twelve ML classifiers out of fifteen can be attributed to the thorough consideration of algorithm performance and effectiveness in this study. The selection of these twelve classifiers was based on the following factors:

• Diverse Set of Algorithms: The initial set of fifteen classifiers likely encompassed a diverse range of ML algorithms, allowing for comprehensive coverage of different approaches. By including a variety of models, the study aimed to explore the strengths and weaknesses of various algorithms, ensuring a more robust understanding of the data and problem domain.

• Performance Evaluation: Initially, we explored fifteen supervised learning classifiers. The decision to narrow down to twelve models suggests that an extensive evaluation was conducted, and the performance of each algorithm was thoroughly assessed. Only the top-performing models were retained for further analysis.

• Elimination of Underperforming Models: The decision to exclude certain models was based on their underperformance and lack of contribution to achieving the study's objectives, emphasizing the importance of selecting the most effective algorithms.

The description of each employed classifier is presented below:

A. Multinomial Logistic Regression (MLR): In binary logistic regression, we estimate the probability of a single class, while in MLR, we extend this concept to estimate probabilities for multiple classes. This approach is particularly effective for dependent variables with three/more unordered categories. MLR applies a logistic function to each category to calculate the probability of belonging to that specific group, with k categories are represented by k-1 logistic functions (Khan et al., 2024). To interpret the results in terms of relative likelihoods, the probabilities are then normalized to ensure that they sum to 1 across all categories (Aguilera et al., 2006; Yang, 2019). MLR calculates a set of coefficients for each category, representing the correlation between predictor variables and log-odds of belonging to that group (Yang, 2019; Reddy et al., 2024). Each category has its intercept, which represents the log-odds of all predictor variables being zero. The MLR is usually trained using maximum likelihood estimation, where the model parameters are calculated to maximize the likelihood of observing the specific set of outcomes (Hedeker, 2003).

B. K-Nearest Neighbors (KNN): The KNN algorithm is a supervised ML technique that is primarily used for classification. The main concept of KNN is to generate predictions by examining the majority class amongst the “K” nearest data points. In the context of classification, when presented with a new data point, KNN examines the “K” data points in the training set that are in closest proximity to it (Li, 2024). The class with the highest prevalence among these nearby data points is then allocated to the new data point. Typically, Euclidean distance (a distance metric) is employed by KNN for measuring the similarity or proximity of data points within a multi-dimensional feature space (Zhang et al., 2022; Li, 2024). The parameter “K” indicates the total number of neighbors to take into account. The choice of an appropriate “K” is essential as it can significantly impact the quality of predictions. A small value of K can result in noisy predictions, whereas a larger value of K can help to identify underlying patterns more accurately (Li, 2024). The KNN classifier identifies a set of neighbors and then counts the number of instances of each class amongst these neighbors. The class with the most occurrences is then assigned to the new data point. This is an instance-based learning algorithm that memorizes the training instances rather than explicitly learning a model. KNN is a non-parametric method which means it makes no assumptions regarding the underlying distribution of the data (Wang et al., 2022).

C. Linear Discriminant Analysis (LDA): It is a supervised algorithm designed to discover the optimal linear combinations of features that efficiently separate different classes within a dataset (Tharwat et al., 2017). By maximizing the separation between these classes, LDA effectively results in the formation of different groups of data points (Seng and Ang, 2017). The algorithm intrinsically performs dimensionality reduction, which is an inherent feature of the algorithm. This involves projecting the data onto a lower-dimensional space while preserving the most informative features for classification (Seng and Ang, 2017; Graf et al., 2024). LDA makes the assumption that the covariance matrix is consistent across all classes and that the features within each class follow a normal distribution (Seng and Ang, 2017; Tharwat et al., 2017). To effectively distinguish the distribution of data points within and between classes, LDA computes mean vectors and scatter matrices. Mean vectors represent the centroids of data points in each class, whereas scatter matrices capture the spread or variability within each class. LDA then performs eigenvalue decomposition on the generalized eigenvalue problem that includes within-class and between-class scatter matrices (Tharwat et al., 2017). The directions of maximum discrimination are revealed by the eigenvectors that correspond to the largest eigenvalues. The data points are then projected onto these discriminant directions, resulting in a new space where the classes are clearly separated (Tharwat et al., 2017).

D. Quadratic Discriminant Analysis (QDA): It is a supervised classification algorithm that aims to determine the optimal boundaries for separating different classes in the feature space (Jiang et al., 2018). In contrast to LDA, QDA provides greater flexibility in capturing variability within each class by allowing distinct covariance matrices for each class (Witten et al., 2005; Tharwat et al., 2017). QDA, analogous to LDA, aims to maximize the degree of segregation among distinct classes. This is achieved through the identification of quadratic decision boundaries, which adequately represent the complex relationships among the features (Witten et al., 2005). Because LDA presumes that all classes utilize the same covariance matrix, QDA allows each class to possess a unique covariance matrix. Thus, QDA is more adaptable when handling classes that may exhibit diverse variability patterns (Siqueira et al., 2017). Similar to LDA, QDA calculates the mean vector for each class that serves as the centroid of the data points in that class. In addition, QDA determines the scatter matrices for each class, which reflect the dispersion or variability present within each class (Witten et al., 2005). Following that, it sets up quadratic decision boundaries that effectively separate classes using data from the class means and scatter matrices. The function of the decision boundaries is to classify the newly acquired data points. Compared to LDA, QDA is better at finding non-linear correlations and complex patterns within each class because it uses different covariance matrices (Siqueira et al., 2017).

E. Decision Tree (DT): A Decision Tree classifier is a tree-like model that follows a hierarchical structure, where input data points are classified based on a series of decisions made at each node (Khan and Zubair, 2022a,b). Leaf nodes represent the predicted class, whereas internal nodes reflect decisions based on particular features. The tree structure is composed of nodes that learn the most important features for classification. In DT, the features that provide the best separation of classes are positioned closer to the root of the tree (Blockeel et al., 2023; Costa and Pedreira, 2023). The root node is the top node in the hierarchy. It represents the complete dataset and is divided according to the feature that best separates classes. The process of selecting features and splitting the dataset recursively continues until a stopping criterion is met (Blockeel et al., 2023; Costa and Pedreira, 2023). When a new data point traverses the tree, it follows the decision path based on the conditions of the features until it gets to a leaf node. The predicted class for the input data point is then determined by the class associated with that leaf node.

F. Random Forest (RF): Random Forest is an ensemble learning ML model that aggregates predictions from multiple ML models to construct a more robust and precise model (Belle and Papantonis, 2021). It employs decision trees as its base model and uses a bagging technique, which trains several trees on various subsets of the training data (Campagner et al., 2023). For classification, the final prediction of the RF algorithm is determined through a voting mechanism. Each tree votes for a class and the class with the most votes is assigned to the input data point. RF produces multiple subsets of training data using random sampling with replacement (bootstrap sampling). Each subset trains a distinct DT. For each DT, a feature subset is randomly selected at each split. This approach guarantees that the generated trees are diverse, thereby minimizing the likelihood of overfitting to specific features (Wang et al., 2021; Campagner et al., 2023). Each DT is trained independently, allowing the ensemble to capture various aspects of the data and lessen overfitting risk (Campagner et al., 2023). The final prediction is determined through a voting method in which the individual tree predictions are aggregated. The class with the maximum votes is the final predicted class.

G. AdaBoost (AB): AdaBoost, an acronym for Adaptive Boosting, is an ensemble learning algorithm in which a robust classifier is constructed by combining several weak learners. The primary aim of the AB algorithm is to train weak classifiers iteratively on different subsets of the data and allocate high weights to instances that have been incorrectly classified during each iteration (Ying et al., 2013). The final model integrates the predictions of all weak learners with varying weights, giving preference to those that perform adequately on training data. AB classifier initializes each data point in the training set with equal weights (Ding et al., 2022). In subsequent iterations, the weights are modified to focus on instances that are difficult to correctly classify. Weak learners are trained sequentially; and at each iteration, a new weak classifier is fitted to the data (Ying et al., 2013; Ding et al., 2022). Instances that are misclassified are assigned greater weights, resulting in the final model being a weighted sum of all weak classifiers. The weights are calculated based on each classifier's accuracy on the training data, with models that demonstrate higher performance contributing proportionally more to the final prediction (Haixiang et al., 2016). The final model is a combination of all weak classifiers, with the weights chosen based on their accuracy on the training data.

H. Principal Component Analysis with Logistic Regression (PC-LR): PC-LR is a method that combines the Principal Component Analysis (PCA) method (meant for dimensionality reduction) with the logistic regression (LR) algorithm (meant for classification tasks). PCA creates a new set of uncorrelated features, known as principal components, which capture the maximum variance in the data, transforming the original features (Khan and Zubair, 2020). By employing this method, the dimension of the feature space is decreased. Logistic Regression is a widely-used classifier that effectively handles linearly separable data and calculates the likelihood of an instance belonging to a specific class (Yang, 2019). By combining the dimensionality reduction capabilities of PCA with the classification power of Logistic Regression, the PC-LR method aims to retain the most informative components while reducing overall dimensionality (Aguilera et al., 2006; Yang, 2019). The input data is first transformed into a lower-dimensional space using PCA, and then logistic regression is applied to make predictions based on the reduced feature set. The LR model is trained to determine the probability that a given instance belongs to a particular class. When generating predictions for new data, the class with the highest probability is assigned as the final prediction.

I. Support Vector Machine - Radial Basis Function (SVM-RBF): SVM is a supervised learning technique that seeks to identify a hyperplane in N-dimensional space (N being the number of features) that best distinguishes between data points from various classes (Siddiqui et al., 2023). The objective is to optimize the margin, defined as the distance between the hyperplane and the closest data points from each class. RBF is a widely utilized kernel function in SVM. Transforming the input space to a higher-dimensional space enables the RBF kernel to effectively capture non-linear correlations in the data (Ding et al., 2021). The transformed space enables SVM to identify a non-linear decision boundary within the original feature space. SVM-RBF searches for the hyperplane in the transformed space that effectively segregates the data points into their respective classes (Ding et al., 2021). The hyperplane is used to maximize the margin, thereby establishing a robust decision boundary. SVM allows for the existence of some misclassified data points to handle cases where a perfect separation is not possible (Siddiqui et al., 2023). The key data points that influence the decision boundary are called support vectors. The RBF kernel has a parameter called the gamma (γ), which determines the shape of the decision boundary (Valero-Carreras et al., 2021). Tuning the gamma parameter is crucial, as a small gamma may lead to underfitting, while a large gamma may lead to overfitting (Sacchet et al., 2015). After establishing the decision boundary, SVM-RBF is capable of classifying newly acquired data points by determining which side of the hyperplane they lie on.

J. Perceptron (PC): The Perceptron classifier, originally developed for binary classification, can be tweaked to perform multi-class classification using a strategy called the One-vs-All and One-vs-Rest (Raju et al., 2021). The OvA strategy involves training multiple Perceptrons, each dedicated to distinguishing one specific class from all the others (Kleyko et al., 2023). For K classes, K Perceptrons are trained, where each Perceptron focusses in recognizing one class and considers the instances of that class as the positive class and all other instances as the negative class. For a multi-class problem with K classes, K Perceptrons are trained. Each Perceptron is allocated to one class, and it aims to correctly classify instances belonging to that class against instances from all other classes (Chaudhuri and Bhattacharya, 2000; Raju et al., 2021). Throughout the training phase, the weights of each Perceptron are tweaked based on the instances belonging to its allocated class. The aim is to locate weights that reduce the classification error for that specific class. During the prediction phase, each Perceptron independently predicts for a given input instance. The class associated with the Perceptron that outputs the highest confidence (largest net input) is then assigned as the predicted class for that instance (Chaudhuri and Bhattacharya, 2000; Kleyko et al., 2023). In essence, the decision for multi-class classification is made by employing a one-vs-rest strategy, where each Perceptron is treated as a binary classifier for its assigned class vs. all other classes.

K. MultiLayer Perceptron (ML-PC): This classifier is a type of artificial neural network that is devised to handle challenging ML tasks, such as multi-class classification (Chaudhuri and Bhattacharya, 2000). An ML-PC is comprised of multiple interconnected layers, such as an input layer, one/more hidden layers, and an output layer (Liu et al., 2023). Data flows through the network in a feedforward manner, with each node in a layer processing information from the previous layer and passing it to the next layer. Each node's weighted sum of inputs is subjected to activation functions such as hyperbolic tangent (tanh), sigmoid, and rectified linear unit (ReLU). These functions add non-linearity, allowing the network to understand intricate connections. The input layer represents the features of the input data, with each node corresponding to a feature and the values being the feature values (Chaudhuri and Bhattacharya, 2000). Hidden layers process information from the input layer, capturing complex, non-linear patterns in the data. The output layer generates the final predictions for each class in a multi-class setting, with each node's output representing the model's confidence in predicting that class. Backpropagation, a supervised learning technique, operates to reduce errors by iteratively modifying weights. The difference between the predicted and actual outputs is determined by the loss function. In order to minimize this difference, optimization algorithms like gradient descent are implemented, which modify the weights (Chaudhuri and Bhattacharya, 2000; Liu et al., 2023). The learning rate determines the size of each weight update.

L. Elastic Nets (EN): Elastic Nets, a regularization technique, were originally designed for linear models. In multi-class classification setups, the linear model is extended to handle multiple classes (Mol et al., 2009). Elastic Networks employ a combination of L1 (Lasso) and L2 (Ridge) methods of regularization. L1 regularization encourages sparsity in the model, promoting feature selection, while L2 regularization prevents large coefficients (Zhan et al., 2023). For multi-class classification, Elastic Nets can be applied to extend linear models to predict probabilities for multiple classes. Elastic Nets can be used in conjunction with strategies like One-vs-All (OvA) and One-vs-One (OvO) to handle multi-class challenges (Chen et al., 2018). OvA trains a distinct model for each class against the rest, whereas OvO trains models for each pair of classes. The hyperparameters alpha and l1_ratio, which control the balance between L1 and L2 regularization, must be carefully chosen in Elastic Nets. In general, the output layer utilizes a softmax activation function to transform raw model outputs into class probabilities, with the certainty that the sum of the predicted probabilities equals 1. During training, Elastic Nets optimize the model's weights using algorithms like gradient descent, to reduce the disparity between predicted and actual class probabilities (Aqeel et al., 2023). The cross-entropy loss function is frequently employed in multi-class classification tasks.

2.5.2 Feature selection

Feature selection is a significant step while preparing data for ML modeling. A subset of the most pertinent features is chosen from the original set. The primary aim is to enhance model performance, simplify the model, and mitigate the risk of overfitting (Pudjihartono et al., 2022). A model with fewer features is often more interpretable, making it easier to understand the relationships between variables (Barnes et al., 2023).

There are 3 types of feature selection strategies i.e., filter, wrapper, and embedded. In this study, we employed the wrapper method (Dokeroglu et al., 2022; Pudjihartono et al., 2022; Kanyongo and Ezugwu, 2023). Wrapper methods involve evaluating the performance of a ML model based on different subsets of features. Unlike filter methods that assess feature relevance independently of the model, wrapper methods use the actual performance of the model as a criterion for selecting features. This involves training the model multiple times, which can be computationally expensive but may yield more accurate results (Kanyongo and Ezugwu, 2023). Furthermore, there are 3 types of wrapper methods viz. forward selection, recursive feature elimination, and backward elimination. In this study, we employed the step forward feature selection method. It is a specific wrapper method that builds the feature set incrementally. It starts with an empty set of features and adds them one at a time, based on how they affect the model's performance. The process involves five steps, illustrated in Figure 3.

2.5.3 Cross-validation and hyperparameter optimization

The objective of this study was to construct such a model that can exhibit optimal generalized performance rather than only for the cases used during training. Consequently, cross-validation gives an estimate of the overall performance for each hyperparameter configuration (Khan and Zubair, 2022a). To achieve this, the train data was divided into 5 folds. Instances from each fold were held-out from the training process, while the remaining cases were trained iteratively. Subsequently, the algorithm was then applied to the held-out samples following their training.

In this study, ten iterations of a 5-fold repeated stratified CV training and testing approach were implemented to maintain the distribution of classes across each fold (Figure 4). The Scikit-Learn library in Python, for instance, provides a RepeatedStratifiedKFold class that was employed for implementing this type of cross-validation. This was employed for evaluating and fine-tuning the models since we were dealing with scenarios where data variability and class imbalance challenges needed to be carefully managed.

An imbalanced classification problem can pose a significant challenge when building a ML model, especially when the data distribution is skewed toward the target variable (Kanyongo and Ezugwu, 2023). In such cases, if not addressed properly, the model may perform poorly, resulting in low accuracy. In the present study, we aimed to resolve the imbalanced classification problem by employing the Synthetic Minority Oversampling Technique (SMOTE) method. SMOTE is a data augmentation technique, designed to handle minority classes (Kohavi, 1995; Chawla et al., 2022). As previously stated, there were 3 classes of the target variable, including cognitively normal, MCI, and AD subjects. But we discovered a significant disparity in the MCI and AD classes. As a result, the built ML model resulted in poor performance and low accuracy. To address this issue, we utilized SMOTE analysis to oversample the minority class, which balanced the class distribution without adding any further information to the ML model.

The hyperparameters were fine-tuned before assessing each classifier. Hyperparameters are essential for structuring ML models and are not learned from the data during training. Hyperparameter optimization is the process of systematically searching for the best combination of hyperparameter values to achieve optimal performance from the model (Yang and Shami, 2020; Khan and Zubair, 2022a). It is crucial to optimize these hyperparameters to obtain the best possible results when applied to unseen instances. ML algorithms often have one or more hyperparameters that can be adjusted during the training process. By varying these hyperparameters, the algorithm's prediction performance can be varied. In this study, each model was trained using specific hyperparameter configurations to optimize the hyperparameters for each employed ML algorithm.

2.6 Analyze results

The purpose of this step was to assess the performance of the base models and determine whether they exhibited improved performance. If not, the model was subjected to additional training for an extended period of time, using a varied set of optimization hyperparameters.

2.7 Evaluate test performance

This stage determined the performance of test splits on the test data. If the performance was found to be poor, the test distribution was reevaluated, and any discrepancies were rectified by creating equitable splits. If the performance proved to be acceptable (with a high level of accuracy), the subsequent step was carried out and a meta-model was constructed.

2.8 Build meta-model

To construct a meta-model, we employed an ensemble learning approach that involved selecting different outputs generated by 12 machine learning classifiers after modeling and optimization. These classifiers were made to run in parallel and subjected to a voting and stacking process. Majority voting and majority stacking are ensemble learning approaches that integrate the predictions of numerous independent models to improve overall predictive performance (Raza, 2019; Dolo and Mnkandla, 2023). Both methods involve aggregating the decisions of multiple models, but they differ in their approaches. Majority voting encompasses multiple models making independent predictions on a given input, with the final prediction being determined by the majority vote or consensus of these individual predictions (Zhao et al., 2023). On the other hand, majority stacking is a more sophisticated method that trains a meta-model, often referred to as a stacker or meta-learner, to combine the predictions of multiple base models (Aboneh et al., 2022; Dolo and Mnkandla, 2023). The process of averaging predictions involves selecting the class with the most votes (the statistical mode) or the class with the highest summed probability. Stacking extends this method by enabling any machine learning model to learn how to integrate predictions from contributing members optimally.

After applying majority voting and majority stacking, we evaluated their performances individually and selected the approach that resulted in the best overall performance. The selected model, whether it was based on majority voting or majority stacking, effectively served as a meta-model in our ensemble learning approach. This approach often results in improved generalization and performance compared to relying on a single model.

2.9 Generalized performance metrics: evaluate and compare

Finally, performance evaluations were conducted and compared to all applied approaches for each of the four controllers established in this study. We employed accuracy, precision, recall, and F1-score as classification performance metrics for both the base model and meta-model. The meta-model was evaluated using two metrics for classification error: Hamming loss and the Jaccard index. Each parameter was determined in this study by employing a 3 × 3 confusion matrix (Figure 5). In ML classification, the confusion matrix is a widely used method that evaluates the performance of a model through a comparison between its predictions to the actual labels in a specific dataset. Table 5 presents a detailed description of each of the metrics derived from the confusion matrix.

Figure 5
www.frontiersin.org

Figure 5. 3 × 3 confusion matrix.

Table 5
www.frontiersin.org

Table 5. Description of the measures used.

2.10 Development environment and base settings of classification algorithms

To ensure the replicability of our study, we documented the development environment, including software and hardware specifications, as well as the configuration settings of the classification algorithms employed. The implementation of our predictive models for the diagnosis of AD pathologies was conducted using the Anaconda distribution of Python, which provides a comprehensive environment for data science and ML tasks. Anaconda includes a wide range of pre-installed libraries and tools, making it well-suited for developing and deploying ML models.

Jupyter Notebook, a web-based interactive computing environment, was utilized for the implementation of the study. Jupyter Notebooks offer a convenient interface for writing code, executing experiments, visualizing results, and cohesively presenting findings. The hardware utilized for model training and evaluation consisted of a system with an Intel Core-6100U CPU running at a base frequency of 2.30 GHz and 4 GB of RAM.

The classification algorithms utilized in our study were configured with base settings to achieve optimal performance. The base settings of these algorithms were carefully selected and fine-tuned to balance model complexity and predictive performance. Hyperparameters were optimized using techniques like grid search and random search. For MLR, the base settings included Cs (regularization strength) set to 10, cv (cross-validation) set to 5, and multi-class set to “multinomial.” For KNN, the base settings included n_neighbors set to 5, weights set to “uniform,” and algorithm set to “auto,” with optimization focused on the number of neighbors.

LDA used the default settings provided by the “svd” (Singular Value Decomposition) solver, while QDA utilized a regularization parameter of 0.0 and did not require specific hyperparameter optimization. Decision Trees were optimized for parameters like max_depth (up to 20), cv set to 5, and n_jobs set to 2 respectively. Furthermore, for Random Forest, the settings consisted of n_estimators set to 50, max_depth set to 20, cv set to 5, and n_jobs set to 4, with optimization performed for the number of estimators and maximum depth. AdaBoost was optimized for parameters like max_depth (up to 20) and learning_rate (0.05, 0.1, 0.5), with a base estimator as Decision Tree.

Utilizing Cs set to 10, cv set to 5, and multi_class set to “ovr” as base settings for logistic regression, with default settings used for PCA for hyperparameter optimization in the PC-LR algorithm. SVM-RBF used parameters like C (ranging from 0.001 to 1,000), gamma (values ranging from 0 to 1), kernel set to “rbf,” and decision_function_shape set to “ovr” for optimization. Perceptron was optimized for the alpha parameter (ranging from 0.0001 to 10,000) and the penalty set to “l2,” with optimization centered around the alpha parameter. Multi-layer Perceptron employed settings for parameters such as hidden_layer_sizes set to 100, activation function set to “relu,” and learning_rate set to “constant.” Elastic Nets were configured with base settings such as l1_ratio (ranging from 0.1 to 1) and alphas (values ranging from 0 to 10). Also, pre-processing steps were applied uniformly to ensure consistency in model training and evaluation.

3 Results

This section presents the findings of the hybrid-clinical ML modeling for each of the individual set controllers. Initially, the base models are discussed, followed by the meta-model.

3.1 For baseline (C = 1)

Table 6 presents the modeling results for the base ML models. These results are for all three approaches to handling missing data, as discussed in Section 2.3.1. The feature selection process identified the following features to be included: Age, Education, Gender, CDR_SB, ADAS13, ADASQ4, MMSE, RAVLT-I, RAVLT-L, RAVLT-PF, TRABSCOR, LDETOTAL, FAQ, mPACCdigit, Ventricles, WholeBrain, Hippocampus, Entorhinal, Fusiform, ICV, MidTemp, Ethnicity, Race category, Married status, and APOE4. Moreover, Table 7 summarizes the outcome of the constructed Meta-Model as well as the metrics used to evaluate performance.

Table 6
www.frontiersin.org

Table 6. Performance results on Base models (C=1).

Table 7
www.frontiersin.org

Table 7. Performance result: meta-model (C = 1).

As shown in Table 6, the AdaBoost algorithm without imputation, Random Forest with mean imputation, Logistic Regression (multinomial), and Random Forest with model imputation techniques all achieved a higher accuracy of 95.20%, 95.26%, and 95.30% on the test set, respectively. Subsequently, as a Meta-Model (Table 7), the improved accuracy was demonstrated using a no-imputation approach with a 95.78% accuracy on the test set. On the test set, mean imputation achieved 95.70% accuracy, whereas model imputation achieved 95.73% accuracy. Additionally, the performance of the Meta-Model with the no-imputation technique was determined to be the best among all, as it gave an accuracy of ~96.0%, a precision value of ~96.0%, a recall value of ~96.0%, and F1-score of 0.96, which was close to 1.0. A high F-measure, close to 1.0, is considered the best measure. Furthermore, the classification error i.e., Hamming Loss, gave a value of 0.04, which was close to 0, considered the best value. Finally, the Jaccard Index value of 0.92, near 1.0, suggested that the created Meta-Model reflected the best classification.

3.2 For baseline + drug (C = 2)

Table 8 shows the modeling results for the base ML models. The feature selection process identified the following features to be included for controller 2 (baseline + drug): Age, Education, Gender, CDR-SB, MMSE, RAVLT-I, RAVLT-L, RAVLT-F, LDETOTAL, TRABSCOR, DIGITSCOR, Ventricles, WholeBrain, Hippocampus, Fusiform, MidTemp, Ethnicity, Race category, Married status, Blood thinner, Calcium, Cholesterol, Cognitive, and Vitamin D. Additionally, Table 9 presents the findings of the built Meta-Model as well as the metrics used to evaluate performance.

Table 8
www.frontiersin.org

Table 8. Outcome-based on base models (C = 2).

Table 9
www.frontiersin.org

Table 9. Result: meta-model (C = 2).

As shown in Table 8, the AdaBoost algorithm produced an accuracy of 94.00% on the test set when used without imputation, 96.18% with mean imputation, and 95.00% with model imputation. Following that, using the no-imputation technique with a 96.40% accuracy on the test set, the improved accuracy was demonstrated using a Meta-Model. On the test set, an improved accuracy of 96.40% was achieved for all three techniques. When compared to the base model, it was established that the performance of the Meta-Model was way better. Other performance metrics including precision, recall, and F-score demonstrated a performance improvement. Hamming loss was calculated and yielded a value of 0.06, 0.05, and 0.04 respectively, which are regarded to be as acceptable. Additionally, the Jaccard Index values of 0.89, 0.89, and 0.91, which were all near 1.0, indicated that the Meta-Model built was accurate.

Following that, we present the results for all of the features above (excluding the drug list), and only one drug at a time was examined. Table 10 summarizes the modeling results for base models, whereas Table 11 summarizes the Meta-Model results when just one particular drug was considered in addition to baseline characteristics.

Table 10
www.frontiersin.org

Table 10. Base model: baseline features + individual drug.

Table 11
www.frontiersin.org

Table 11. Meta-model: baseline features + individual drug.

As shown in Table 10, all of the algorithms studied had an accuracy between 94.00% and 96.00%. The following conclusions were drawn: Comparing Tables 8, 10 results for Base Models revealed that when only the cognitive drug was considered, the no-imputation technique (96.00%) using the AdaBoost classifier produced better accurate results. If just cholesterol-lowering supplements and baseline characteristics were considered, the accuracy of the imputation with the mean technique was 96.02%. When compared to Table 8, which included the performance results for the selected baseline features and the five medications, it was discovered that the AdaBoost classifier had an accuracy of 96.18% when using the Linear Regression imputation approach. If just blood thinners or cholesterol-lowering supplements were considered in addition to baseline characteristics, the accuracy for imputation with mean and Linear Regression techniques was 95.80% and 95.80%, respectively. When compared to Table 8, the accuracy of the AdaBoost classifier for the model imputation approach was 95.00%.

Following that, Table 11 summarizes the Meta-Model developed for all of the baseline characteristics assessed, as well as for a single drug. It revealed that the performance accuracy of the five Meta-Models was close to 96.0% in three of the scenarios (Table 11). The precision, recall, and F1-score values, as well as the classification errors, were all highly commendable as well. This indicated that the created model performed well for both if all drugs were taken into consideration or if any single drug was included for the diagnosis of the response variable as CN, AD, or MCI, which was a positive factor. When compared to the results in Table 9, this indicated that the Meta-Model yielded rather similar outcomes regardless of whether a specific drug was employed or all five medications were included. No one medication had a discernible effect on the diagnosis of AD, CN, or MCI. Each of them produced an identical result.

3.3 For baseline + protein (C = 3)

The features that were selected post-feature selection mechanism are Age, Education, Gender, CDR-SB, ADAS13, MMSE, RAVLT-L, LDETOTAL, DIGITSCOR, TRABSCOR, mPACCdigit, Fusiform, MidTemp, Ethnicity, Race category, Married status, Aβ, Tau, and PTau. To categorize the response variables CN, MCI, and AD based on these features, ML modeling was performed. Table 12 presents the results for the base ML models. In addition, Table 13 highlights the findings of the Meta-Model that was developed as a result of this process.

Table 12
www.frontiersin.org

Table 12. Performance result on base model for C = 3.

Table 13
www.frontiersin.org

Table 13. Result: meta-model (C = 3).

Table 12 exhibited that when the AdaBoost technique was employed without imputation on the test set, it delivered a high accuracy of 96.43%. When compared to the Meta-Model, it achieved an accuracy of 97.60% for the same applied approach. The accuracy of the Base Model was 86.0% with mean imputation and 87.30% with model imputation techniques, respectively. Following that, using an imputation with mean approach with an accuracy of 88.40% and an imputation with Linear Regression technique with an accuracy of 89.00%, the enhanced accuracy was demonstrated using a Meta-Model. As shown in Table 13, the results achieved so far employing baseline characteristics and protein biomarkers are commendable on other measures. As a result of this, we present the results for a reduced collection of baseline characteristics such as education and gender, CDR-SSB, ADASQ4, RAVLT-I, RAVLT-L, RAVLT-PF, mPACCdigit, and mPACCtrailsB, along with only one protein biomarker (at a time). When only one specific protein biomarker (Aβ/Tau/Ptau) was included in addition to the aforesaid baseline characteristics, the results of the modeling are summarized in Table 14, whilst the findings of the Meta-Modeling are summarized in Table 15.

Table 14
www.frontiersin.org

Table 14. Base model: baseline features + individual protein biomarker.

Table 15
www.frontiersin.org

Table 15. Meta-model: baseline features + individual protein biomarker.

Comparing Tables 12, 14 results for Base Models demonstrated that when only the Tau protein biomarker was evaluated, the no-imputation strategy offered more accurate findings with a 95.40% accuracy using the Decision Trees classifier. When only the Aβ/Tau/PTau protein biomarker and baseline parameters were included, the imputation with mean approach achieved an accuracy of 88.00% (Table 14). When compared to Table 12, which contained performance results for the specified baseline features and the three protein indicators, it was determined that the Logistic Regression (Multinomial) classifier had an accuracy of 86.00% when employing a similar imputation strategy. When only the Aβ protein biomarker was included in addition to baseline features, the accuracy of the model imputation technique on the AdaBoost classifier was 88.40%, respectively. When compared to Table 14, the Logistic Regression (Multinomial) classifier had an accuracy of 87.30% for the Linear Regression imputation strategy.

Table 15 presents the Meta-Model built for all baseline features along with a single protein biomarker. It demonstrated that the three meta-models' performance accuracy ranged from 87.00% to 96.40% in three of the cases (Table 15). Additionally, the precision, recall, F1-score, and classification errors were all noteworthy. This suggested that the developed model performed efficiently on both employed techniques i.e., when all protein biomarkers were included and when a single protein biomarker was used to diagnose the response variable as CN, AD, or MCI, which was a significant indicator. When compared to the results in Table 13, this implied that the Meta-Model resulted in a modest increase in outcomes when Aβ, Tau, and PTau were taken into account in the mean imputation technique. However, the Meta-Model (Table 13) achieved an accuracy of 97.60% for the no-imputation technique, which was greater than the no-imputation method in Table 15 for all Aβ/Tau/PTau protein biomarkers.

3.4 For baseline + medication + protein (C = 4)

The selected features after applying the feature selection technique for controller 4 included: Education, Gender, CDR-SB, ADAS13, ADASQ4, MMSE, RAVLT-I, RAVLT-L, RAVLT-PF, DIGITSCOR, LDETOTAL, mPACCdigit, Ventricles, Entorhinal, Aβ, Tau, PTau, Ethnicity, Race category, Married status, APOE4, Blood thinner, Calcium, Cholesterol, Cognitive, and Vitamin D. The results of the modeling for the base ML models are summarized in Table 16. Table 17 presents the results of the Meta-Model construction process as well as the metrics used to evaluate performance.

Table 16
www.frontiersin.org

Table 16. Performance result on base model for C = 4.

Table 17
www.frontiersin.org

Table 17. Result: meta-model (C = 4).

As demonstrated in Table 16, when employed without imputation on the test set, the Logistic Regression (Multinomial) algorithm produced an accuracy of 95.20%. In comparison to the Meta-Model, it attained a high accuracy of 96.20% when the identical approach was used. When the mean imputation approach was applied, the base model's accuracy was 87.50%. Following that, an imputation with a mean technique with an accuracy of 88.40% was used to demonstrate the enhanced accuracy for a Meta-Model. Additionally, model imputation yielded an accuracy of 88.40% on the Logistic Regression (Multinomial) classifier for a Base Model, while the Meta-Model yielded an accuracy of 89.30% for the same applied approach. As shown in Table 17, the findings thus far based on baseline characteristics, medication, and protein biomarkers were notable in terms of additional performance evaluation criteria.

3.5 Comparative analysis

Figure 6 presents a comparative analysis of all four developed Meta-Models based on performance accuracy. We can comprehend the following from Figure 6: the highest accuracy of 97.60% was shown by no imputation approach when the controller C = 3 (baseline + protein); for controller C = 4 (baseline + drug + protein), the no imputation approach gave an improved accuracy of 96.20% amongst the three employed techniques; for C = 2 (baseline + drug), the accuracy of 96.40% for all the three applied approaches was observed; and when C = 1 (baseline data), the accuracy for all the three methods was found to be close to 96.0% respectively.

Figure 6
www.frontiersin.org

Figure 6. Comparative analysis of meta-models.

4 Discussion

The timely and precise diagnosis of AD is critical to reducing the consequences of this disorder, whose incidence has been on the rise. The application of ML can enhance the clinical diagnosis of AD in its early stages and provide valuable insights into research on this damaging and progressive disease. Early Alzheimer's disease was diagnosed in this study by considering characteristics associated with specific AD medications such as calcium, vitamin D supplements, blood thinner medicines, cholesterol-lowering drugs, and cognitive drugs, including a substantial protein biomarker (Aβ, tau, and ptau) as a predictor. To simulate the MRI-based data, we proposed a hybrid-clinical model. The diagnostic group in this study comprised five categories. We designed a pipeline that integrated exhaustive approaches for detecting AD across a broad range of input values and parameters. We evaluated the association between the diagnosis of AD and the use of several drugs in particular. Aiming to better understand how Alzheimer's is diagnosed, we looked at the importance of three cerebrospinal fluid biomarkers: tau, ptau, and Aβ. The proposed design generated four Meta-Models for four different sets of criteria. The diagnostic criteria were established based on baseline features, baseline and drug characteristics, baseline and protein features, and baseline, drug, and protein features. The developed model incorporated a range of methodologies, including data collection and integration, multi-step data pre-processing, feature selection, development of machine learning models employing wide-ranging methods, optimization and analysis of results, measuring of test performance, construction of Meta-Model, generalized performance metrics, and comparative analysis.

Our results indicate that patients' age, education, gender, CDR_SB, ADAS13, ADASQ4, MMSE, RAVLT-I, RAVLT-L, RAVLT-PF, TRABSCOR, LDETOTAL, mPACCdigit, Ventricles, WholeBrain, Hippocampus, Entorhinal, Fusiform, MidTemp, ICV, ethnicity, race category, marital status, and APOE4 are associated with a higher likelihood of being diagnosed with AD based on our analysis of the baseline visit data. In addition, our results reveal that CSF biomarkers, tau, ptau, and Aβ, when added to the baseline model, could be the significant predictors. We were able to attain a maximum accuracy of 97.60% for baseline and protein data without any imputation technique. We observed that the constructed model functioned effectively when all five drugs were included, as well as when any single drug was used for the diagnosis of the response variable (CN, AD, or MCI). Interestingly, the constructed Meta-Model worked well when all three protein biomarkers were included, as well as when a single protein biomarker was utilized to diagnose the response variable. Thus, our developed model not only has the potential to aid clinicians and medical professionals in advancing Alzheimer's diagnosis but also serves as a valuable starting point for future research into AD and other neurodegenerative disorders. With further refinement and exploration, it could pave the way for innovative diagnostic techniques and therapeutic interventions in the field of neurology.

Researchers are increasingly employing machine learning and deep learning techniques in their work to classify and evaluate patients and the associated risks and predict treatment outcomes. One area where these methods have been particularly useful is in the classification of neurodegenerative conditions caused by AD, as well as their different stages, using imaging-based detection. Additionally, researchers have constructed automated pipelines that employ feature extraction approaches based on a range of biomarker methodologies to improve the quality of their findings (Shukla et al., 2023). ML models have become ubiquitous in real-time clinical applications, diagnostics, and the treatment of AD. Numerous recent studies (as reported in Table 18) have integrated MRI data into ML models for predicting AD as evidenced by the works of Gopi et al. (2020), Khan and Zubair (2020), Liu et al. (2020), Diogo et al. (2022), Kavitha et al. (2022), Khan and Zubair (2022b), and Uddin et al. (2023).

Table 18
www.frontiersin.org

Table 18. Comparison with the state-of-the-art approaches.

Ensuring optimal accuracy for cognitive assessments in the context of AD continues to be a pressing challenge despite ongoing efforts. To bridge this gap, a novel clinical-hybrid model has been introduced to enhance the accuracy of Alzheimer's detection. To demonstrate the improvements and potential contributions of our new model in making cognitive tests more accurate for individuals with AD, Table 18 presents a comparison between our proposed method and prior research. However, it is essential to highlight that the datasets, the number of patients in each research, the classifiers used, and the modeling technique are all highly distinct, making direct comparison challenging.

Recent years have witnessed significant advances in research, most notably the discovery of biomarkers (especially brain imaging technologies) that enable the diagnosis and monitoring of AD-related processes months, years, and even decades before clinical problems appear. Alzheimer's disease biomarkers are divided into two types: early biomarkers, which measure amyloid accumulation in the brain (e.g., PET imaging, CSF amyloid), and late biomarkers, which measure neurodegeneration [e.g., structural MRI, fluorodeoxyglucose-positron emission tomography (FDG PET), CSF tau]. Few recent studies have found that AD biomarkers are associated with cognitive decline including Stamate et al. (2019) and Gaetani et al. (2021), as presented in Table 18.

Biomarkers are widely exploited in the diagnostic framework and help in designing appropriate therapy (when available), albeit these are largely intended for research use. There are two types of biomarkers: those that directly influence the pathology of AD, like the A beta-amyloid (Aβ) and tau proteins; and those that provide an indirect or non-specific indication of the disorder by locating indices of neuronal damage, which are considered to be the main cause of AD. While these indicators are generally related to Alzheimer's disease, they have been related to other types of illnesses as well. The concomitant presence of both proteins in an individual indicates a strong case of AD.

In the present circumstances, both advanced-stage AD and dementia are considered incurable. It could be attributed to the failure in the early stage of the disease. Currently, the goal of treatment is to reduce the course of the disease and also to control its symptoms. While this is extremely difficult, it is achievable to some extent if the disease is detected reasonably early. Treatments focus on symptom management, such as cognitive and psychological difficulties, as well as behavioral difficulties; environmental alteration to enable patients to do everyday activities more effectively; and caregiver support, such as family and friends.

The CSF analysis also provides information regarding blood-brain barrier damage and inflammatory diseases that resemble or make a contribution to dementia (Blennow et al., 2010). Nonetheless, more recently, approaches used in determining levels of biomarkers such as Aβ1-42, pTau, and total tau (tTau) have been established in investigating essential Alzheimer's disease pathology. The levels of Aβ1-42 are decreased in AD, although tTau (a more general sign of neuronal degeneration) and pTau (a more specific diagnostic for AD) have been found to get elevated during the progression of AD. Current research suggests that a low Aβ1-42 value combined with a high pTau or tTau value provides the most diagnostic specificity (Hansson et al., 2006). In the absence of established pathology, it is difficult to define reference ranges as well as cut points for these parameters. A lack of uniformity in CSF gathering and processing has been reported to make it more difficult (Mattsson et al., 2009; Bartlett et al., 2012). However, efforts are being made to resolve these concerns, and CSF analysis has been inducted into Alzheimer's disease research diagnostic guidelines (Dubois et al., 2014).

Moreover, the imminent application of plasma biomarkers, including plasma Aβ and tau, holds promise for enhancing AD diagnosis (Sun et al., 2022). With advancements in research and the potential approval of disease-modifying therapies targeting MCI-AD or AD-dementia, the imperative for earlier and more accurate diagnosis of AD becomes increasingly critical (Cummings, 2019; Arafah et al., 2023). The advent of disease-modifying drugs highlights the importance of identifying individuals at risk of developing AD in preclinical stages, paving the way for proactive interventions aimed at delaying or preventing disease progression (Crous-Bou et al., 2017).

5 Conclusion

This study presents a notable improvement in early AD detection through the integration of machine learning, statistical modeling, and biomarker indicators. The optimized predictive models demonstrate a robust diagnostic framework that takes into account various factors, including patient drugs, protein biomarkers, and baseline features. The proposed hybrid-clinical model and the in-depth analysis of the correlations between demographic, clinical, and biomarker variables and AD diagnosis underscore its potential to revolutionize clinical detection. The high level of accuracy obtained with both baseline and protein data serves as validation for the efficacy of the developed models. The comprehensive pipeline and Meta-Models designed for various diagnostic criteria, offer a versatile approach for clinicians. Additionally, the comparative analysis with existing studies highlights the novel contributions of this research while acknowledging the challenges in direct comparisons due to variations in datasets and methodologies. Overall, this study not only provides valuable insights into Alzheimer's diagnosis but also sets a basis for future research into neurodegenerative disorders, emphasizing the essential role of advanced technologies in transforming diagnostic approaches. It could be hypothesized that incorporating patients' drugs or protein biomarkers into an ML model to supplement clinical diagnoses could enhance diagnostic accuracy. Our improved model with drug features could mitigate the challenge of diagnosing the early stages of AD when other symptoms are not as easily observable, which remains a pervasive issue for both clinical and research professionals.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

AK: Conceptualization, Formal analysis, Writing – original draft. SZ: Investigation, Methodology, Supervision, Validation, Writing – review & editing. MS: Formal analysis, Validation, Visualization, Writing – review & editing. AS: Conceptualization, Funding acquisition, Investigation, Supervision, Validation, Writing – original draft. SA: Project administration, Supervision, Visualization, Writing – review & editing. BA: Investigation, Project administration, Validation, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. The authors gratefully acknowledge the funding of the Deanship of Graduate Studies and Scientific Research, Jazan University, Saudi Arabia through Project number GSSRD-24.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

1. ^ADNI | About. Available at: http://adni.loni.usc.edu/about/.

2. ^Alzheimer's Dementia Causes, Risk Factors | Research Center. Alzheimer's Association. Available at: https://www.alz.org/alzheimers-dementia/what-is-alzheimers/causes-and-risk-factors.

References

Aboneh, T., Rorissa, A., and Srinivasagan, R. (2022). Stacking-based ensemble learning method for multi-spectral image classification. Technologies 10:17. doi: 10.3390/technologies10010017

Crossref Full Text | Google Scholar

Aguilera, A. M., Escabias, M., and Valderrama, M. J. (2006). Using principal components for estimating logistic regression with high-dimensional multicollinear data. Comput. Stat. Data Anal. 50, 1905–1924. doi: 10.1016/j.csda.2005.03.011

Crossref Full Text | Google Scholar

Ahmed, Z., Mohamed, K., Zeeshan, S., and Dong, X. (2020). Artificial intelligence with multi-functional machine learning platform development for better healthcare and precision medicine. Database 2020:baaa010. doi: 10.1093/database/baaa010

PubMed Abstract | Crossref Full Text | Google Scholar

Alowais, S. A., Alghamdi, S. S., Alsuhebany, N., Alqahtani, T., Alshaya, A. I., Almohareb, S. N., et al. (2023). Revolutionizing healthcare: the role of artificial intelligence in clinical practice. BMC Med. Educ. 23:689. doi: 10.1186/s12909-023-04698-z

PubMed Abstract | Crossref Full Text | Google Scholar

Alqahtani, N., Alam, S., Aqeel, I., Shuaib, M., Mohsen Khormi, I., Khan, S. B., et al. (2023). Deep belief networks (DBN) with IoT-based alzheimer's disease detection and classification. Appl. Sci. 13:7833. doi: 10.3390/app13137833

Crossref Full Text | Google Scholar

Aqeel, I., Khormi, I. M., Khan, S. B., Shuaib, M., Almusharraf, A., Alam, S., et al. (2023). Load balancing using artificial intelligence for cloud-enabled internet of everything in healthcare domain. Sensors 23:5349. doi: 10.3390/s23115349

PubMed Abstract | Crossref Full Text | Google Scholar

Arafah, A., Khatoon, S., Rasool, I., Khan, A., Rather, M. A., Abujabal, K. A., et al. (2023). The future of precision medicine in the cure of Alzheimer's disease. Biomedicines 11:335. doi: 10.3390/biomedicines11020335

PubMed Abstract | Crossref Full Text | Google Scholar

Assiri, B., and Hossain, M. A. (2023). Face emotion recognition based on infrared thermal imagery by applying machine learning and parallelism. Mathem. Biosci. Eng. 20, 913–929. doi: 10.3934/mbe.2023042

PubMed Abstract | Crossref Full Text | Google Scholar

Barnes, H., Humphries, S. M., George, P. M., Assayag, D., Glaspole, I., Mackintosh, J. A., et al. (2023). Machine learning in radiology: the new frontier in interstitial lung diseases. Lancet Digital Health 5, e41–e50. doi: 10.1016/S2589-7500(22)00230-8

PubMed Abstract | Crossref Full Text | Google Scholar

Bartlett, J. W., Frost, C., Mattsson, N., Skillbäck, T., Blennow, K., Zetterberg, H., et al. (2012). Determining cut-points for Alzheimer's disease biomarkers: statistical issues, methods and challenges. Biomark. Med. 6, 391–400. doi: 10.2217/bmm.12.49

PubMed Abstract | Crossref Full Text | Google Scholar

Belle, V., and Papantonis, I. (2021). Principles and practice of explainable machine learning. Front. Big Data 4:688969. doi: 10.3389/fdata.2021.688969

PubMed Abstract | Crossref Full Text | Google Scholar

Better, M. A. (2023). Alzheimer's disease facts and figures. Alzheimer's Dement. 19, 1598–1695. doi: 10.1002/alz.13016

PubMed Abstract | Crossref Full Text | Google Scholar

Bhatia, S., Alam, S., Shuaib, M., Hameed Alhameed, M., Jeribi, F., Alsuwailem, R. I., et al. (2022). Retinal vessel extraction via assisted multi-channel feature map and U-net. Front. Public Health 10:858327. doi: 10.3389/fpubh.2022.858327

PubMed Abstract | Crossref Full Text | Google Scholar

Blennow, K., Hampel, H., Weiner, M., and Zetterberg, H. (2010). Cerebrospinal fluid and plasma biomarkers in Alzheimer disease. Nat. Rev. Neurol. 6, 131–144. doi: 10.1038/nrneurol.2010.4

PubMed Abstract | Crossref Full Text | Google Scholar

Blockeel, H., Devos, L., Frénay, B., Nanfack, G., and Nijssen, S. (2023). Decision trees: from efficient prediction to responsible AI. Front. Artif. Intell. 6:1124553. doi: 10.3389/frai.2023.1124553

PubMed Abstract | Crossref Full Text | Google Scholar

Brookmeyer, R., Johnson, E., Ziegler-Graham, K., and Arrighi, H. M. (2007). Forecasting the global burden of Alzheimer's disease. Alzheimer's Dement. 3, 186–191. doi: 10.1016/j.jalz.2007.04.381

PubMed Abstract | Crossref Full Text | Google Scholar

Campagner, A., Ciucci, D., and Cabitza, F. (2023). Aggregation models in ensemble learning: a large-scale comparison. Inf. Fusion 90, 241–252. doi: 10.1016/j.inffus.2022.09.015

Crossref Full Text | Google Scholar

Chaudhuri, B. B., and Bhattacharya, U. (2000). Efficient training and improved performance of multilayer perceptron in pattern classification. Neurocomputing 34, 11–27. doi: 10.1016/S0925-2312(00)00305-2

Crossref Full Text | Google Scholar

Chawla, N. V., Bowyer, K. W., Hall, L. O., and Kegelmeyer, W. P. (2022). Smote: Synthetic minority over-sampling technique. doi: 10.1613/jair.953

Crossref Full Text | Google Scholar

Chen, S. B., Zhang, Y., Ding, C. H., Zhou, Z. L., and Luo, B. (2018). A discriminative multi-class feature selection method via weighted l2, 1-norm and extended elastic net. Neurocomputing 275, 1140–1149. doi: 10.1016/j.neucom.2017.09.055

Crossref Full Text | Google Scholar

Ciurea, A. V., Mohan, A. G., Covache-Busuioc, R. A., Costin, H. P., Glavan, L. A., Corlatescu, A. D., et al. (2023). Unraveling molecular and genetic insights into neurodegenerative diseases: advances in understanding Alzheimer's, Parkinson's, and Huntington's diseases and amyotrophic lateral sclerosis. Int. J. Mol. Sci. 24:10809. doi: 10.3390/ijms241310809

PubMed Abstract | Crossref Full Text | Google Scholar

Costa, V. G., and Pedreira, C. E. (2023). Recent advances in decision trees: an updated survey. Artif. Intell. Rev. 56, 4765–4800. doi: 10.1007/s10462-022-10275-5

Crossref Full Text | Google Scholar

Crous-Bou, M., Minguillón, C., Gramunt, N., and Molinuevo, J. L. (2017). Alzheimer's disease prevention: from risk factors to early intervention. Alzheimer's Res. Ther. 9, 1–9. doi: 10.1186/s13195-017-0297-z

PubMed Abstract | Crossref Full Text | Google Scholar

Cummings, J. (2019). The role of biomarkers in alzheimer's disease drug development. Adv. Exp. Med. Biol. 1118, 29–61. doi: 10.1007/978-3-030-05542-4_2

PubMed Abstract | Crossref Full Text | Google Scholar

Ding, X., Liu, J., Yang, F., and Cao, J. (2021). Random radial basis function kernel-based support vector machine. J. Franklin Inst. 358, 10121–10140. doi: 10.1016/j.jfranklin.2021.10.005

Crossref Full Text | Google Scholar

Ding, Y., Zhu, H., Chen, R., and An, L. i. R. (2022). efficient adaboost algorithm with the multiple thresholds classification. Appl. Sci. 12:5872. doi: 10.3390/app12125872

Crossref Full Text | Google Scholar

Diogo, V. S., Ferreira, H. A., Prata, D., and Alzheimer's Disease Neuroimaging Initiative (2022). Early diagnosis of Alzheimer's disease using machine learning: a multi-diagnostic, generalizable approach. Alzheimers. Res. Ther. 14:107. doi: 10.1186/s13195-022-01047-y

PubMed Abstract | Crossref Full Text | Google Scholar

Dokeroglu, T., Deniz, A., and Kiziloz, H. E. (2022). A comprehensive survey on recent metaheuristics for feature selection. Neurocomputing 494, 269–296. doi: 10.1016/j.neucom.2022.04.083

Crossref Full Text | Google Scholar

Dolo, K. M., and Mnkandla, E. (2023). “Weighted voting stacking ensemble method for highly skewed binary data distribution,” in International Conference on Wireless Intelligent and Distributed Environment for Communication (Cham: Springer International Publishing), 107–120. doi: 10.1007/978-3-031-33242-5_8

Crossref Full Text | Google Scholar

Dubois, B., Feldman, H. H., Jacova, C., Hampel, H., Molinuevo, J. L., Blennow, K., et al. (2014). Advancing research diagnostic criteria for Alzheimer's disease: the IWG-2 criteria. Lancet Neurol. 13, 614–629. doi: 10.1016/S1474-4422(14)70090-0

PubMed Abstract | Crossref Full Text | Google Scholar

Feigin, V. L., Vos, T., Nichols, E., Owolabi, M. O., Carroll, W. M., Dichgans, M., et al. (2020). The global burden of neurological disorders: translating evidence into policy. Lancet Neurol. 19, 255–265. doi: 10.1016/S1474-4422(19)30411-9

PubMed Abstract | Crossref Full Text | Google Scholar

Gaetani, L., Bellomo, G., Parnetti, L., Blennow, K., Zetterberg, H., Filippo, D., et al. (2021). M. Neuroinflammation and Alzheimer's disease: a machine learning approach to CSF proteomics. Cells 10:1930. doi: 10.3390/cells10081930

PubMed Abstract | Crossref Full Text | Google Scholar

Gauthier, S., Rosa-Neto, P., Morais, J. A., and Webster, C. (2021). World Alzheimer Report 2021: Journey through the diagnosis of dementia. Alzheimer's Dis. Int. 2022:30. doi: 10.1016/j.jns.2023.121394

PubMed Abstract | Crossref Full Text | Google Scholar

Gopi, B., Nalini, C., and Francesco, A. (2020). Late-life Alzheimer's disease (AD) detection using pruned decision trees. Int. J. Brain Disord. Treat. 6:033. doi: 10.23937/2469-5866/1410033

PubMed Abstract | Crossref Full Text | Google Scholar

Graf, R., Zeldovich, M., and Friedrich, S. (2024). Comparing linear discriminant analysis and supervised learning algorithms for binary classification—A method comparison study. Biom. J. 66:2200098. doi: 10.1002/bimj.202200098

PubMed Abstract | Crossref Full Text | Google Scholar

Haixiang, G., Yijing, L., Yanan, L., Xiao, L., and Jinling, L. (2016). BPSO-Adaboost-KNN ensemble learning algorithm for multi-class imbalanced data classification. Eng. Appl. Artif. Intell. 49, 176–193. doi: 10.1016/j.engappai.2015.09.011

Crossref Full Text | Google Scholar

Hansson, O., Zetterberg, H., Buchhave, P., Londos, E., Blennow, K., Minthon, L., et al. (2006). Association between CSF biomarkers and incipient Alzheimer's disease in patients with mild cognitive impairment: a follow-up study. Lancet Neurol. 5, 228–234. doi: 10.1016/S1474-4422(06)70355-6

PubMed Abstract | Crossref Full Text | Google Scholar

Hedeker, D. (2003). A mixed-effects multinomial logistic regression model. Stat. Med. 22, 1433–1446. doi: 10.1002/sim.1522

PubMed Abstract | Crossref Full Text | Google Scholar

Hossain, M. A., and Assiri, B. (2020). “Emotion specific human face authentication based on infrared thermal image,” in 2020 2nd International Conference on Computer and Information Sciences (ICCIS) (IEEE), 1–6. doi: 10.1109/ICCIS49240.2020.9257683

Crossref Full Text | Google Scholar

Jaul, E., and Barron, J. (2017). Age-related diseases and clinical and public health implications for the 85 years old and over population. Front. Public Health 5:335. doi: 10.3389/fpubh.2017.00335

PubMed Abstract | Crossref Full Text | Google Scholar

Javaid, M., Haleem, A., Pratap Singh, R., Suman, R., and Rab, S. (2022). Significance of machine learning in healthcare: Features, pillars and applications. Int. J. Intell. Netw. 3, 58–73. doi: 10.1016/j.ijin.2022.05.002

Crossref Full Text | Google Scholar

Jiang, B., Wang, X., and Leng, C. A. (2018). Direct approach for sparse quadratic discriminant analysis internet. J. Mach. Learn. Res. 19:285. doi: 10.5555/3291125.3291156

Crossref Full Text | Google Scholar

Jiang, F., Jiang, Y., Zhi, H., Dong, Y., Li, H., Ma, S., et al. (2017). Artificial intelligence in healthcare: past, present and future. Stroke Vasc. Neurol. 2, 230–243. doi: 10.1136/svn-2017-000101

PubMed Abstract | Crossref Full Text | Google Scholar

Kanyongo, W., and Ezugwu, A. E. (2023). Feature selection and importance of predictors of non-communicable diseases medication adherence from machine learning research perspectives. Inform. Med. Unlocked 38:101232. doi: 10.1016/j.imu.2023.101232

Crossref Full Text | Google Scholar

Kavitha, C., Mani, V., Srividhya, S. R., Khalaf, O. I., and Tavera Romero, C. A. (2022). Early-stage Alzheimer's disease prediction using machine learning models. Front. Public Health 10:853294. doi: 10.3389/fpubh.2022.853294

PubMed Abstract | Crossref Full Text | Google Scholar

Khan, A., and Zubair, S. (2019). Usage of random forest ensemble classifier based imputation and its potential in the diagnosis of Alzheimer's disease. Int. J. Sci. Technol. Res 8, 271–275.

Google Scholar

Khan, A., and Zubair, S. (2020). “A Machine Learning-based robust approach to identify Dementia progression employing Dimensionality Reduction in Cross-Sectional MRI data,” in 2020 First International Conference of Smart Systems and Emerging Technologies (SMARTTECH) (Riyadh, Saudi Arabia), 237–242. doi: 10.1109/SMART-TECH49988.2020.00060

Crossref Full Text | Google Scholar

Khan, A., and Zubair, S. (2022a). An improved multi-modal based machine learning approach for the prognosis of Alzheimer's disease. J. King Saud Univ. Comput. Inf. Sci. 34, 2688–2706. doi: 10.1016/j.jksuci.2020.04.004

Crossref Full Text | Google Scholar

Khan, A., and Zubair, S. (2022b). Development of a three tiered cognitive hybrid machine learning algorithm for effective diagnosis of Alzheimer's disease. J. King Saud Univ. Comput. Inf. Sci. 34, 8000–8018. doi: 10.1016/j.jksuci.2022.07.016

Crossref Full Text | Google Scholar

Khan, R., Akbar, S., Mehmood, A., Shahid, F., Munir, K., Ilyas, N., et al. (2023). A transfer learning approach for multiclass classification of Alzheimer's disease using MRI images. Front. Neurosci. 16:1050777. doi: 10.3389/fnins.2022.1050777

PubMed Abstract | Crossref Full Text | Google Scholar

Khan, W., Ishrat, M., Khan, A. N., Arif, M., Shaikh, A. A., Khubrani, M. M., et al. (2024). Detecting anomalies in attributed networks through sparse canonical correlation analysis combined with random masking and padding. IEEE Access. 12, 65555–65569. doi: 10.1109/ACCESS.2024.3398555

Crossref Full Text | Google Scholar

Kleyko, D., Rosato, A., Frady, E. P., Panella, M., and Sommer, F. T. (2023). Perceptron theory can predict the accuracy of neural networks. IEEE Trans. Neural Netw. Learn. Syst. 35, 9885–9899. doi: 10.1109/TNNLS.2023.3237381

PubMed Abstract | Crossref Full Text | Google Scholar

Kohavi, R. (1995). “A study of cross-validation and bootstrap for accuracy estimation and model selection,” in International Joint Conference on Artificial Intelligence (IJCAI), 1137–1143.

Google Scholar

Li, H. (2024). “K-Nearest Neighbor,” in Machine Learning Methods (Springer, Singapore). doi: 10.1007/978-981-99-3917-6_3

Crossref Full Text | Google Scholar

Liu, L., Zhao, S., Chen, H., and Wang, A. (2020). A new machine learning method for identifying Alzheimer's disease. Simul. Model. Pract. Theory 99:102023. doi: 10.1016/j.simpat.2019.102023

Crossref Full Text | Google Scholar

Liu, M., Li, L., Wang, H., Guo, X., Liu, Y., Li, Y., et al. (2023). A multilayer perceptron-based model applied to histopathology image classification of lung adenocarcinoma subtypes. Front. Oncol. 13:1172234. doi: 10.3389/fonc.2023.1172234

PubMed Abstract | Crossref Full Text | Google Scholar

Maharana, K., Mondal, S., and Nemade, B. (2022). A review: data pre-processing and data augmentation techniques. Global Trans. Proc. 3, 91–99. doi: 10.1016/j.gltp.2022.04.020

Crossref Full Text | Google Scholar

Martí-Juan, G., Sanroma-Guell, G., and Piella, G. (2020). A survey on machine and statistical learning for longitudinal analysis of neuroimaging data in Alzheimer's disease. Comput. Methods Programs Biomed. 189:105348. doi: 10.1016/j.cmpb.2020.105348

PubMed Abstract | Crossref Full Text | Google Scholar

Mattap, S. M., Mohan, D., McGrattan, A. M., Allotey, P., Stephan, B. C., Reidpath, D. D., et al. (2022). The economic burden of dementia in low- and middle-income countries (LMICs): a systematic review. BMJ Glob Health. 7:e007409. doi: 10.1136/bmjgh-2021-007409

PubMed Abstract | Crossref Full Text | Google Scholar

Mattsson, N., Zetterberg, H., Hansson, O., Andreasen, N., Parnetti, L., Jonsson, M., et al. (2009). CSF biomarkers and incipient Alzheimer disease in patients with mild cognitive impairment. JAMA 302, 385–393. doi: 10.1001/jama.2009.1064

PubMed Abstract | Crossref Full Text | Google Scholar

Mol, D. e, De Vito, C., and Rosasco, E. L. (2009). Elastic-net regularization in learning theory. J. Complex. 25, 201–230. doi: 10.1016/j.jco.2009.01.002

Crossref Full Text | Google Scholar

Pires, I. M., Hussain, F., Garcia, N. M., Lameski, P., and Zdravevski, E. (2020). Homogeneous data normalization and deep learning: a case study in human activity classification. Fut. Internet 12, 1–14. doi: 10.3390/fi12110194

Crossref Full Text | Google Scholar

Pudjihartono, N., Fadason, T., Kempa-Liehr, A. W., and O'Sullivan, J. M. (2022). A review of feature selection methods for machine learning-based disease risk prediction. Front. Bioinform. 2:927312. doi: 10.3389/fbinf.2022.927312

PubMed Abstract | Crossref Full Text | Google Scholar

Raju, M., Gopi, V. P., and Anitha, V. S. (2021). “Multi-class classification of Alzheimer's Disease using 3DCNN features and multilayer perceptron,” in 2021 Sixth International Conference on Wireless Communications, Signal Processing and Networking (WiSPNET) (IEEE), 368–373. doi: 10.1109/WiSPNET51692.2021.9419393

Crossref Full Text | Google Scholar

Raza, K. (2019). “Improving the prediction accuracy of heart disease with ensemble learning and majority voting rule,” in U-Healthcare Monitoring Systems (Academic Press), 179–196. doi: 10.1016/B978-0-12-815370-3.00008-6

Crossref Full Text | Google Scholar

Reddy, C. K. K., Kaza, V. S., Anisha, P. R., Khubrani, M. M., Shuaib, M., Alam, S., et al. (2024). Optimising barrier placement for intrusion detection and prevention in WSNs. PLoS ONE 19:e0299334. doi: 10.1371/journal.pone.0299334

PubMed Abstract | Crossref Full Text | Google Scholar

Ritchie, C. W., and Ritchie, K. (2012). The PREVENT study: a prospective cohort study to identify mid-life biomarkers of late-onset Alzheimer's disease. BMJ Open 2:1893. doi: 10.1136/bmjopen-2012-001893

PubMed Abstract | Crossref Full Text | Google Scholar

Sacchet, M. D., Prasad, G., Foland-Ross, L. C., Thompson, P. M., and Gotlib, I. H. (2015). Support vector machine classification of major depressive disorder using diffusion-weighted neuroimaging and graph theory. Front. Psychiatry 6:21. doi: 10.3389/fpsyt.2015.00021

PubMed Abstract | Crossref Full Text | Google Scholar

Saleem, T. J., Zahra, S. R., Wu, F., Alwakeel, A., Alwakeel, M., Jeribi, F., et al. (2022). Deep learning-based diagnosis of Alzheimer's disease. J. Pers. Med. 12:815. doi: 10.3390/jpm12050815

PubMed Abstract | Crossref Full Text | Google Scholar

Seng, J. K. P., and Ang, K. L. M. (2017). Big feature data analytics: split and combine linear discriminant analysis (SC-LDA) for integration towards decision making analytics. IEEE Access 5, 14056–14065. doi: 10.1109/ACCESS.2017.2726543

Crossref Full Text | Google Scholar

Shukla, A., Tiwari, R., and Tiwari, S. (2023). Review on Alzheimer disease detection methods: automatic pipelines and machine learning techniques. Science 5:13. doi: 10.3390/sci5010013

Crossref Full Text | Google Scholar

Siddiqui, A. J., Jahan, S., Siddiqui, M. A., Khan, A., Alshahrani, M. M., Badraoui, R., et al. (2023). Targeting monoamine oxidase b for the treatment of Alzheimer's and Parkinson's diseases using novel inhibitors identified using an integrated approach of machine learning and computer-aided drug design. Mathematics 11:1464. doi: 10.3390/math11061464

Crossref Full Text | Google Scholar

Siqueira, L. F., Júnior, R. F. A., de Araújo, A., Morais, A., and Lima, C. L. K. M. (2017). LDA vs. QDA for FT-MIR prostate cancer tissue classification. Chemometr. Intell. Labor. Syst. 162, 123–129. doi: 10.1016/j.chemolab.2017.01.021

Crossref Full Text | Google Scholar

Stamate, D., Kim, M., Proitsi, P., Westwood, S., Baird, A., Nevado-Holgado, A., et al. (2019). A metabolite-based machine learning approach to diagnose Alzheimer-type dementia in blood: results from the European Medical Information Framework for Alzheimer disease biomarker discovery cohort. Alzheimer's Dementia 5, 933–938. doi: 10.1016/j.trci.2019.11.001

PubMed Abstract | Crossref Full Text | Google Scholar

Sun, Q., Ni, J., Wei, M., Long, S., Li, T., Fan, D., et al. (2022). Plasma β-amyloid, tau, neurodegeneration biomarkers and inflammatory factors of probable Alzheimer's disease dementia in Chinese individuals. Front. Aging Neurosci. 14:963845. doi: 10.3389/fnagi.2022.963845

PubMed Abstract | Crossref Full Text | Google Scholar

Tartaglia, M. C., Rosen, H. J., and Miller, B. L. (2011). Neuroimaging in dementia. Neurotherapeutics 8, 82–92. doi: 10.1007/s13311-010-0012-2

PubMed Abstract | Crossref Full Text | Google Scholar

Tharwat, A., Gaber, T., Ibrahim, A., and Hassanien, A. E. (2017). Linear discriminant analysis: a detailed tutorial. AI Commun. 30, 169–190. doi: 10.3233/AIC-170729

Crossref Full Text | Google Scholar

Uddin, K. M. M., Alam, M. J., Uddin, M. A., and Aryal, S. (2023). A novel approach utilizing machine learning for the early diagnosis of Alzheimer's disease. Biomed. Mater. Devices 1, 882–898. doi: 10.1007/s44174-023-00078-9

PubMed Abstract | Crossref Full Text | Google Scholar

Valero-Carreras, D., Aparicio, J., and Guerrero, N. M. (2021). Support vector frontiers: a new approach for estimating production functions through support vector machines. Omega 104:102490. doi: 10.1016/j.omega.2021.102490

Crossref Full Text | Google Scholar

Wang, D., Li, J., Sun, Y., Ding, X., Zhang, X., Liu, S., et al. (2021). A machine learning model for accurate prediction of sepsis in ICU patients. Front. Public Health 9:754348. doi: 10.3389/fpubh.2021.754348

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, Y., Pan, Z., and Dong, J. (2022). A new two-layer nearest neighbor selection method for kNN classifier. Knowl. Based Syst. 235:107604. doi: 10.1016/j.knosys.2021.107604

Crossref Full Text | Google Scholar

Whiteford, H. A., Ferrari, A. J., Degenhardt, L., Feigin, V., and Vos, T. (2015). The global burden of mental, neurological and substance use disorders: an analysis from the global burden of disease study 2010. PLoS ONE 10:e0116820. doi: 10.1371/journal.pone.0116820

PubMed Abstract | Crossref Full Text | Google Scholar

Witten, I. H., Frank, E., Hall, M. A., Pal, C. J., and Data, M. (2005). “Practical machine learning tools and techniques,” in Data mining (Amsterdam, The Netherlands: Elsevier), 403–413.

Google Scholar

Yang, L., and Shami, A. (2020). On hyperparameter optimization of machine learning algorithms: theory and practice. Neurocomputing 415, 295–316. doi: 10.1016/j.neucom.2020.07.061

Crossref Full Text | Google Scholar

Yang, X. S. (2019). “Logistic regression, PCA, LDA, and ICA,” in Introduction to Algorithms for Data Mining and Machine Learning (Elsevier), 91–108. doi: 10.1016/B978-0-12-817216-2.00012-0

PubMed Abstract | Crossref Full Text | Google Scholar

Ying, C., Qi-Guang, M., Jia-Chen, L., and Lin, G. (2013). Advance and prospects of adaboost algorithm. Acta Autom. Sinica 36, 745–758. doi: 10.1016/S1874-1029(13)60052-X

Crossref Full Text | Google Scholar

Zhan, W., Wang, K., and Cao, J. (2023). Elastic-net based robust extreme learning machine for one-class classification. Signal Proc. 211:109101. doi: 10.1016/j.sigpro.2023.109101

Crossref Full Text | Google Scholar

Zhang, S., Li, J., and Li, Y. (2022). Reachable distance function for KNN classification. IEEE Trans. Knowl. Data Eng. 35, 7382–7396. doi: 10.1109/TKDE.2022.3185149

Crossref Full Text | Google Scholar

Zhao, T., Zheng, Y., and Wu, Z. (2023). Feature selection-based machine learning modeling for distributed model predictive control of nonlinear processes. Comput. Chem. Eng. 169:108074. doi: 10.1016/j.compchemeng.2022.108074

Crossref Full Text | Google Scholar

Keywords: Alzheimer's disease, biomarker, early diagnosis, drug, hybrid clinical model, machine learning, multinomial classification, protein

Citation: Khan A, Zubair S, Shuaib M, Sheneamer A, Alam S and Assiri B (2024) Development of a robust parallel and multi-composite machine learning model for improved diagnosis of Alzheimer's disease: correlation with dementia-associated drug usage and AT(N) protein biomarkers. Front. Neurosci. 18:1391465. doi: 10.3389/fnins.2024.1391465

Received: 25 February 2024; Accepted: 12 August 2024;
Published: 06 September 2024.

Edited by:

Peter Kokol, University of Maribor, Slovenia

Reviewed by:

Miren Altuna, Fundacion CITA Alzheimer, Spain
Bojan Žlahtič, University of Maribor, Slovenia

Copyright © 2024 Khan, Zubair, Shuaib, Sheneamer, Alam and Assiri. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Abdullah Sheneamer, asheneamer@jazanu.edu.sa

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.