Skip to main content

ORIGINAL RESEARCH article

Front. Psychiatry, 24 June 2024
Sec. Aging Psychiatry
This article is part of the Research Topic Neuroimaging in Psychiatry 2023: Neurodegenerative Disorders View all 6 articles

Enhancing Alzheimer’s disease diagnosis and staging: a multistage CNN framework using MRI

Muhammad Umair Ali&#x;Muhammad Umair Ali1†Kwang Su Kim,&#x;Kwang Su Kim2,3†Majdi KhalidMajdi Khalid4Majed FarrashMajed Farrash4Amad Zafar*Amad Zafar1*Seung Won Lee*Seung Won Lee5*
  • 1Department of Artificial Intelligence and Robotics, Sejong University, Seoul, Republic of Korea
  • 2Department of Scientific Computing, Pukyong National University, Busan, Republic of Korea
  • 3Interdisciplinary Biology Laboratory (iBLab), Division of Biological Science, Graduate School of Science, Nagoya University, Nagoya, Japan
  • 4Department of Computer Science and Artificial Intelligence, College of Computing, Umm Al-Qura University, Makkah, Saudi Arabia
  • 5Department of Precision Medicine, Sungkyunkwan University School of Medicine, Suwon, Republic of Korea

This study addresses the pervasive and debilitating impact of Alzheimer’s disease (AD) on individuals and society, emphasizing the crucial need for timely diagnosis. We present a multistage convolutional neural network (CNN)-based framework for AD detection and sub-classification using brain magnetic resonance imaging (MRI). After preprocessing, a 26-layer CNN model was designed to differentiate between healthy individuals and patients with dementia. After detecting dementia, the 26-layer CNN model was reutilized using the concept of transfer learning to further subclassify dementia into mild, moderate, and severe dementia. Leveraging the frozen weights of the developed CNN on correlated medical images facilitated the transfer learning process for sub-classifying dementia classes. An online AD dataset is used to verify the performance of the proposed multistage CNN-based framework. The proposed approach yielded a noteworthy accuracy of 98.24% in identifying dementia classes, whereas it achieved 99.70% accuracy in dementia subclassification. Another dataset was used to further validate the proposed framework, resulting in 100% performance. Comparative evaluations against pre-trained models and the current literature were also conducted, highlighting the usefulness and superiority of the proposed framework and presenting it as a robust and effective AD detection and subclassification method.

1 Introduction

Alzheimer’s disease (AD) is a progressive neurodegenerative disease that causes irreversible cognitive dysfunction, amnesia, and progressive loss of brain function, eventually resulting in an inability to function independently in daily life (1). AD is the most prevalent type of dementia and requires a high level of medical attention. Global projections anticipate a significant increase in the number of individuals affected by AD, with an estimated 152 million affected by 2050 compared to the current 47 million. This poses challenges across the economic, medical, and societal domains (2). Every 3 seconds, one person worldwide is affected by dementia, with AD accounting for 60% of all dementia cases (3). The phases of dementia associated with AD can be broadly divided into the following categories: i) mild cognitive impairment (MCI), ii) mild dementia, iii) moderate dementia, and iv) severe dementia. MCI, which is often characterized by memory loss with increasing age, can lead to dementia in certain individuals. Individuals with mild dementia occasionally struggle with cognitive deficits that affect their daily activities. The symptoms include disorientation, memory loss, uncertainty, personality changes, and difficulties in performing daily chores. Moderate dementia significantly complicates daily life and requires greater assistance and care. These symptoms are more prominent and resemble those observed in patients with mild dementia. Individuals may require assistance with basic tasks, such as brushing their hair, and experience significant personality changes, including sudden onset of irritation or anxiety. Sleep disturbances were also frequent. Patients with severe dementia experience a marked decline in their condition as symptoms progress. Loss of communication skills may necessitate full-time care. The inability to perform basic activities, such as sitting in a chair or holding one’s head up, and loss of bladder control are characteristics of this stage.

Currently, there are no viable treatments to cure or decrease the progression of AD, and a complete understanding of its pathogenesis remains elusive. MCI is the transitional stage between AD and normal cognitive aging. Individuals with MCI are more likely to develop AD than those with age-matched healthy cognition (4). Preventive strategies to limit disease progression as well as efficient treatment and care procedures depend on the early detection of AD.

Medical history, physical examination, and further diagnostic tests such as neurological screenings that examine reflexes, coordination, and muscle tone are all common components of AD diagnostic evaluation (5). Magnetic resonance imaging (MRI), computed tomography (CT), and positron emission tomography (PET) are important imaging techniques for the diagnosis of AD. In particular, fluid-attenuated inversion recovery sequences in MRI are used to suppress the cerebrospinal fluid, allowing for a more thorough evaluation of anatomical structures and examination of gliotic alterations (6, 7). The advantages of MRI include improved flexibility, clear tissue contrast, lack of ionizing radiation, and the capacity to provide insightful information regarding the structure of the human brain (8). Developing an improved computer-aided diagnostic system capable of analyzing MRI images to determine whether individuals have AD or are in good health is crucial.

A wide range of machine-learning methodologies that integrate neuroimaging have improved the accuracy of identifying distinct dementia subtypes (9). Conventional machine-learning methods such as support vector machine (SVM) (10), random forest (11), and linear program boosting algorithms (12) have been used to detect AD using MRI (13). Furthermore, the variants of SVM and the ensemble of classifiers were also proposed. However, conventional machine-learning approaches frequently require the manual selection of predefined brain areas of interest based on established MRI markers linked to AD. Due to an inadequate understanding of definitive MRI biomarkers for AD, predetermined regions are likely unable to include all the information required to unravel the complexities of AD. In addition to being labor-intensive and time-consuming, manual selection also has the potential for subjective errors. Likewise, when it comes to choosing ensemble methods, managing computational expenses poses a significant challenge.

In contrast, deep-learning networks adopt a more advanced approach, including methods such as convolutional neural networks (CNNs), deep belief networks, recurrent neural networks, long-short term memory, stacked autoencoders, and restricted Boltzmann machine (1417). These techniques combine low-level features in the data to automatically create a higher-level, more abstract representation of a learning system (18). Because of high image classification accuracy, deep learning state-of-the-art techniques are preferred over conventional machine-learning methods. In general, CNN and its variants are the most popular deep-learning algorithms due to their better performance in identifying AD. CNN models are widely used in segmentation, object recognition, and classification (1921). This popularity can be attributed to several benefits including leveraging the spatial information of neighboring pixels, direct acceptance of image data as input, and efficient reduction of model parameters by utilizing weight-sharing, subsampling, and local receptive fields. A CNN trained using MRI slices can automatically extract features from images, thereby eliminating the requirement for manual feature selection during the learning phase (22). Furthermore, they also display higher generalization skills when dealing with scans from various sources or scanners.

Several CNN models have recently been proposed as diagnostic tools (23). Lu et al. (24) presented a multimodal CNN model using MRI and PET. Their proposed framework yielded an accuracy of 82.4% for patients with MCI who were subsequently diagnosed with AD. The model achieved a classification accuracy of 86.3% for individuals without dementia. In another study (25), accuracies of 90.05 and 85.55% were achieved for different datasets using CNN-based features and softmax as classifiers for binary classification (normal controls and AD). A pretrained AlexNet CNN model was used to retrieve deep features, and conventional machine-learning methods were used as classifiers (26). The results showed that the proposed methodology outperformed other handcrafted features with an accuracy of 99.21%. A VGG-16-based CNN model was used to classify the MRI slices (27). The model achieved a high accuracy of 95.73% for tertiary classification problems (early MCI, normal control, and late MCI). Pan et al. (28) hybridized ensemble learning and CNN models to classify brain MRI for various classification problems. Their models exhibited reasonable classification performance. Murugan et al. (29) presented the DEMNET model to classify various stages of dementia using MRI. They used the synthetic minority oversampling technique (SMOTE) approach to resolve class imbalance issues. Their model achieved a high accuracy of 95.23% for four classes. Although DEMNET shows high classification performance, the accuracy of an augmented dataset is not considered reliable for real-time applications. Recently, Fathi et al. (30) introduced a weighted probability-based ensemble method to combine six 2D-CNN architectures and obtain a high classification rate of 93.88 for four classes. Furthermore, they compared different ensemble methods and showed that the ensemble methods yielded better results than individual architectures. Kang et al. (31) proposed a three-round learning strategy based on a 3D deep convolutional generative adversarial network model and obtained an accuracy of 92.8% for two classes.

To improve the prediction performance, numerous studies have recently included attention models. The attention models focus on the most informative image regions. By combining two distinct attention modules (i.e., enhanced non-local attention and coordinate attention), Illakiya et al. (32) presented an adaptive hybrid attention network to enhance the performance of the DenseNet architecture, resulting in a higher classification accuracy of 98.53%. Similarly, in another study (33), an integrated model consisting of a depthwise group shuffle, global context network, hybrid multi-focus attention block, and EfficientNEt-B0 was developed to improve the prediction performance of MCI classification. Zhang et al. (34) developed an end-to-end 3D CNN framework based on the ResNet architecture, which employs multi-modality brain images to perform AD diagnostic and MCI prediction tasks by integrating 3D attention processes with multi-layer feature fusion algorithms. They showed that their multimodal model outperformed a single modality in predicting AD and MCI, with superior results of 6.37% and 3.51%, respectively. Some studies have also combined the transformer and attention networks. Hu et al. (35) designed a classifier model by combining a CNN with a swine transformer. In addition, they added a shift window attention mechanism to the transformer to improve the feature extraction. They achieved an accuracy of 93.5% for the two classes using their proposed model. Illakiya et al. (36) utilized a swine transformer, a dimension-centric proximity-aware attention network, and an age deviation factor to improve feature extraction from brain MRI images. The proposed network improves the classification results by utilizing a novel feature fusion strategy that incorporates global, local, and proximal characteristics, as well as dimensional dependencies. The literature describes various methods for classifying AD using conventional machine-learning and deep-learning models. However, there are challenges related to the large number of model parameters, training time, and high performance without augmentation in multiclass AD classification.

To address these issues, the primary contributions and steps of this study are outlined below:

● We hypothesized that leveraging the frozen weights of the developed CNN on correlated medical images facilitated the transfer learning process for sub-classifying dementia classes.

● To prove the hypothesis, a lightweight CNN model was developed in stage 1 to detect dementia using MRI images after preprocessing.

● In stage 2, a new model was built by reutilizing the frozen weight of the developed model for further classification of dementia into mild, moderate, and very mild dementia using transfer learning.

● Various online AD datasets were used to validate the proposed model.

● Various pre-trained models were trained using the same parameters and datasets for a fair comparison.

The results were compared with those reported in the literature.

2 Materials and methods

2.1 Proposed framework

The proposed CNN-based framework is depicted in Figure 1. In the proposed framework, dementia detection and sub-classification were divided into two stages (i.e., stage 1 and stage 2). In stage 1, the brain MRI scans were classified into two classes (healthy and dementia) using a proposed 26-layer CNN model (presented in Section 2.4). In stage 2, transfer learning was used to reutilize the frozen weights of the 26-layer CNN model (i.e., developed in stage 1) to fine-tune the new transfer-learned model by replacing the last three layers of the developed CNN for dementia subclassification (mild, moderate, and very mild dementia).

Figure 1
www.frontiersin.org

Figure 1 CNN-based proposed framework for dementia detection and subclassification.

2.2 Datasets

Two different datasets (the Alzheimer’s dataset (four classes of images) and ADNI_Extracted_Axial) were used to validate the proposed approach. Both datasets are publicly available (https://www.kaggle.com/datasets/tourist55/alzheimers-dataset-4-class-of-images and https://www.kaggle.com/datasets/katalniraj/adni-extracted-axial; accessed November 13, 2023). The specifications of this dataset are listed in Table 1.

Table 1
www.frontiersin.org

Table 1 Details of various online available datasets.

2.3 Pre-processing

In CNN applications, irrelevant information in an image can adversely affect the subsequent image-processing steps. Preprocessing is imperative to address these issues and ensure the accuracy of subsequent steps in image processing. Therefore, a cropping and zero-center approach was applied to remove unwanted information and normalization (37). After preprocessing and normalization of the dataset, the images were input into a developed CNN, which identified the AD-affected area by extracting discriminating features.

2.4 Design of CNN for AD Detection

CNNs are a type of deep-learning model developed specifically for analyzing structured grid data, such as images. Their ability to learn hierarchical feature representations autonomously has revolutionized computer vision tasks. CNNs comprise layers that perform convolutional operations to capture local patterns and pooling operations to reduce spatial dimensions. Weight sharing characterizes these networks, allowing them to recognize comparable features across the input space. Convolutional layers are often followed by fully connected layers to achieve a high level of feature integration and classification. CNNs excel at image identification, object detection, and segmentation and demonstrate superior performance across various domains. The ability of CNNs to automatically extract significant characteristics from raw data makes them valuable tools for complicated pattern detection, leading to the development of artificial intelligence.

In this study, a 26-layer CNN model was developed from scratch to detect dementia in stage 1. This architecture comprised five blocks of convolutional layers, batch normalization, ReLU activation functions, and max-pooling layers. Fully connected layers and a softmax output layer were used for the classification. The model uses dropout for regularization, and the final output layer employs cross-entropy loss. Complete information regarding the designed CNN model is presented in Table 2. The details of the layers are presented in the subsequent sections.

Table 2
www.frontiersin.org

Table 2 Details of the designed CNN model for AD detection.

2.4.1 Input layer

The input layer of the developed model represents the initial layer and accepts normalized images. This layer sets the input size and normalization strategy for the subsequent processing.

2.4.2 Convolutional layer

The foundation of any CNN model is comprised of convolutional layers. Convolutional layers are the core layers of any CNN model and are responsible for the extensive computational work. The input image is passed through this layer to produce a feature map or response by convolving it with weight filters and adding bias values. Subsequently, the feature response is passed through the following layers. Mathematically, convolution involves taking the element-wise product of the filter and a patch of the input and summing up all these products. The input (x), can be expressed as Equation (1).

y=Wx+bi(1)

where W and bi are the filter and bias of each filter, respectively.

2.4.3 Batch normalization

Batch normalization was applied to the output of the convolutional layer. This layer normalizes the activation and enhances convergence and training stability. This introduced learnable parameters for scaling and shifting. x is assumed to be the convolutional layer output. Batch normalization normalizes x across batch dimensions using Equation (2) (38):

x^=xμσ2+(2)

where μ is the mean, σ is the variance, and is a small constant used to avoid division by zero. It scales and shifts the normalized output y=αx^+β, where α and β are learnable parameters.

2.4.4 Rectified linear unit activation

The rectified linear unit (ReLU) activation function was applied element-wise. ReLUs introduce nonlinearity, which allows the model to capture more complex data patterns. If the input value is positive, the ReLU activation function immediately outputs the value. If not positive, it outputs zero. This can be mathematically expressed as Equation (3) (29):

f(x)=max(0,x)(3)

2.4.5 Max pooling

Pooling layers were used between the convolutional layers to reduce the representation in the spatial domain and computation space. Max pooling helps retain essential information while reducing the computational complexity by reducing the spatial dimensions. Max pooling was calculated using Equation (4) (29):

Max Pooling=Floor(Input xPooling window sizestride+1)(4)

2.4.6 Fully connected layer

The fully connected layer is densely connected. It captures high-level features from convolutional layers and prepares a classification model. Mathematically, this can be expressed as Equation (5).

y=Wx+bi(5)

2.4.7 Dropout layer

Dropout prevents overfitting by randomly deactivating neurons during training, enhancing model generalization.

2.4.8 Softmax

Softmax applies an activation function to convert logs into class probabilities. Softmax ensures that the sum of probabilities for all classes is one.

2.4.9 Classification output

It uses cross-entropy loss, specifically ‘crossentropyex’ in MATLAB, for model training. Cross-entropy measures the dissimilarity between predicted and actual class probabilities as Equation (6).

H(y,p)=i=1npi·log(yi)(6)

where H denotes the cross-entropy loss, yi denotes the predicted probability distribution, and pi denotes the true probability distribution.

2.5 Design of transfer learned CNN for subclassification

Transfer learning, a machine-learning technique, leverages pre-existing models to expedite learning in new tasks. That is, the model developed for a task is reused as the starting point for the model for a second task. This approach is particularly useful when data are scarce. The main concept of transfer learning is to leverage the features learned from tasks with a large amount of available data to improve the performance of tasks with less data. This is based on the idea that tasks share commonalities that can be utilized to improve performance. Assume that a domain is composed of two elements (39, 40), as shown in Equation (7).

Am=Y+prob(y)(7)

In Equation (7), Y and prob(y) denote the feature space and marginal probability, respectively. Suppose a task has two components:

Lr=X+α(8)

In Equation (8), X and α symbolize the label space and the objective function, respectively. Here, Ams and Lrs represent the source domain and task, respectively, and Amt and Lrt represent the target domain and task, respectively. The goal of transfer learning is to utilize knowledge from the source domain to understand the conditional probability in the target domain.

In stage 2 of this study, transfer learning is applied to the 26-layer CNN model that was developed in the first stage (section 2.4). Reutilizing the frozen weights of the 26-layer CNN model, the model was retrained by replacing the last three layers (i.e., fully connected, softmax, and classification layers) for dementia subclassification (mild, moderate, and very mild dementia). Figure 2 shows the concept of knowledge sharing from dementia to dementia subclassification.

Figure 2
www.frontiersin.org

Figure 2 The concept of knowledge sharing.

3 Results

MATLAB 2023a was used to perform all simulations and analyses on a personal computer with the following specifications: Core i7, 12th Generation, 32 GB RAM, NVIDIA GeForce RTX 3050, 1 TB SSD, and 64-bit Windows 11 operating system. The dataset was randomly divided into 80 and 20 ratios for model training and testing, respectively. The images used for model testing were not used to train the CNN. The following initial parameters were utilized: 100 Epochs, 0.9 momentum, 128 mini batch-size, and 0.001 learning rate. The stochastic gradient descent with momentum (SGDM) solver was utilized to train and test the model.

First, various publicly available pre-trained CNNs, such as ResNet50, Inception-v3, GoogleNet, EfficientNet-b0, and DenseNet-201, were used to categorize the brain MRI dataset. Subsequently, the proposed CNN model was trained to classify the brain MRI scans using the same parameters. Table 3 shows a performance comparison of various pre-trained models with the developed 26-layer CNN; and the confusion matrix of all models is presented in Figure 3, which also shows the true positive rate (TPR), false negative rate (FNR), positive predictive value (PPV), and false discovery rate (FDR).

Table 3
www.frontiersin.org

Table 3 Performance comparison of various CNN models for Alzheimer’s Dataset.

Figure 3
www.frontiersin.org

Figure 3 Confusion matrix of various CNN models for Alzheimer’s Dataset. (A) ResNet50, (B) Inception-v3, (C) GoogleNet, (D) EfficientNet-b0, (E) DenseNet-201, and (F) Developed 26-layer CNN. MID, mild dementia; MOD, moderate dementia; ND, non-dementia; VMD, very mild dementia; CNN convolutional neural networks.

After thoroughly analyzing the results presented in Table 3 and Figure 3, it was found that the developed 26-layer CNN model has the best classification rate, with minimal training time and a high true positive rate for each class compared to all other pre-trained models. The learning curves of the various pre-trained models with the developed multistage 26-layer CNN are presented in Figure 4.

Figure 4
www.frontiersin.org

Figure 4 Learning curves of various convolutional neural networks. (A, C) shows the training accuracy and loss curves, (B, D) of training and depict the validation accuracy and loss curves.

After a comprehensive analysis of the results presented in Figure 4A, it was found that DenseNet-201 was 100% trained in fewer epochs than ResNet50, Inception-v3, GoogleNet, EfficientNet-b0, and 26-layer CNN; it took approximately 5 epochs to stabilize the results. DenseNet-201 also demonstrates the highest validation accuracy of 93.93% among all pre-trained models for dementia classification (Figure 4B). In contrast, GoogleNet took almost 32 epochs to train the model and also showed a reasonable validation performance (92.57%) than the remaining pre-trained models. The developed 26-layer CNN took almost 47 epochs to reach 100% training accuracy but had the best validation accuracy of 97.45% for dementia classification. Furthermore, the time consumed for the training of the developed 26-layer CNN was only 8 min 57 s for 100 epochs, which was the fastest among all methods. It validates the robustness and high classification performance of the developed CNN model compared to other pre-trained models.

To further enhance the dementia detection rate and subclassification performance, the proposed framework is divided into two stages. The performance of the proposed framework for both stages is presented in Table 4. Figures 5A, B show the confusion matrix for dementia detection and sub-classification, respectively.

Table 4
www.frontiersin.org

Table 4 Performance of proposed CNN-based framework for dementia detection and subclassification.

Figure 5
www.frontiersin.org

Figure 5 Confusion matrices using the proposed 2-stage framework. (A) Dementia detection and (B) dementia subclassification.

After carefully evaluating the result presented in Table 4, it was found that the proposed 2-stage approach performs efficiently to differentiate between healthy and dementia persons with a high classification rate of 98.24% with only 6 min 42 s training time. The comprehensive performance of the model is presented in Figure 5 using a confusion matrix. The proposed model only misclassified 9 samples of each class (Figure 5A), resulting in a high true positive rate of 98.2% for each class. After that, a new transfer-learned model was used for the subclassification of the dementia class and yielded a high accuracy of 99.7%, with only one sample of mild dementia misclassified, as shown in Figure 5B. The learning curves of the dementia detection and subclassification are shown in Figures 6A, B, respectively.

Figure 6
www.frontiersin.org

Figure 6 Training and loss curves. (A) for dementia detection and (B) for dementia subclassification. MID, mild dementia; MOD, moderate dementia; VMD, very mild dementia.

To further validate the performance of the proposed 2-stage approach against overfitting, the results of 10-fold cross-validation are presented in Figure 7.

Figure 7
www.frontiersin.org

Figure 7 Confusion matrices using 10-fold cross-validation for the proposed 2-stage framework. (A) Dementia detection and (B) dementia subclassification. AD, Alzheimer’s disease; CI, mild cognitive impairment.

Furthermore, other data were also used to validate the approach’s reliability, adaptability, and accuracy. The results of dementia detection and subclassification are presented in Figures 8A, B, respectively.

Figure 8
www.frontiersin.org

Figure 8 Confusion matrices using the proposed 2-stage framework for another dataset. (A) Dementia detection and (B) dementia subclassification. AD, Alzheimer’s disease; CI, mild cognitive impairment.

4 Discussion

This study investigated the application of CNNs to identify AD and differentiate between the different stages of dementia by analyzing MRI. Recently, there has been an increase in the use of computer-aided systems for early AD detection, using both machine learning and CNNs. This study contributes to the development of an automated AD detection system for improving the operating efficiency of medical centers.

The ablation study was performed for the layers selection of developed CNN. The effect of changing the number of layers (from 22 to 34 layers) is reported in Table 5. All models attain 100% training accuracy showing that the deep network models match the training data closely. However, 26-layer CNN yielded the less training loss. Furthermore, the 26-layer CNN model also gives higher validation accuracy and lower validation loss as compared to other models. This shows that the 26-layer model seems to provide a greater generalization of the unknown validation data. The 26-layer CNN model was chosen because it appears to achieve an optimal balance between model complexity and generalization at this depth. Too few or too many layers may result in suboptimal performance on validation data.

Table 5
www.frontiersin.org

Table 5 Results of ablation study for selection of layers.

The developed 26-layer CNN model achieved an impressive classification accuracy of 97.45% for directly categorizing MRI scans into four classes, demonstrating superior performance with minimal training time compared to several pre-trained CNNs (Table 3). DenseNet-201 yielded better results than the other pre-trained networks. In the validation, the proposed model correctly classified 998 of the 1023 images (see Figure 3F). The TPR of all classes was higher than 95%, with a very low FDR. All pre-trained models converged faster during the training, indicating that transfer learning facilitates fast convergence in the learning of pre-trained models (see Figures 4A, C).

The proposed technique for advancing diagnostic capabilities comprises two key stages. First (i.e., stage 1), a 26-layer CNN model was developed to detect dementia using MRI slices. Next (i.e., stage 2), the weights of the developed model were reutilized to subclassify the dementia class. In the first stage, to detect only dementia, the results showed that the proposed model yielded a high accuracy of 98.24% for binary classification, with a TPR of more than 98% for both classes (see Figure 5A). In the second stage, the developed CNN model was reused using the transfer learning concept for dementia subclassification. The results showed that only one sample out of 512 samples was misclassified, and the model produced a high classification rate of 99.7%, with a very high TPR and low FDR (see Figure 5B). Furthermore, fast convergence was observed as a result of transfer learning (Figure 6B). These results support our hypothesis that the frozen weight of a trained model from correlated images benefits transfer learning and results in a high classification performance. To further evaluate the performance of the model against data leakage issues, the authors have further performed 10-fold cross-validation. The proposed multistage framework shows almost similar high classification accuracy, further validating the effectiveness of the proposed approach (Figure 7). Another ADNI MRI scan dataset was used to validate the efficacy of the proposed framework. The framework validated the 100% classification rate of the developed CNN for dementia detection and subclassification (Figure 8). Table 6 compares the proposed multistage framework with those of recent studies.

Table 6
www.frontiersin.org

Table 6 Comparison of the proposed multistage CNN-based framework with recent research.

After deeply analyzing the results presented in Table 6, it can be concluded that the proposed CNN-based framework has the highest classification rate compared to others. These outcomes underscore the efficacy of the proposed model in efficiently and accurately handling the classification task, emphasizing its potential as a robust solution in AD detection and subclassification.

A single-modality dataset was used to evaluate the performance of the proposed network. In the future, multiple-modality datasets may be utilized to improve the classification performance for AD diagnosis. In addition, this study proposes a simple architecture; however, more intuitive architectures, such as transformers or the incorporation of attention networks, may be tested in the future. Confounding variables, such as the independent variable (imaging data in this case) and the dependent variable (presence or absence of AD), should be considered in designing AD studies. These confounding variables can introduce spurious correlations, leading to reduced AD identification accuracy. The key clinical confounding variables to be considered include age, education, vascular health, genetic factors, and lifestyle. For instance, a CNN model trained on an older-skewed dataset might learn age-related features instead of AD-specific ones, resulting in misdiagnosis for younger AD patients and overdiagnosis in healthy older adults. Other important confounding variables related to the imaging data of AD identification include data collection, preprocessing, model designing, and the evaluation of model performance for variables unseen during training. Image quality can be affected and data inconsistencies introduced by differences in acquisition paradigms, spatial resolution, and magnetic field strength. Furthermore, inadequate preprocessing can result in artifacts, spatial distortions, and data inconsistencies, all of which can have an impact on the accuracy and reliability of the study. Model architecture, hyperparameter tuning, and regularization techniques are also important in reducing confounding effects and maximizing model performance. Finally, the evaluation of the model must be done by employing a rigorous validation process. For example, k-fold cross-validation can assist in reducing the impact of data variability and produce more reliable estimates of model performance. Evaluation criteria should also be carefully chosen to take into consideration confounding variables unique to AD identification tasks, such as class imbalance and susceptibility to false positives.

5 Conclusion

AD is a common and devastating neurological condition that substantially reduces the quality of life in affected individuals. These effects affect not only patients but also their families and society. Timely diagnosis is critical to adequately control AD and reduce its socioeconomic consequences. This paper proposes a multistage CNN-based AD detection and subclassification framework. A 26-layer CNN model was developed from scratch using MRI images to detect dementia. The model yielded a high accuracy of 98.24% in dementia detection using an online AD dataset. Subsequently, the developed CNN model was reutilized for the subclassification of dementia classes using transfer learning. This yielded a high accuracy rate of 99.70%, with only one misclassified sample. Moreover, another AD dataset was used to validate the model, and the results showed a 100% performance rate. The proposed framework was also compared with various pre-trained models and the latest literature to prove the effectiveness and superiority of the proposed model.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding author/s.

Ethics statement

Ethical approval was not required for the study involving humans in accordance with the local legislation and institutional requirements. Written informed consent to participate in this study was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and the institutional requirements.

Author contributions

MA: Conceptualization, Methodology, Software, Writing – original draft. KK: Conceptualization, Methodology, Software, Writing – review & editing. MK: Conceptualization, Formal analysis, Writing – review & editing. MF: Formal analysis, Investigation, Writing – review & editing. AZ: Formal analysis, Investigation, Validation, Visualization, Writing – review & editing. SL: Funding acquisition, Project administration, Resources, Supervision, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work is supported by a National Research Foundation (NRF) grant funded by the Ministry of Education (MOE), South Korea, through the "Development Research Program" NRF2021R1I1A2059735 (SL). This work was supported by National Research Foundation of Korea (NRF) grants funded by the Korean government (MSIT) (2022R1C1C2003637) (to KK).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Ulep MG, Saraon SK, McLea S. Alzheimer disease. J Nurse Practitioners. (2018) 14:129–35. doi: 10.1016/j.nurpra.2017.10.014

CrossRef Full Text | Google Scholar

2. Salvatore C, Cerasa A, Battista P, Gilardi MC, Quattrone A, Castiglioni I. Magnetic resonance imaging biomarkers for the early diagnosis of alzheimer's disease: A machine learning approach. Front Neurosci. (2015) 9:307. doi: 10.3389/fnins.2015.00307

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Patterson C. World alzheimer Report 2018. (2018) (London, UK: Alzheimer’s Disease International (ADI)).

Google Scholar

4. Liu S, Liu S, Cai W, Pujol S, Kikinis R, Feng D eds. (2014). Early diagnosis of alzheimer's disease with deep learning, in: 2014 IEEE 11th international symposium on biomedical imaging (ISBI), Beijing, China: IEEE.

Google Scholar

5. Sabbagh MN, Lue L-F, Fayard D, Shi J. Increasing precision of clinical diagnosis of alzheimer's disease using a combined algorithm incorporating clinical and novel biomarker data. Neurol Ther. (2017) 6:83–95. doi: 10.1007/s40120-017-0069-5

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Noor MBT, Zenia NZ, Kaiser MS, Mamun SA, Mahmud M. Application of deep learning in detecting neurological disorders from magnetic resonance images: A survey on the detection of alzheimer’s disease, parkinson’s disease and schizophrenia. Brain Inf. (2020) 7:11. doi: 10.1186/s40708-020-00112-2

CrossRef Full Text | Google Scholar

7. Baratti C, Barkhof F, Hoogenraad F. Valk J. Partially saturated fluid attenuated inversion recovery (Flair) sequences in multiple sclerosis: comparison with fully relaxed flair and conventional spin-echo. Magnetic Resonance Imaging. (1995) 13:513–21. doi: 10.1016/0730-725X(95)00009-6

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Moser E, Stadlbauer A, Windischberger C, Quick HH, Ladd ME. Magnetic resonance imaging methodology. Eur J Nucl Med Mol Imaging. (2009) 36:30–41. doi: 10.1007/s00259-008-0938-3

CrossRef Full Text | Google Scholar

9. Mateos-Pérez JM, Dadar M, Lacalle-Aurioles M, Iturria-Medina Y, Zeighami Y, Evans AC. Structural neuroimaging as clinical predictor: A review of machine learning applications. NeuroImage: Clin. (2018) 20:506–22. doi: 10.1016/j.nicl.2018.08.019

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Rabeh AB, Benzarti F, Amiri H eds. (2016). Diagnosis of alzheimer diseases in early step using svm (Support vector machine), in: 2016 13th International Conference on Computer Graphics, Imaging and Visualization (CGiV); 29 March-1 April 2016. Beni Mellal, Morocco.

Google Scholar

11. Tripoliti EE, Fotiadis DI, Argyropoulou M. A supervised method to assist the diagnosis and monitor progression of alzheimer's disease using data from an fmri experiment. Artif Intell Med. (2011) 53:35–45. doi: 10.1016/j.artmed.2011.05.005

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Hinrichs C, Singh V, Mukherjee L, Xu G, Chung MK, Johnson SC. Spatially augmented lpboosting for ad classification with evaluations on the adni dataset. NeuroImage. (2009) 48:138–49. doi: 10.1016/j.neuroimage.2009.05.056

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Thayumanasamy I, Ramamurthy K. Performance analysis of machine learning and deep learning models for classification of alzheimer's disease from brain mri. Traitement Du Signal. (2022) 39:1961–70. doi: 10.18280/ts.390608

CrossRef Full Text | Google Scholar

14. Hinton GE. Deep belief networks. Scholarpedia. (2009) 4:5947.

Google Scholar

15. Vincent P, Larochelle H, Lajoie I, Bengio Y, Manzagol P-A, Bottou L. Stacked denoising autoencoders: learning useful representations in a deep network with a local denoising criterion. J Mach Learn Res. (2010) 11:3371–408.

Google Scholar

16. Rawat W, Wang Z. Deep convolutional neural networks for image classification: A comprehensive review. Neural Comput. (2017) 29:2352–449. doi: 10.1162/neco_a_00990

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Illakiya T, Karthik R. Automatic detection of alzheimer's disease using deep learning models and neuro-imaging: current trends and future perspectives. Neuroinformatics. (2023) 21:339–64. doi: 10.1007/s12021-023-09625-7

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Khamparia A, Singh KM. A systematic review on deep learning architectures and applications. Expert Syst. (2019) 36:e12400. doi: 10.1111/exsy.12400

CrossRef Full Text | Google Scholar

19. Ali MU, Kallu KD, Masood H, Tahir U, Gopi CVVM, Zafar A, et al. A cnn-based chest infection diagnostic model: A multistage multiclass isolated and developed transfer learning framework. Int J Intelligent Syst. (2023) 2023:6850772. doi: 10.1155/2023/6850772

CrossRef Full Text | Google Scholar

20. Alanazi MF, Ali MU, Hussain SJ, Zafar A, Mohatram M, Irfan M, et al. Brain tumor/mass classification framework using magnetic-resonance-imaging-based isolated and developed transfer deep-learning model. Sensors. (2022) 22:372. doi: 10.3390/s22010372

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Almalki YE, Ali MU, Kallu KD, Masud M, Zafar A, Alduraibi SK, et al. Isolated convolutional-neural-network-based deep-feature extraction for brain tumor classification using shallow classifier. Diagnostics. (2022) 12:1793. doi: 10.3390/diagnostics12081793

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Lin W, Tong T, Gao Q, Guo D, Du X, Yang Y, et al. Convolutional neural networks-based mri image analysis for the alzheimer’s disease prediction from mild cognitive impairment. Front Neurosci. (2018) 12:777. doi: 10.3389/fnins.2018.00777

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Wen J, Thibeau-Sutre E, Diaz-Melo M, Samper-González J, Routier A, Bottani S, et al. Convolutional neural networks for classification of alzheimer's disease: overview and reproducible evaluation. Med image Anal. (2020) 63:101694. doi: 10.1016/j.media.2020.101694

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Lu D, Popuri K, Ding GW, Balachandar R, Beg MF, Weiner M, et al. Multimodal and multiscale deep neural networks for the early diagnosis of alzheimer’s disease using structural mr and fdg-pet images. Sci Rep. (2018) 8:5697. doi: 10.1038/s41598-018-22871-z

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Ahmed S, Choi KY, Lee JJ, Kim BC, Kwon GR, Lee KH, et al. Ensembles of patch-based classifiers for diagnosis of alzheimer diseases. IEEE Access. (2019) 7:73373–83. doi: 10.1109/ACCESS.2019.2920011

CrossRef Full Text | Google Scholar

26. Nawaz H, Maqsood M, Afzal S, Aadil F, Mehmood I, Rho S. A deep feature-based real-time system for alzheimer disease stage detection. Multimedia Tools Appl. (2021) 80:35789–807. doi: 10.1007/s11042-020-09087-y

CrossRef Full Text | Google Scholar

27. Jain R, Jain N, Aggarwal A, Hemanth DJ. Convolutional neural network based alzheimer’s disease classification from magnetic resonance brain images. Cogn Syst Res. (2019) 57:147–59. doi: 10.1016/j.cogsys.2018.12.015

CrossRef Full Text | Google Scholar

28. Pan D, Zeng A, Jia L, Huang Y, Frizzell T, Song X. Early detection of alzheimer’s disease using magnetic resonance imaging: A novel approach combining convolutional neural networks and ensemble learning. Front Neurosci. (2020) 14:259. doi: 10.3389/fnins.2020.00259

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Murugan S, Venkatesan C, Sumithra MG, Gao XZ, Elakkiya B, Akila M, et al. Demnet: A deep learning model for early diagnosis of alzheimer diseases and dementia from mr images. IEEE Access. (2021) 9:90319–29. doi: 10.1109/ACCESS.2021.3090474

CrossRef Full Text | Google Scholar

30. Fathi S, Ahmadi A, Dehnad A, Almasi-Dooghaee M, Sadegh M. for the alzheimer’s disease neuroimaging I. A deep learning-based ensemble method for early diagnosis of alzheimer’s disease using mri images. Neuroinformatics. (2024) 22:89–105. doi: 10.1007/s12021-023-09646-2

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Kang W, Lin L, Sun S, Wu S. Three-round learning strategy based on 3d deep convolutional gans for alzheimer’s disease staging. Sci Rep. (2023) 13:5750. doi: 10.1038/s41598-023-33055-9

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Illakiya T, Ramamurthy K, Siddharth MV, Mishra R, Udainiya A. Ahanet: adaptive hybrid attention network for alzheimer’s disease classification using brain magnetic resonance imaging. Bioengineering. (2023) 10:714. doi: 10.3390/bioengineering10060714

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Illakiya T, Karthik R. A deep feature fusion network with global context and cross-dimensional dependencies for classification of mild cognitive impairment from brain mri. Image Vision Computing. (2024) 144:104967. doi: 10.1016/j.imavis.2024.104967

CrossRef Full Text | Google Scholar

34. Zhang Y, He X, Liu Y, Ong CZL, Liu Y, Teng Q. An end-to-end multimodal 3d cnn framework with multi-level features for the prediction of mild cognitive impairment. Knowledge-Based Syst. (2023) 281:111064. doi: 10.1016/j.knosys.2023.111064

CrossRef Full Text | Google Scholar

35. Hu Z, Li Y, Wang Z, Zhang S, Hou W. Conv-swinformer: integration of cnn and shift window attention for alzheimer’s disease classification. Comput Biol Med. (2023) 164:107304. doi: 10.1016/j.compbiomed.2023.107304

PubMed Abstract | CrossRef Full Text | Google Scholar

36. Illakiya T, Karthik R. A dimension centric proximate attention network and swin transformer for age-based classification of mild cognitive impairment from brain mri. IEEE Access. (2023) 11:128018–31. doi: 10.1109/ACCESS.2023.3332122

CrossRef Full Text | Google Scholar

37. Helaly HA, Badawy M, Haikal AY. Deep learning approach for early detection of alzheimer’s disease. Cogn Comput. (2022) 14:1711–27. doi: 10.1007/s12559-021-09946-2

CrossRef Full Text | Google Scholar

38. Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. Int Conf Mach Learn. (2015) 37:448–56.

Google Scholar

39. Fatima M, Khan MA, Shaheen S, Almujally NA, Wang S-H. B2c3netf2: breast cancer classification using an end-to-end deep learning feature fusion and satin bowerbird optimization controlled newton raphson feature selection. CAAI Trans Intell Technol. (2023) 8(4):1374–90. doi: 10.1049/cit2.12219

CrossRef Full Text | Google Scholar

40. Zahoor S, Shoaib U, Lali IU. Breast cancer mammograms classification using deep neural network and entropy-controlled whale optimization algorithm. Diagnostics. (2022) 12:557. doi: 10.3390/diagnostics12020557

PubMed Abstract | CrossRef Full Text | Google Scholar

41. Wang H, Shen Y, Wang S, Xiao T, Deng L, Wang X, et al. Ensemble of 3d densely connected convolutional network for diagnosis of mild cognitive impairment and alzheimer’s disease. Neurocomputing. (2019) 333:145–56. doi: 10.1016/j.neucom.2018.12.018

CrossRef Full Text | Google Scholar

42. Mohammed BA, Senan EM, Rassem TH, Makbol NM, Alanazi AA, Al-Mekhlafi ZG, et al. Multi-method analysis of medical records and mri images for early diagnosis of dementia and alzheimer&Rsquo;S disease based on deep learning and hybrid methods. Electronics. (2021) 10:2860. doi: 10.3390/electronics10222860

CrossRef Full Text | Google Scholar

43. Acharya H, Mehta R, Singh DK eds. (2021). Alzheimer disease classification using transfer learning, in: 2021 5th International Conference on Computing Methodologies and Communication (ICCMC); 8-10 April 2021. Erode, India.

Google Scholar

44. El-Latif AAA, Chelloug SA, Alabdulhafith M, Hammad M. Accurate detection of alzheimer's disease using lightweight deep learning model on mri data. Diagnostics. (2023) 13:1216. doi: 10.3390/diagnostics13071216

PubMed Abstract | CrossRef Full Text | Google Scholar

45. Loddo A, Buttau S, Di Ruberto C. Deep learning based pipelines for alzheimer's disease diagnosis: A comparative study and a novel deep-ensemble method. Comput Biol Med. (2022) 141:105032. doi: 10.1016/j.compbiomed.2021.105032

PubMed Abstract | CrossRef Full Text | Google Scholar

46. Kaplan E, Dogan S, Tuncer T, Baygin M, Altunisik E. Feed-forward lpqnet based automatic alzheimer's disease detection model. Comput Biol Med. (2021) 137:104828. doi: 10.1016/j.compbiomed.2021.104828

PubMed Abstract | CrossRef Full Text | Google Scholar

47. Ching WP, Abdullah SS, Shapiai MI, Islam AM. Transfer learning for alzheimer's disease diagnosis using efficientnet-B0 convolutional neural network. J Adv Res Appl Sci Eng Technol. (2024) 35:181–91. doi: 10.37934/araset.34.3.181191

CrossRef Full Text | Google Scholar

48. Mohammad F, Al Ahmadi S. Alzheimer’s disease prediction using deep feature extraction and optimization. Mathematics. (2023) 11:3712. doi: 10.3390/math11173712

CrossRef Full Text | Google Scholar

49. Hasan ME, Wagler A. New convolutional neural network and graph convolutional network-based architecture for AI applications in alzheimer’s disease and dementia-stage classification. AI. (2024) 5:342–63. doi: 10.3390/ai5010017

CrossRef Full Text | Google Scholar

50. Shukla A, Tiwari R, Tiwari S. Alz-convnets for classification of alzheimer disease using transfer learning approach. SN Comput Sci. (2023) 4:404. doi: 10.1007/s42979-023-01853-7

CrossRef Full Text | Google Scholar

Keywords: Alzheimer’s disease, MRI, convolutional neural network, dementia, neuroimaging

Citation: Ali MU, Kim KS, Khalid M, Farrash M, Zafar A and Lee SW (2024) Enhancing Alzheimer’s disease diagnosis and staging: a multistage CNN framework using MRI. Front. Psychiatry 15:1395563. doi: 10.3389/fpsyt.2024.1395563

Received: 04 March 2024; Accepted: 07 June 2024;
Published: 24 June 2024.

Edited by:

Gaelle Eve Doucet, Boys Town National Research Hospital, United States

Reviewed by:

Karthik R, Vellore Institute of Technology (VIT), India
Jannik Prasuhn, Johns Hopkins University, United States

Copyright © 2024 Ali, Kim, Khalid, Farrash, Zafar and Lee. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Amad Zafar, amad@sejong.ac.kr; Seung Won Lee, swleemd@g.skku.edu

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.