- 1Department of Computer Science, HITEC University, Taxila, Pakistan
- 2Department of Artificial Intelligence (AI), College of Computer Engineering and Science, Prince Mohammad Bin Fahd University, Al Khobar, Saudi Arabia
- 3College of Computer Science, King Khalid University, Abha, Saudi Arabia
- 4Department of Internal Medicine, Soonchunhyang University Cheonan Hospital, Cheonan, Republic of Korea
- 5ICT Convergence Research Centre, Soonchunhyang University, Asan, Republic of Korea
One of the most prevalent disorders relating to neurodegenerative conditions and dementia is Alzheimer's disease (AD). In the age group 65 and older, the prevalence of Alzheimer's disease is increasing. Before symptoms showed up, the disease had grown to a severe stage and resulted in an irreversible brain disorder that is not treatable with medication or other therapies. Therefore, early prediction is essential to slow down AD progression. Computer-aided diagnosis systems can be used as a second opinion by radiologists in their clinics to predict AD using MRI scans. In this work, we proposed a novel deep learning architecture named DenseIncepS115for for AD prediction from MRI scans. The proposed architecture is based on the Inception Module with Self-Attention (InceptionSA) and the Dense Module with Self-Attention (DenseSA). Both modules are fused at the network level using a depth concatenation layer. The proposed architecture hyperparameters are initialized using Bayesian Optimization, which impacts the better learning of the selected datasets. In the testing phase, features are extracted from the depth concatenation layer, which is further optimized using the Catch Fish Optimization (CFO) algorithm and passed to shallow wide neural network classifiers for the final prediction. In addition, the proposed DenseIncepS115 architecture is interpreted through Lime and Gradcam explainable techniques. Two publicly available datasets were employed in the experimental process: Alzheimer's ADNI and Alzheimer's classes MRI. On both datasets, the proposed architecture obtained an accuracy level of 99.5% and 98.5%, respectively. Detailed ablation studies and comparisons with state-of-the-art techniques show that the proposed architecture outperforms.
1 Introduction
Brain-related disorders are among the most challenging conditions due to their high cost, their sensitivity, and the difficulty in treating them. The most prevalent brain disease affecting people is Alzheimer's disease, which causes various degrees of memory loss and knowledge loss (1). According to the most recent World Alzheimer Report (2), there are 55 million clinically diagnosed AD patients worldwide, and by 2050, that number is expected to reach 139 million (3). People over the age of 65 are the most likely to suffer from this irreversible disorder (4). Alzheimer's disease is more prevalent in individuals with diabetes, cardiovascular issues, and hypertension. This neurological condition begins slowly and gets worse every day. The early signs and symptoms of Alzheimer's include memory loss, difficulties completing daily tasks, confusion about where you are, visual or spatial difficulties, language difficulties, poor decisions, withdrawal from work, mood swings like depression, as well as behavioral and personality changes (5). As a result of the most recent advances in multimodal neuroimaging data, early disease detection has been enabled in neuroscience (6, 7). The pattern of brain shrinkage and image intensities, however, are so similar that it has been difficult to distinguish between healthy and Alzheimer's brains (8). The body gradually loses its ability to function, which may finally result in death, even though the exact cause of this disorder is unknown (9). AD does not have a cure at present, but if detected early, its progression may be slowed.
In order to diagnose brain disorders, neuroimaging techniques such as magnetic resonance imaging (MRI) (10) or computed tomography (CT) (11) with positron emission tomography (PET) (11) use images to provide three-dimensional (3D) images of the brain (12). It was estimated that AD patients would live only 3.1 years, especially if they were diagnosed at an early stage. Mild cognitive impairment (MCI) is a state of amnesia that may be an early indicator of Alzheimer's disease (13), but it is constantly getting worse. Non-sympathetic (generalized psychosis) Alzheimer's disease progresses in three stages: mild (stage 1), severe (stage 2), and moderate dementia (stage 3) (14). It is essential for AD diagnoses to be automated since clinical treatments are highly expensive (15). Machine learning (ML) paradigms have been expanded with new methods for learning in healthcare, especially for medical imaging (16). The traditional ML techniques focused on handcrafted features such as shape, color, and geometric; however, the complex imaging datasets required more patterns for an accurate prediction. Issues in traditional ML techniques include extra middle steps for the extraction of individual patterns, large number of extracted patterns (extracted features), and irrelevant information extraction.
Recent advancements in the area of deep learning (DL) techniques are becoming more and more popular in several applications, especially in medical imaging (17, 18). Convolutional neural networks (CNN), a specialized DL method, significantly outperform state-of-the-art ML techniques in various applications (19). These studies highlight CNN's superior ability to automatically extract and learn complex features from data, leading to improved accuracy and performance in tasks such as image recognition, classification, and other predictive modeling scenarios (20, 21). The DL algorithms work based on the hidden layers such as convolutional layer, pooling layer, and fully connected layers (22). There are several pre-trained DL architectures available for classification purposes such as GoogleNet (23), Alexnet (24), VGG (25), ResNet, and Inception-ResNet (25). Each model has a different way of learning mechanism. These models performed well for several classification tasks; however, in medical imaging, it degrades the performance when there is an issue of data imbalance and complex patterns of the disease. In addition, these models degraded the performance due to similar patterns of different classes, as shown in Figure 1. Therefore, for the customized models, it is always required that these are designed from scratch for a specific problem (26). The customized CNN models are usually designed based on the structure of the hidden layers and the number of learnable (27). This work proposed a novel network-level fused deep architecture based on dense and inception modules with self-attention mechanism for the classification of brain MRIs to predict and diagnose Alzheimer's disease. In the proposed network, deep features are extracted from brain MRI images using DenseNet and multi-scale Inception modules. These architectures are further integrated at the network level and can accurately predict patients with AD such as EMCI, MCI, and LMCI and those who are cognitively normal.
Figure 1. Five different categories of ADNI dataset: Alzheimer's disease (AD), cognitively normal (CN), early mild cognitive impairment (EMCI), late mild cognitive impairment (LMCI), and subjective memory complaint (MCI).
Following is a summary of the major contributions of this work:
● Develop a novel network-level fused convolutional neural network architecture called Dense with Multiscale Inception Self Module (DenseIncepS115) for the classification of Alzheimer's disease.
● The inception module with multiscale heads is designed to improve model accuracy and generalization. The concept of inverted bottlenecks is also considered to optimize network learning on the selected datasets.
● To increase model robustness, a data augmentation technique is also designed based on color enhancement.
● Extracting deep features with the use of multi-head self-attention mechanisms that are later optimized using Catch Fish Optimization Algorithm (CFOA).
● Demonstrating the efficiency of the proposed architecture, several ablation studies have been performed such as computed results using several pre-trained networks, selection of learning rate through manual approach, and confidence interval-based analysis.
2 Literature review
Alzheimer's disease has been attributed to genetics, but its underlying causes remain unknown. There are numerous issues related to social cognitive abilities that are affected by Alzheimer's, which also cause a number of neurological issues related to memory (28). Sarraf et al. (29) discussed about the very first use of MRI data in deep learning applications for Alzheimer's disease prediction and medical image analysis. The suggested pipelines produced average accuracy rates of 94.32% for fMRI and 97.88% for MRI, whereas the high accuracy rate of 98.84% for MRI was attained by subject-level classification. Hamdi et al. (12) addressed the problem of differentiating Alzheimer's disease patients from normal controls by developing a novel and improved CAD system based on a convolutional neural network (CNN). Using the 18FDG-PET images of 855 patients—220 Alzheimer's disease patients and 635 normal controls—from the ADNI database, the presented method was assessed. The findings demonstrated that the CAD system produced 96% accuracy, 96% sensitivity, and 94% specificity, respectively. Zhang et al. (30) developed an attention-based CNN that was trained using multilevel data from brain MRI to classify Alzheimer's disease. Using the ADNI dataset, this method identified Alzheimer's disease patients with 97.35% accuracy, MCI converters with 87.82% accuracy, and non-converters with 78.79% accuracy, respectively.
Divya et al. (31) utilized MRI features from the ADNI dataset to classify normal control (NC), mild cognitive impairment (MCI), and Alzheimer's disease (AD) using feature selection techniques and supervised learning algorithms. The best results were obtained with a support vector machine (SVM) with a radial basis function kernel, achieving 96.82%, 89.39%, and 90.40% accuracy for binary classifications of NC/AD, NC/MCI, and MCI/AD, respectively. The LeNet model was modified in this study (32) by concatenating Min-Pooling layers with Max-Pooling layers so that low-intensity pixels are preserved. This new model performed best when compared with 20 different DNN models, with an average accuracy of 96.64% for Alzheimer's disease classification compared to 80% for the original LeNet.
Shamrat et al. (33) developed a fine-tuned CNN architecture for Alzheimer's disease prediction into five stages. They modify the model based on the layers and hyperparameters. After that, they used 60,000 MRI scans from the ADNI database and obtained highest accuracy of 96.31%. Tanveer et al. (34) presented a computationally efficient ensemble of neural networks trained with transfer learning. The classification accuracy for mild cognitive impairment (MCI) vs. Alzheimer's disease (AD) was 98.71% and 83.11%, respectively, on two independent datasets split by cognitively normal (NC) vs. AD. Hajamohideen et al. (35) developed an architecture for a Siamese Convolutional Neural Network (SCNN) which embeds input images as k-dimensional embedding's with a triplet loss function. This embedding space was used to classify Alzheimer's disease using both pre-trained and non-trained CNNs. Model effectiveness was evaluated using ADNI and OASIS datasets, which yielded accuracy rates of 91.83% and 93.85%, respectively.
Alp et al. (36) investigated the use of Vision Transformer (ViT) for processing MRIs in the diagnosis of Alzheimer's disease. ViT used as a time series transformer to classify the MRI features once they had been extracted and modeled. On ADNI T1-weighted MRIs, the model was evaluated for binary and multiclass classification. In comparison to deep learning models such as CNN with BiL-STM and ViT with Bi-LSTM, the model demonstrated a high level of accuracy, scoring above 95% for binary and 96% for multiclass classification. Shaffi et al. (37) offered an ensemble classifier based on machine learning for the prediction of AD from MRI scans. It demonstrated an amazing accuracy of 96.52%, which is 3-5% better than the best individual classifier. They used the Alzheimer's Disease Neuroimaging Initiative and Open Access Series of Imaging Studies datasets to assess well-known machine learning classifiers and obtained improved accuracy above 94%. The study (16) suggested a bilateral filtering and histogram equalization image enhancement strategy to enhance the quality of the dataset. Then, to classify dementia into three groups, a custom CNN architecture has been designed. Using the designed custom architecture, the presented architecture obtained an accuracy of 93.45% and 95.62% for multiclass and binary class, respectively.
In summary, the above mentioned studies focused on the pre-trained models and supervised learning algorithms. In addition, they also focused on the traditional augmentation techniques such as flip and rotate methods. These methods not focused on the networks fusion and features optimization. Also, they did not focus on the shallow neural network classifiers for the prediction of AD from the MRI scans. In this work, we proposed a novel network-level fused CNN architecture for the prediction of AD from the MRI scans.
3 Proposed methodology
In this work, we proposed novel DenseIncepS115 architecture for the prediction of Alzheimer's disease from MRI scans. The proposed architecture comprises two novel modules—inception module with multiscale self-heads and dense module with self-attention. Figure 2 illustrates the proposed architecture of Alzheimer's disease prediction. In the proposed architecture, a dataset augmentation step has been performed at the first step to increase the training diversity. In the second step, both modules are fused at the network level using depth-concatenation layer. The model is trained on the selected dataset, whereas Bayesian optimization (BO) has been opted for the hyperparameters' initialization. In the testing phase, features are extracted from the testing data and optimized using an improved Catch Fish Optimization Algorithm technique. The best features are selected and passed to the shallow wide neural network (SWNN) classifier for final classification. Furthermore, the fused model is interpreted using GRAD-CAM and LIME explainable techniques.
Figure 2. Proposed automated framework for classifying Alzheimer's disease using network-level fusion.
3.1 Dataset collection and augmentation
ADNI dataset: Data from the Alzheimer's disease neuroimaging initiative (ADNI) database were used in this investigation. In 2003, the ADNI was developed as a public–private partnership under the direction of principal investigator Michael W. Weiner, MD. The purpose of this project was to explore the potential utility of positron emission tomography (PET), magnetic resonance imaging (MRI), and other biological markers in monitoring the development of early diagnosis of Alzheimer's disease and mild cognitive impairment (38). The dataset used in this work, which was collected from Kaggle, contains five categories: AD, CN, EMCI, LMCI and MCI genetic information, and the results from cognitive tests are included in the ADNI collection (39). Figure 3 illustrates the sample images of this dataset. The dataset contains both and female patients, wherein 100 out of 416 patients aged 60+ have been diagnosed with AD, ranging from very mild to moderate. A brief description of the dataset is given in Table 1.
Alzheimer MRI dataset: The AD preprocessed magnetic resonance imaging (MRI) images (ADMPIs) make up the dataset. The dataset contains four classes such as mildly demented, moderately demented, non-demented, and very mildly demented. The image details in each class are given in Table 1. In this work, 1,564 ADMPIs were used from the Kaggle source (40). A few sample images are shown in Figure 3.
During dataset collection, it was observed that the samples in both datasets have a class imbalance problem and are insufficient to train a deep learning model efficiently. This can lead to biases and overfitting during the model's training. To address these challenges, we performed an augmentation process to normalize the samples in each class and increase the diversity in the selected datasets. In the augmentation step, we performed three operations, such as flip left, flip right, and rotations. The overall description of the datasets is presented in Table 1.
3.2 Proposed DenseIncepS115 architecture
Deep learning is an important research area in the area of computer vision for classification and detection tasks. Medical imaging is an important research area, and many deep learning techniques are introduced for the classification and detection of medical diagnosis such as breast cancer, skin cancer, brain tumor, Alzheimer's disease (AD), and a few more. AD draws much attention from researchers working in the area of neuro-related diseases. Through DL, a better precision rate can be achieved for diagnosis and classification. In this paper, we proposed a novel DenseIncepS115 Architecture with explainable AI (XAI) for the classification and interpretation of AD from MRI scans. The dense and inception modules were chosen because the densely connected layers allowed the network to capture fine-grained information, such as subtle tissue degradations and gray matter changes, which indicate early Alzheimer's disease. In addition, inception modules have the capability to capture inclusive and specific regions, such as overall brain shrinkage and cortical folding–shifting in Alzheimer's disease. In this study, the capabilities of both modules are integrated using a network-level fusion method in the classification of Alzheimer's disease through the different phases.
The proposed architecture consists of two modules such as an inception module with multiscale self-heads and a dense module with self-attention. The total number of parameters of the proposed architecture is 6.9 million, whereas the total numbers of layers is 115. Figure 4 shows the architecture of the proposed DenseIncepS115 model. The input size of the proposed architecture is 227 × 227 × 3.
Dense module with self-attention: The self-module consists of 71 layers, including nine residual blocks and 22 convolution layers. At the start of the architecture, a convolutional layer has been added with a 3 × 3 filter size, stride value of 2, and depth of 32. Then, ReLU activation has been added as a nonlinearity function that attached with the batch normalization layer. After this, a first residual block has been added. Each block contains five layers such as convolutional layer, BN (batch normalization), ReLU, convolutional layer, and ReLU layers. In the first residual block, the first convolutional layer has a depth size of 64 with a stride value of 1 and 1×1 filter size. A ReLU activation layer is added with the batch normalization layer. The first residual block concludes by adding the last convolutional layer with depth of 64, stride value of 1, and filter size of 1 × 1. Then, these residual blocks are connected to the other layers by adding the first depth concatenation layer that concatenate inputs along the channel dimension which involves taking inputs with the same height and width.
The second residual block is added in same layer pattern in which the convolutional layers have a depth size of 96 and 64 with a stride value of 1 and filter size 1 × 1. Moreover, a depth concatenation layer has been added to connect it with the other layers, and there is a skip connection between both depth concatenation layers. A convolutional layer with ReLU has been added after this, with filter size of 2 × 2 and stride value of 2. The depth size is increased in this block to 128 from what was previously 96. A max-pool layer is added after this to get the most activated neurons with a stride value of 2 and filter size of 3 × 3. Three more residual blocks are added after the max-pooling layer. All of these blocks are connected with skip connection through the depth concatenation layer. The number of layers in each block is five, which is similar to the first residual and second residual block. The convolutional layers in the first block have a depth size of 128 and filter size of 1 × 1. In the second block, the convolutional layers have a depth size of 128 and 64, whereas the stride is 1 and the filter size is 1 × 1. A third block follows the first block, of which both convolutional layers have depth of 128, filter size of 1 × 1, and stride of 1. After that, a convolutional layer with ReLU is added, which has a depth of 256, a filter size of 2 × 2, and a stride value of 2. Another max-pool layer has been attached with a filter size of 3 × 3 and a stride value of 2.
There are four more residual blocks attached to the network after the second max-pool layer that are connected via depth concatenation. The max-pooling layer and the depth concatenation layer, as well as the remaining four depth concatenation layers, are interconnected by skip connections. These blocks follow the same number of layers (five) as the previous blocks in the same pattern but with different parameters. In the first block, both convolutional layers have a depth of 256 with different filter sizes of 1 × 1 and 3 × 3 and a stride value of 1 × 1. With the second residual block, both convolutional layers have a depth size of 128 and constant filter sizes of 1 × 1 and stride of 1. The third block follows the first block in which both convolutional layers have depth of 256 and a stride value of 1 with the same filter size of 1 × 1. At the end, a convolutional layer has been added that has a filter size of 2 × 2, a stride value of 2, and a depth size of 256. Another max-pool layer is added with a filter size of 3 × 3 and a stride of 2. After that, a self-attention module is added for the in-depth information extraction of each image. The self-attention module is visually shown in Figure 5. The self-attention module is mathematically formulated as follows:
As the inputs of attention module are queries, keys, and values which are defined with , , and , respectively, these notations created a linear transformation as follows:
where denotes the input feature matrix of the convolutional layer and . After that, the attention score is computed among and using a dot product, which is later scaled down by the factor of . This scaling is performed to prevent the dot product from growing too large during the training process. The attention score can be formulated by using the following equation:
The computed score is passed to softmax function to obtain the final attention weights as follows:
On each row of the attention score, the softmax function has been applied, which sums up to 1 for the final score. Hence, the weighted sum is applied to create the attention weights as final features.
Inception module with multiscale self-heads: The proposed inception module with multiscale self-heads consists of 44 layers and follows the inverted bottleneck pattern in residual blocks. The first convolutional layer has a depth size of 16 and a filter size of 1 × 1 with a stride value of 2. Then, the first inception module has been added, in which the first convolutional layer is added with a depth value of 16 and a filter size of 1 × 1. The second convolutional layer also has a depth size of 16 and a filter size of 1 × 1 with a single stride. For the third convolutional layer, a depth size of 32 is opted, with a filter size of 1 × 1 and a stride value of 1. In the next or third convolutional layer, a max-pool layer has been added with a filter size of 3 × 3 and a stride value of 1. The first convolutional layer is followed by a second convolution layer with a depth size of 16 and a filter size of 1 × 1. The output is divided in two convolutional layers of filter size 1 × 3 and 3 × 1, respectively. Moreover, 1 × 1 and 3 × 1 filters are added before the max-pooling layer of filter size 1 × 1. When convolutions do not significantly change the dimensions of the input, neural networks performed better work and reduced the dimensions. An excess of small dimensional reductions may result in information loss; therefore, convolutions and smart factorization techniques can be made more computationally efficient. This module is connected by adding one depth concatenation layer. After that, a convolutional layer with ReLU is added, which has a filter size of 3 × 3 with a stride value of 2 and depth of 64. A max-pool layer that has a stride value of 1 and filter size of 3 × 3 was attached.
Three inverted residual blocks are added after the inception module, where each block consists of six layers such as convolutional, ReLU, batch normalization, grouped convolutional, convolutional, and batch normalization layer. In the first block, the convolutional layer has a depth size of 128, with a filter size of 1 × 1 and a stride value of 1. Then, a batch normalization layer is added with ReLU activation. After that, the grouped convolutional layer is added with a 3 × 3 filter size. The first residual block concludes by adding the last convolutional layer with a depth of 256, a stride value of 1, and a filter size of 1 × 1. These residual blocks are connected to the other layers by adding the first depth concatenation layer. The second block is identical to the previous block with the same parameters. In the third block, convolutional layer with a filter size of 1 × 1 was employed, with a stride value of 1 and depth size of 256 and 312, respectively. After that, a convolutional layer with ReLU is added, which has a filter size of 3 × 3 and a stride value of 2 with a depth of 256. A max-pooling layer has been added with 3 × 3 filter size that followed a global average pool layer. At the end, we added a multi-headed self-attention module, as illustrated in Figure 5. The flattened layer is always added before the self-attention module.
The outputs of both modules are fused at the network level using a depth concatenation layer, as shown in Figure 4. Through the depth concatenation layer, information of different layers can be stacked into a single layer based on the depth dimension. It enriches the information in the form of feature vector and, as an output, accurate AD prediction.
Consider that we have two feature vectors denoted by and of dimensions and , respectively. The depth concatenation (DC) is formulated through the following equation:
Here and denote the depth of both feature vectors that is 256. Hence, the final feature vector dimension after the DC layer is 512, which passed to the fully connected layer and softmax for the final classification. The softmax is employed as a classification layer in the proposed architecture. Mathematically, the combined loss function of softmax and cross entropy is defined as follows:
where denotes the raw score for the true class. The detailed architecture is described under Algorithm 1.
Training the proposed model: In training the proposed architecture, we used 70% of each dataset to train the model and the remaining 30% for the testing phase. In this process, several hyperparameters are required to initialize, such as initial learning rate, momentum, and batch size. These hyperparameters are usually initialized based on a random process or hit and trial; however, we employed Bayesian optimization (BO) (40) that returned the best-fit values after 100 iterations. There are a few other hyperparameters of this network, such as stride and filter size, that are selected based on literature knowledge. After the initialization process, the network has been trained with five folds and 100 epochs for each dataset.
3.3 Testing the proposed framework
The proposed trained model DenseIncepS115 is tested on the testing image set and explained in this section. Deep learning features are extracted from the depth concatenation layer, and the performance was analyzed. In the analysis process, some redundant features are found; therefore, we employed a feature selection algorithm—Catch Fish Optimization (CFO). The CFO algorithm is applied in order to select the best features and pass them to the shallow wide neural network and traditional neural network classifiers. The shallow neural network has an advantage as it requires less computational power and is easier to train compared to deep neural networks. This simplicity often results in faster processing times, which is crucial for timely diagnoses. The visual illustration of the testing phase is shown in Figure 6.
3.4 Catch Fish Optimization Algorithm
The extracted features from Alzheimer's disease are often complex and multidimensional. The motivation behind choosing the CFO algorithm is due to its adaptive search; it allows broad exploration across all feature dimensions. Once a promising region is found, it can refine (exploit) that region to select the most appropriate features. This makes it effective to avoid local optima and ensure significant patterns, which other methods may have missed, and the complexity of Alzheimer's disease might vary considerably, enabling the algorithm to react dynamically to various features without requiring user intervention. This facilitates partial expansion throughout the feature selection process, ensuring that the optimal features are selected. After feature extraction, it is observed that the extracted information has some redundant features, which leads to the ineffective use of computational resources, and weak features do not contribute expressively to the model's decision. Therefore, we employed the CFO algorithm for optimal feature selection. The Catch Fish Optimization (CFO) algorithm introduced in (41) is based on the easy and traditional fishing technique called "fishing in water bodies," which is frequently used in rural regions. In the past, fishermen used the phrase "trouble the water to catch the fish" to describe a method of disrupting the water in a pond to confuse the fish into thinking that they were in clear water and to capture them easily. The basic tenet of CFOA is teamwork among members to maximize fish catches. It is possible for individuals to use tools and share stories about their individual experiences when they fish in diverse bodies of water. The optimization process is based on the following steps:
Step 1: The first step is to initialize the population. Every fisherman in CFOA serves as a search agent. Mathematically, it is defined as follows:
Fisher is a complete matrix of location data for search agents in a space of dimension . The locations can be updated based on the following formula:
The location of the fisher in the dimension is denoted by , the upper and lower limits of the dimension are denoted by and , and is a random number between 0 and 1.
Step 2: In the second step, the fitness value and update fitness and optimal position are computed. The fitness values of each fisherman are determined via the fitness value evaluation function using their current position information. Mathematically, fitness can be determined as follows:
The fitness values of the first and second fishermen are represented by and , respectively, in the fitness matrix. The ratio of exploration to exploitation is evenly distributed among the iterations using 0.5. In this procedure, global search was performed during the first phase (when ) and exploited during the second phase (when ).
Step 3: In this step, we determine whether exploration or exploitation is the present stage. This phase is based on two sub-stages.
a) Phase of exploration (EFs/MaxEFs<0.5): Update the position and randomly reshuffle each fisherman's position using the formulas (14)–(16) or (17) and (18). Fish population and capture rate drop with continuous fish capturing. Fishermen shift from solo exploration to group encirclement using individual skills in support. This transition can be modeled using the capture rate parameter, represented by â.
Fishermen have the option of either a group catch or an independent search. Individual search is preferred when the rate of capture increases. They move to group capture as decreases. Random integers (0, 1) are used to simulate this: when , independent search is selected, and when , group capture is selected. Here is a random value between 0 and 1.
Independent search : Update the new position by applying the following equations:
For each of them, the range of values is (-1, 1), and is the empirical analysis value that the fisherman acquired using any (= with fishermen as the point of reference. The fitness values that show the optimal and worse results for every position update, following are, respectively, and . The number of iterations of the fishermen's positions is represented by .
and indicate the location of fisherman in k-dimension after the iterations of and . The random number is in the interval 0–1. The individual and the reference are separated by a Euclidean distance . A random unit vector of dimension is the . Identify the primary direction of travel and the distance using empirical analysis; positive guidance should be taken from the fisherman to the reference person. The range of exploration , which is less than or equal to , also varies according to the absolute value of and the total number of evaluations (EFs) carried out at the moment.
Group capture ƿ ≥ â: Nets are used by fishermen to cooperate with one another and maximize their fishing capability. They form up in groups of three to four at random to surround suspected areas. Following is how the model and formula are defined:
The consists of three or four people whose places have not been changed. The red-orange point of the group has as its target point. and indicate the state of the fisherman in group in the k-dimension after the and updates. The fisherman's speed as he reaches the center is a variable that differs from individual to individual and can range between (0, 1). The move offset has a value range of (-1, 1), a value range of (-1, 1) that decreases with an increase in the number of EFs.
b) Phase of exploitation (EFs/MaxEFs ≥ 0.5): Some fish escape capture during the search phase, but fishermen employ an effective tactic that attracts stray and hidden fish to a focal point, eventually surrounding them. Fishermen are dispersed with the greatest density in the center and the lowest density at the edges. This arrangement guarantees that those in the center capture the majority of the fish population while those in the edges grab escaping fish, hence enhancing capture rates through tactical cooperation. The positions are updated using the Gaussian distribution as follows:
● Among these, the Gaussian distribution function or GD has an overall mean of µ at 0 and an overall variance or that grows with an increase in evaluations and approaches 0 from 1, the position of the jth fisherman after the update. The is the fishermen's center matrix, which shows the average value of each dimension. A random integer , having a value of , is used to disperse fishermen into three ranges with Gbest representing the global optimal location.
Step 4: If the termination condition is not satisfied, repeat steps 2–4. In step 5, the final fitness and optimal positions are obtained. The local region is explored by encircling, and the independent inquiry is converted according to the catch rate into group capture, making the global exploration thorough and effective. To enable the CFO algorithm to identify the optimal solution with good robustness and efficacy, the optimal position is continually refined using group capture. The final selected features are passed to the shallow wide neural network classifier for the final AD prediction, as shown in Figure 6. The shallow neural network has an advantage as it requires less computational power, and it is easier to perform classification compared to machine learning and traditional multi-layered perception (MLP). This simplicity often results in faster processing times, which is crucial for timely diagnoses. The shallow neural network of this work consists of an input layer that accepts best features as input, a single hidden layer that is fully connected, and an output layer for the final classification.
4 Results and discussions
4.1 Experimental setup and performance measures
This section presents the experimental setup for the presented framework. There are two publicly available datasets, such as ADNI and MRI Alzheimer, that have been employed for the validation of the proposed framework. There is a 70:30 split between training and testing data, as mentioned in the section on the training of the proposed architecture. In the training phase, the initial learning rate was initialized through CFO algorithm that is 0.00014, momentum value of 0.7002, and batch size of 64. All of the experiments were carried out using 10-fold cross-validation. Cross-validation is performed at 10 k-folds in order to balance the computational bias and variance. Also, it increases the generalizability of performance estimates. The classifiers are chosen based on hidden layers, including neural networks such as narrow neural network (NNN), medium neural network (MNN), bi-layered neural network (BNN), SWNN, and tri-layered neural network (TNN), respectively. Each classifier performance is computed using several performance measures such as sensitivity, precision, F1-score, accuracy, and false negative rate (FNR). The simulation was run on a workstation with a 12-GB RTX 3000 graphics card and 256 GB of RAM, and all experiments were conducted using MATLAB 2023b.
4.2 Proposed framework results using ADNI dataset
The proposed DenseIncepS115 classification results for the Alzheimer's ADNI dataset are presented in this subsection. Features are extracted from the depth-wise concatenation layer and passed to the several classifiers including the base classifier SWNN. Table 2 discusses the results of this experiment, whereas the maximum obtained accuracy is 99.2% for the SWNN classifier. The precision and sensitivity rate of this classifier is also 99.2%. In addition to that, the F1-score is also computed to be of the value 99.22%. The rest of the listed classifiers in this table also obtained an accuracy rate of 98.4%, 85.6%, 99.1%, and 98.5%, respectively. The obtained precision rates for these classifiers are 98.4%, 85.94%, 99.12%, and 98.52%, respectively. Hence, it is noted that the NNN, SWNN, and TNN classifiers obtained better precision rates. Time is also noted during the final classification process, as mentioned in this table. The minimum required time is 14.05 (s) for SWNN classifier, whereas the maximum consumed time is 22.03 (s) by TNN. The SWNN classifier performance can be further verified through a confusion matrix, as shown in Figure 7. From this figure, it can be observed that the AD class predicted 100% correctly, whereas the CN class' correct prediction is at 96.3%.
The obtained accuracy and computational time are further optimized by employing a CFO algorithm that selects the best features. The results are presented in Table 3, showing the best accuracy of 99.5% for the SWNN classifier, whereas the time is reduced to 5.422 (s). The precision rate of this classifier is improved to 99.5% after the optimization process. In addition, the other classifiers, precision rates are 99.48%, 99.24%, 98.52%, and 97.36%, respectively. The precision rates are significantly improved for all of the classifiers after the optimization process. Moreover, time is also reduced after employing the optimization process as mentioned in this table, such as 6.623, 7.801, 8.218, and 9.503 (s), respectively. Figure 7 (DenseIncepS115 with CFO on ADNI dataset) illustrates the confusion matrix of SWNN classifier and shows the improved prediction accuracy. Overall, the SWNN classifier performed better for this dataset using both experiments.
Table 3. Classification results of the proposed fused network after employing Catch Fish Optimization Algorithm on the ADNI dataset.
4.3 Proposed results using the Alzheimer MRI dataset
The prediction results for the Alzheimer MRI dataset using the proposed DenseIncepS115 architecture are presented in Table 4. Features are extracted from the depth concatenation layer and passed to the optimization algorithm for the selection of best features. In the first experiment, the results are computed and listed in Table 4 for the originally deep extracted feature vector. The SWNN classifier obtained the highest accuracy of 98.1%, followed by the F1-score of 98.26%, recall rate of 98.62%, and precision rate of 97.92%, respectively. The values of these obtained measures can be confirmed by a confusion matrix, as shown in Figure 8 (DenseIncepS115 on Alzheimer MRI dataset). The rest of the classifiers' precision rates are 96.67%, 97.27%, 96.85%, and 96.80%, respectively. Time is also noted for this experiment during the prediction process, and the minimum noted time is 10.2 (s) for the SWNN classifier, whereas the maximum consumed time is 28.7 (s) for the TNN classifier. To improve the prediction rate and reduce the computational time, we employed the CFO algorithm, and the results are presented in Table 5. The SWNN classifier had the highest accuracy of 98.7%, whereas the noted F1-score of 98.2% was improved compared with experiment 1 for this dataset. The recall rate is 98.6%, and the precision rate is 98.2%, respectively. These measures can be confirmed by a confusion matrix (illustrated in Figure 8). In this figure, it is observed that the correct prediction rate of each class has been improved after employing the optimization algorithm. Time is also noted for each classifier during the prediction step, and SWNN classifier executed the fastest with an execution time of 7.654 (s). Overall, the precision rate is improved, and the time is reduced after employing the optimization algorithm.
Table 4. Classification results of the proposed fused network DenseIncepS115 on the Alzheimer MRI dataset.
Table 5. Classification results of the proposed fused network after employing Catch Fish Optimization Algorithm on the Alzheimer MRI dataset.
4.4 Discussion
A detailed discussion of the proposed architecture is described in this section in terms of ablation studies and comparison with state-of-the-art (SOTA) techniques. The proposed AD prediction framework that consists of two phases—training phase and testing phase—is illustrated in Figure 2. The proposed architecture is trained on AD MRI image datasets such as ADNI and Alzheimer MRI. The hyperparameters of the proposed architecture are initialized using the BO algorithm instead of manual selection. In the testing phase, features are extracted in the first phase and passed to the classifiers, and the results are noted in Tables 2, 4. In order to improve the precision rate and reduce the computational time, the CFO algorithm has been employed, and the results are given in Tables 3, 5. The SWNN classifier shows the improved performance that can be confirmed through confusion matrices, as illustrated in Figures 7, 8. To further validate the proposed architecture, we performed several ablation studies as follows:
Ablation study 1: In the first ablation study, we performed four experiments for each dataset. In the first experiment, features are extracted from the self-attention layer of the Proposed InceptionSA Module and passed to classifiers. In the second experiment, the Proposed DenseSA Module is opted, and self-attention layer features that are fused at the network level in the third experiment are extracted. In the last experiment, optimization is performed on the fused network. Tables 6, 7 present the precision values of these experiments for the selected datasets—ADNI and Alzheimer MRI. In Table 6, the ADNI dataset precision values are presented for each experiment. For the Proposed InceptionSA Module, the highest precision value is 95.20% which was achieved by the SWNN classifier. The precision rate is improved by the Proposed DenseSA Module to 96.16%, which is further improved in the network-level fusion step (99.21%). In the optimization, the precision rate is a little decreased, but, overall, the SWNN classifier performed well and the fusion process shows strength.
Table 7. Precision-based analysis of the proposed architecture performance using the Alzheimer MRI dataset.
Table 7 presents the obtained precision rates for each experiment using the Alzheimer MRI dataset. In this table, the Proposed InceptionSA Module obtained the highest precision rate of 93.60%, whereas the Proposed DenseSA Module improved the precision value to 94.76%. The precision rate of this experiment is further improved in the fusion step to 97.92%, which is a significant strength of this experiment. After the optimization process (experiment 4), the highest obtained precision rate is 98.20% by the SWNN classifier. Hence, the SWNN classifier shows the improved precision rate for the proposed fused network architecture and optimization algorithm.
Ablation study 2: In the second ablation study, we compared the proposed fused network with several pre-trained deep learning architecture including Alexnet, GoogleNet, Resnet, and InceptionV3. Figures 9, 10 illustrate the detailed comparison of several pre-trained models with the proposed architecture based on the precision value. In Figure 9, the comparison is conducted in two different experiments for ADNI dataset: (i) precision value obtained on the original deep architectures and (ii) precision value is computed after employing the optimization algorithm on features extracted from these deep models. The proposed architecture obtained an improved precision rate of 99.21% and 98.52%, respectively, for both experiments. Figure 10 shows the precision-based analysis for Alzheimer MRI dataset. In this figure, the proposed architecture obtained an improved accuracy of 97.92% and 98.2%, respectively, for both experiments performed.
Figure 9. Analysis in the base of accuracy for the proposed architecture and state-of-the-art pre-trained models using ADNI dataset. The results are computed with and without optimization algorithm.
Figure 10. Analysis in the base of accuracy for the proposed architecture and state-of-the-art pre-trained models using the Alzheimer MRI dataset.
4.5 Comparison with SOTA
A detailed comparison is conducted in this section with recent state-of-the-art (SOTA) techniques. The recent SOTA techniques' accuracy value is mentioned in Table 8. In this table, it is observed that the obtained accuracy by Odusami et al. (42) is 73.90%, which was later improved by Raza et al. (43) and El-Assy et al. (21) to 97.84% and 99.1%, respectively. The proposed architecture obtained improved an accuracy rate of 99.5%. For the Alzheimer's Kaggle MRI dataset, Gupta et al. (44) obtained an accuracy rate of 93.7%, whereas our proposed architecture obtained an accuracy rate of 98.7%. Hence, the proposed architecture shows the improved accuracy and precision rates based on the detailed ablation studies. The proposed architecture's interpretation is also presented in Figure 11. In this figure, the LIME explainable AI technique is employed for the interpretation that shows the insight strength of our work.
Computational overheads: The proposed architecture is lightweight based on its number of learnable and number of hidden layers; however, the training time is almost double after the augmentation process. The proposed architecture required 253 min and 45 s; however, we also trained our model on the original images, and it takes 122 min and 12 s. The accuracy is the main difference that was noted during the training process among augmented and original images—99.1% and 99.5% (training accuracy using augmented datasets such as Alzheimer's Kaggle MRI and ADNI Kaggle) and 93.1% and 95.2% (before augmentation).
5 Conclusion
Alzheimer's disease (AD) is the most frequent cause of dementia, affecting millions of people globally. It causes memory loss, cognitive decline, and breakdown in day-to-day functioning, which have a major detrimental impact on people and their families. In this paper, we presented a novel deep learning architecture for AD prediction from MRI images. Dataset augmentation has been performed at the initial phase, and then a novel CNN architecture called DenseIncepS115 that is based on the fusion of two modules—Proposed InceptionSA Module and Proposed DenseSA Module—was designed. The hyperparameters of the proposed architecture are initialized using BO instead of manual selection. The trained model is later validated in the testing phase, whereas the depth concatenation layer features are extracted and optimized using the CFO algorithm. The selected features are passed to a shallow wide neural network classifier and obtained improved accuracy levels of 98.7% and 99.5%, respectively, for the selected datasets. Based on the detailed results, analysis, and ablation studies, we conclude with the following points:
● Using data argumentation step, we overcome the problems of small data samples and class imbalance.
● The proposed InceptionSA extracted information from the multiscales that was further improved in the self-attention module.
● The proposed DenseSA module improved the weights information of an image that, in return, improved the prediction accuracy and reduced the computational time.
● The proposed network-level fused architecture improved the prediction accuracy that was further optimized using the CFO algorithm. The CFO algorithm improved the precision rate and reduced the computational time.
● Model interpretation through LIME and GRAD-CAM shows how accurately the proposed architecture is trained on selected datasets for AD prediction.
Hence, the well-being of Alzheimer's disease patients may benefit from improved AD diagnosis, which is one way in which this study benefits the scientific community. In terms of accuracy and prediction, the proposed framework outperforms the current approaches. In the future, the current work shall be shifted to vision transformers.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding authors.
Author contributions
FR: Methodology, Software, Writing – original draft. MK: Methodology, Project administration, Software, Writing – original draft. GB: Investigation, Methodology, Software, Writing – original draft. WA: Conceptualization, Methodology, Validation, Writing – original draft. AA: Formal analysis, Investigation, Methodology, Writing – review & editing. MM: Conceptualization, Data curation, Funding acquisition, Project administration, Software, Writing – review & editing. SJ: Funding acquisition, Project administration, Software, Visualization, Writing – review & editing. YN: Funding acquisition, Project administration, Supervision, Visualization, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. The authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Research Project under grant number RGP2/283/45. This research was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (no. RS-2023-00218176) and the Soonchunhyang University Research Fund.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Chimthanawala NM, Haria A, Sathaye S. Non-invasive biomarkers for early detection of Alzheimer’s disease: a new-age perspective. Mol Neurobiol. (2024) 61:212–23. doi: 10.1007/s12035-023-03578-3
2. Gauthier S, Webster C, Servaes S, Morais J, Rosa-Neto P. World Alzheimer report 2022: life after diagnosis: navigating treatment, care and support. London, England: Alzheimer’s Disease International (2023).
3. Viswan V, Shaffi N, Mahmud M, Subramanian K, Hajamohideen F. Explainable artificial intelligence in Alzheimer’s disease classification: A systematic review. Cogn Comput. (2024) 16:1–44. doi: 10.1007/s12559-023-10192-x
4. Illakiya T, Ramamurthy K, Siddharth M, Mishra R, Udainiya A. AHANet: Adaptive hybrid attention network for Alzheimer’s disease classification using brain magnetic resonance imaging. Bioengineering. (2023) 10:714. doi: 10.3390/bioengineering10060714
5. Asgharzadeh-Bonab A, Kalbkhani H, Azarfardian S. An alzheimer’s disease classification method using fusion of features from brain magnetic resonance image transforms and deep convolutional networks. Healthcare Analytics. (2023) 4:100223. doi: 10.1016/j.health.2023.100223
6. Khojaste-Sarakhsi M, Haghighi SS, Ghomi SF, Marchiori E. Deep learning for Alzheimer's disease diagnosis: A survey. Artif Intell Med. (2022) 130:102332. doi: 10.1016/j.artmed.2022.102332
7. Wu Y, Zhou Y, Zeng W, Qian Q, Song M. An attention-based 3D CNN with multi-scale integration block for Alzheimer's disease classification. IEEE J Biomed Health Inf. (2022) 26:5665–73. doi: 10.1109/JBHI.2022.3197331
8. Wong PC, Abdullah SS, Shapiai MI. Exceptional performance with minimal data using a generative adversarial network for alzheimer's disease classification. Sci Rep. (2024) 14:17037. doi: 10.1038/s41598-024-66874-5
9. Pallawi S, Singh DK. Review and analysis of deep neural network models for Alzheimer's disease classification using brain medical resonance imaging. Cogn Comput Syst. (2023) 5:1–13. doi: 10.1049/ccs2.12072
10. Voineskos AN, Hawco C, Neufeld NH, Turner JA, Ameis SH, Anticevic A, et al. Functional magnetic resonance imaging in schizophrenia: current evidence, methodological advances, limitations and future directions. World Psychiatry. (2024) 23:26–51. doi: 10.1002/wps.21159
11. Bonakdarpour B, Takarabe C. Brain networks, clinical manifestations, and neuroimaging of cognitive disorders: The role of computed tomography (CT), magnetic resonance imaging (MRI), positron emission tomography (PET), and other advanced neuroimaging tests. Clinics Geriatric Med. (2023) 39:45–65. doi: 10.1016/j.cger.2022.07.004
12. Hamdi M, Bourouis S, Rastislav K, Mohmed F. Evaluation of neuro images for the diagnosis of Alzheimer's disease using deep learning neural network. Front Public Health. (2022) 10:834032. doi: 10.3389/fpubh.2022.834032
13. Leela M, Helenprabha K, Sharmila L. Prediction and classification of Alzheimer disease categories using integrated deep transfer learning approach. Measurement: Sensors. (2023) 27:100749. Available at: https://www.sciencedirect.com/science/article/pii/S2665917423000855.
14. Aparna M, Rao BS. A novel automated deep learning approach for Alzheimer’s disease classification. IAES Int J Artif Intell. (2023) 12:451. doi: 10.11591/ijai.v12.i1.pp451-458
15. Kaur S, Gupta S, Singh S, Gupta I. Detection of Alzheimer’s disease using deep convolutional neural network. Int J Image Graphics. (2022) 22:2140012. doi: 10.1142/S021946782140012X
16. Awarayi NS, Twum F, Hayfron-Acquah JB, Owusu-Agyemang K. A bilateral filtering-based image enhancement for Alzheimer disease classification using CNN. PloS One. (2024) 19:e0302358. doi: 10.1371/journal.pone.0302358
17. Gulhare KK, Shukla S, Sharma L. Deep neural network classification method to Alzheimer’s disease detection. Int J Adv Res Comput Sci Softw Eng. (2017) 7:1–4. doi: 10.23956/ijarcsse/V7I6/0259
18. Uyguroğlu F, Toygar Ö., Demirel H. CNN-based Alzheimer’s disease classification using fusion of multiple 3D angular orientations. Signal Image Video Process. (2024) 18:2743–51. Available at: https://link.springer.com/article/10.1007/s11760-023-02945-w.
19. Sorour SE, Abd El-Mageed AA, Albarrak KM, Alnaim AK, Wafa AA, El-Shafeiy E. Classification of Alzheimer’s disease using MRI data based on Deep Learning Techniques. J King Saud University-Computer Inf Sci. (2024) 36:101940. doi: 10.1016/j.jksuci.2024.101940
20. Mohi ud din dar G, Bhagat A, Ansarullah SI, Othman MTB, Hamid Y, Alkahtani HK, et al. A novel framework for classification of different Alzheimer’s disease stages using CNN model. Electronics. (2023) 12:469. doi: 10.3390/electronics12020469
21. El-Assy A, Amer HM, Ibrahim H, Mohamed M. A novel CNN architecture for accurate early detection and classification of Alzheimer’s disease using MRI data. Sci Rep. (2024) 14:3463. doi: 10.1038/s41598-024-53733-6
22. Kolhar M, Kazi RNA, Mohapatra H, Al Rajeh AM. AI-driven real-time classification of ECG signals for cardiac monitoring using i-alexNet architecture. Diagnostics. (2024) 14:1344. doi: 10.3390/diagnostics14131344
23. Ma L, Wu H, Samundeeswari P. GoogLeNet-AL: A fully automated adaptive model for lung cancer detection. Pattern Recognition. (2024), 110657. doi: 10.1016/j.patcog.2024.110657
24. Chen X, Chen H, Wan J, Li J, Wei F. An enhanced AlexNet-Based model for femoral bone tumor classification and diagnosis using magnetic resonance imaging. J Bone Oncol. (2024), 100626. doi: 10.1016/j.jbo.2024.100626
25. Davila A, Colan J, Hasegawa Y. Comparison of fine-tuning strategies for transfer learning in medical image classification. Image Vision Computing. (2024) 146:105012. doi: 10.1016/j.imavis.2024.105012
26. Mohammed FA, Tune KK, Assefa BG, Jett M, Muhie S. Medical image classifications using convolutional neural networks: A survey of current methods and statistical modeling of the literature. Mach Learn Knowledge Extraction. (2024) 6:699–735. doi: 10.3390/make6010033
27. Srinivasan S, Francis D, Mathivanan SK, Rajadurai H, Shivahare BD, Shah MA. A hybrid deep CNN model for brain tumor image multi-classification. BMC Med Imaging. (2024) 24:21. doi: 10.1186/s12880-024-01195-7
28. Khan R, Akbar S, Mehmood A, Shahid F, Munir K, Ilyas N, et al. A transfer learning approach for multiclass classification of Alzheimer's disease using MRI images. Front Neurosci. (2023) 16:1050777. doi: 10.3389/fnins.2022.1050777
29. Sarraf S, Tofighi G. Deepad: Alzheimer’s disease classification via deep convolutional neural networks using mri and fmri. bioRxiv. (2017), 070441. https://www.biorxiv.org/content/early/2017/01/14/070441.
30. Zhang J, Zheng B, Gao A, Feng X, Liang D, Long X. A 3D densely connected convolution neural network with connection-wise attention mechanism for Alzheimer's disease classification. Magnetic Resonance Imaging. (2021) 78:119–26. doi: 10.1016/j.mri.2021.02.001
31. Divya R, Shantha Selva Kumari R, Alzheimer’s Disease Neuroimaging Initiative. Genetic algorithm with logistic regression feature selection for Alzheimer’s disease classification. Neural Computing Appl. (2021) 33:8435–44. doi: 10.1007/s00521-020-05596-x
32. Hazarika RA, Abraham A, Kandar D, Maji AK. An improved LeNet-deep neural network model for Alzheimer’s disease classification using brain magnetic resonance images. IEEE Access. (2021) 9:161194–207. doi: 10.1109/ACCESS.2021.3131741
33. Shamrat FJM, Akter S, Azam S, Karim A, Ghosh P, Tasnim Z, et al. AlzheimerNet: An effective deep learning based proposition for alzheimer’s disease stages classification from functional brain changes in magnetic resonance images. IEEE Access. (2023) 11:16376–95. doi: 10.1109/ACCESS.2023.3244952
34. Tanveer M, Rashid AH, Ganaie M, Reza M, Razzak I, Hua K-L. Classification of Alzheimer’s disease using ensemble of deep neural networks trained through transfer learning. IEEE J Biomed Health Inf. (2021) 26:1453–63. Available at: https://ieeexplore.ieee.org/abstract/document/9440810.
35. Hajamohideen F, Shaffi N, Mahmud M, Subramanian K, Al Sariri A, Vimbi V, et al. Four-way classification of Alzheimer’s disease using deep Siamese convolutional neural network with triplet-loss function. Brain Inf. (2023) 10:5. doi: 10.1186/s40708-023-00184-w
36. Alp S, Akan T, Bhuiyan MS, Disbrow EA, Conrad SA, Vanchiere JA, et al. Joint transformer architecture in brain 3D MRI classification: its application in Alzheimer’s disease classification. Sci Rep. (2024) 14:8996. doi: 10.1038/s41598-024-59578-3
37. Shaffi N, Subramanian K, Vimbi V, Hajamohideen F, Abdesselam A, Mahmud M. Performance evaluation of deep, shallow and ensemble machine learning methods for the automated classification of Alzheimer’s disease. Int J Neural Syst. (2024), 2450029. doi: 10.1142/S0129065724500291
38. Weiner M, Aisen P, Jagust W, Landau S, Okonkwo O, Shaw LM, et al. Acknowledgement list for adni publications. Available at: https://adni.loni.usc.edu/wp-content/uploads/how_to_apply/ADNI_Acknowledgement_List.pdf.
39. Shukla GP, Kumar S, Pandey SK, Agarwal R, Varshney N, Kumar A. Diagnosis and detection of Alzheimer's disease using learning algorithm. Big Data Min Analytics. (2023) 6:504–12. doi: 10.26599/BDMA.2022.9020049
41. Jia H, Wen Q, Wang Y, Mirjalili S. Catch fish optimization algorithm: a new human behavior algorithm for solving clustering problems. Cluster Computing. (2024), 1–38. doi: 10.1007/s10586-024-04618-w
42. Odusami M, Maskeliūnas R, Damaševičius R, Misra S. Explainable deep-learning-based diagnosis of Alzheimer’s disease using multimodal input fusion of PET and MRI Images. J Med Biol Eng. (2023) 43:291–302. doi: 10.1007/s40846-023-00801-3
43. Raza N, Naseer A, Tamoor M, Zafar K. Alzheimer disease classification through transfer learning approach. Diagnostics. (2023) 13:801. doi: 10.3390/diagnostics13040801
Keywords: neuroimaging, Alzheimer's disease, MRI, network-level fusion, multiscale inception module, dementia stages classification, optimization, shallow neural network
Citation: Rauf F, Khan MA, Brahim GB, Abrar W, Alasiry A, Marzougui M, Jeon S and Nam Y (2024) DenseIncepS115: a novel network-level fusion framework for Alzheimer's disease prediction using MRI images. Front. Oncol. 14:1501742. doi: 10.3389/fonc.2024.1501742
Received: 25 September 2024; Accepted: 11 November 2024;
Published: 03 December 2024.
Edited by:
Abhishek Mahajan, The Clatterbridge Cancer Centre, United KingdomReviewed by:
Muhammad Yaqub, Hunan University, ChinaMenaka Radhakrishnan, Vellore Institute of Technology (VIT), India
Copyright © 2024 Rauf, Khan, Brahim, Abrar, Alasiry, Marzougui, Jeon and Nam. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Yunyoung Nam, eW5hbUBzY2guYWMua3I=; Muhammad Attique Khan, YXR0aXF1ZS5raGFuQGllZWUub3Jn