Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 11 July 2023
Sec. Brain Imaging Methods
This article is part of the Research Topic Advanced Machine Learning Approaches for Brain Mapping View all 16 articles

Development and validation of a deep-broad ensemble model for early detection of Alzheimer's disease

\r\nPeixian MaPeixian Ma1Jing WangJing Wang2Zhiguo Zhou
Zhiguo Zhou3*C. L. Philip ChenC. L. Philip Chen4the Alzheimer&#x;s Disease Neuroimaging Initiative the Alzheimer's Disease Neuroimaging Initiative5Junwei Duan,
Junwei Duan1,6*
  • 1College of Information Science and Technology, Jinan University, Guangzhou, China
  • 2College of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
  • 3School of Integrated Circuits and Electronics, Beijing Institute of Technology, Beijing, China
  • 4School of Computer Science and Engineering, South China University of Technology, Guangzhou, China
  • 5Steering Committee of Alzheimer's Disease Neuroimaging Initiative, Bethesda, MD, United States
  • 6Guangdong Provincial Key Laboratory of Traditional Chinese Medicine Informatization, Guangzhou, China

Introduction: Alzheimer's disease (AD) is a chronic neurodegenerative disease of the brain that has attracted wide attention in the world. The diagnosis of Alzheimer's disease is faced with the difficulties of insufficient manpower and great difficulty. With the intervention of artificial intelligence, deep learning methods are widely used to assist clinicians in the early recognition of Alzheimer's disease. And a series of methods based on data input with different dimensions have been proposed. However, traditional deep learning models rely on expensive hardware resources and consume a lot of training time, and may fall into the dilemma of local optima.

Methods: In recent years, broad learning system (BLS) has provided researchers with new research ideas. Based on the three-dimensional residual convolution module and BLS, a novel broad-deep ensemble model based on BLS is proposed for the early detection of Alzheimer's disease. The Alzheimer's Disease Neuroimaging Initiative (ADNI) MRI image dataset is used to train the model and then we compare the performance of proposed model with previous work and clinicians' diagnosis.

Results: The result of experiments demonstrate that the broad-deep ensemble model is superior to previously proposed related works, including 3D-ResNet and VoxCNN, in accuracy, sensitivity, specificity and F1.

Discussion: The proposed broad-deep ensemble model is effective for early detection of Alzheimer's disease. In addition, the proposed model does not need the pre-training process of its depth module, which greatly reduces the training time and hardware dependence.

1. Introduction

Alzheimer's disease (AD) is a chronic neurodegenerative disease of the brain that develops insidiously. People diagnosed with Alzheimer's will suffer from the disease for remaining lifespan (Todd et al., 2013; Weiner et al., 2013; Fink et al., 2020). The main symptoms of AD includes memory impairment, executive dysfunction, aphasia, impairment of visuospatial skills and so on, and the etiology remains unknown (Mayeux and Sano, 1999; Mimura and Yano, 2006). Thus, Millions of people around the world suffer from Alzheimer's disease. The long-term treatment of these patients consumes huge medical resources and costs (Cummings and Cole, 2002; Scheltens et al., 2016). The diagnosis of AD can be divided into three main types: AD (Alzheimer's disease), MCI (Mild Cognitive Impairment), NC (Normal Control).

Radiographic images are important in medical diagnosis of Alzheimer's disease. These include positron emission tomography (PET), magnetic resonance imaging (MRI), computed tomography (CT) and so on (Prince and Links, 2006; Doi, 2007; Johnson et al., 2012). Due to low cost and high efficiency, MRI imaging play an important role in diagnosing AD related pathological brain changes and researching (Jack et al., 1999). The understanding of the pathological information provided by these radiographic images depends on the knowledge and experience of the front-line clinicians. As the number of professional medical staff is far less than the actual patient treatment needs, they can not timely diagnose some early hidden symptoms of Alzheimer's disease. At the same time, the imbalance of medical resources also leads to the inability of patients in rural areas to obtain effective early diagnosis locally for follow-up treatment.

Currently, with the proposal of VGG (Simonyan and Zisserman, 2014), ResNet (He et al., 2016), ResNest (Zhang et al., 2022), ResNext (Xie et al., 2017) and a series of deep neural networks, many medical and artificial intelligence researchers used these model to conduct corresponding training on radiographic images. The popularity of high-performance hardware maked it possible to deploy these frameworks in some large hospitals and enable medical departments to actually use these methods to assist physicians in clinical diagnosis and reduce patient care costs. Due to the complex 3-Dimensional (3D) spatial feature of radiographic images, which is extremely different from the traditional 2-Dimensional (2D) images, a variety of model were proposed based on different inputs. Compared to 2D-input models, 3D-input models get more structure information from data obtained by continuous scanning, thus they can extract more complex three-dimensional spital feature. Consequently, in most application scenarios, 3D-input models performs better than the former in recognition tasks. These methods can be divided into 2D deep learning method and 3D deep learning method. The 2D method mainly divides the original medical image into multiple slices on a specific axis and then inputs them into the classical convolutional neural network for training. However, the 2D method cannot learn the correlation feature between these slices, so the model performance is limited. The 3D method directly input the original image into the 3D improved convolutional neural network for training, in order to learn more comprehensive feature information and make up for the above defects.

However, deep models contained a large number of hyperparameters, which required huge hardware resources. The gradient descent method is also prone to fall into the optimal solution, leading to the failure of weight. Researchers need to find a quick and effective way to solve this problem. In recent years, on the basis of Random Vector Functional-Link Neural Network (RVFLNN) (Pao et al., 1994) and Single-layer Linear Feedforward Network (SLFN) (Sanger, 1989), Chen et al. proposed the Broad Learning System (BLS) (Chen and Liu, 2017a,b) and proved its approximation. BLS showed good accuracy and excellent calculation speed in various classification tasks.

Therefore, on the basis of BLS, we try to combine it with deep learning to establish a depth-broad ensemble model. The 3D deep convolution module will enable the model to have the capacity to initially extract features of 3D inputs, while the broad learning module, as a key part of feature fitting, greatly reduces the resource consumption of the model and can maintain a good performance. While Alzheimer's image recognition is a emblematic 3D image processing task, it has had a profound influence on medicine and computer science. There have been a lot of research on the application of depth model in this aspect. Applying our proposed depth-broad ensemble model to the early detetion of Alzheimer's disease will help drive technological innovation, reduce the cost of future applications, and better facilitate the adoption of machine learning technologies in this field.

In this study, we proposed an improved deep-broad ensemble model for the detection of AD. This model combined the 3D extraction capability with the fast operation speed and low dependence on hardware. It firstly extracted spatial features of different levels of images, and then fused multi-level features based on a novel BLS to get better classification results. We applied this model to the task of MRI image recognition in Alzheimer's disease and compared it with some previous work and the work of radiology readers. Experimental results demonstrate that the proposed model has excellent accuracy and computational efficiency.

The main contributions of our study for the early detection of AD can be reported as follows:

1. We constructed a novel deep-broad ensemble model based on 3D residual convolution module and Broad Learning System.

2. The proposed model outperforms previous single deep models, and has higher training efficiency, less dependence in hardwares.

3. There is no need to pre-train the deep modules of the proposed, which greatly reduces the training time.

2. Related works

Early detection of Alzheimer's disease is a chronic and significant research topic in computer science area. At present, a large number of computer-aided early detection methods for Alzheimer's disease have been developed. According to the dimension division of input data, related works can be divided into two-dimensional input-based research methods and three-dimensional input-based research methods. Two-dimensional input methods consist of traditional machine learning model and two-dimensional deep learning models. The three-dimensional input method basically takes the three-dimensional deep learning model as the main backbone. Due to the abundant spatial pathological information in medical examination images, three-dimensional methods generally have more advantages in the detection effect. In addition, some researchers have studied the characteristics of small sample size of medical test images, or introduced different types of data to establish a multi-modal fusion algorithm.

Rieke et al. trained on a 3D CNN model and applied four gradient-based and occlusion-based approaches to visualization, promoting clinical impact and trust in computer-based decision support systems (Rieke et al., 2018). But 3DCNN contains the risk of network degradation and gradient disappearance/gradient explosion after increasing the number of middle layers. Rieke et al. trained on a 3D CNN model and applied four gradient-based and occlusion-based approaches to visualization, promoting clinical impact and trust in computer-based decision support systems. But 3DCNN contains the risk of network degradation and gradient disappearance/gradient explosion after increasing the number of middle layers. Based on convolutional autoencoder (CAE), Kanghan et al. proposed supervised and unsupervised classification methods for the diagnosis of Alzheimer's disease. The combination of convolutional layer and pooling layer of CAE is relatively fixed, which limits to construct more complex network structure. Guan et al. constructed a preliminary standardized model framework based on ResNet, VGG, DenseNet and other networks, and comprehensively tested and compared these models using standard MRI image data sets of Alzheimer's disease. They found that these simple architectures performed similarly on the task, and the pre-training process of these methods has less impact on accuracy. Korolev et al. proposed VoxCNN based on ResNet to classify MRI images. This model can achieve better performance using a small training dataset, and be applied to 3D MRI images without the need of intermediate handcrafted feature extraction. However, VoxCNN includes the module of 3D-Resnet, which needs to consume more training time and more computing resources in training.

In the above reports, these methods have achieved excellent performance in their selected datasets. However, these studies lacked comparability and robustness among themselves. Most studies needs lots of GPU resources to train the model, which makes it difficult to apply the research results widely.

3. Methods

3.1. Approvement statement of institutional review board

This study is approved by institutional board with written informed consent waived. All experiments including any relevant details are approved by institutional and/or licensing committee. All experiments on humans and/or the use of human tissue samples were performed in accordance with relevant guidelines and regulations. All experimental protocols were approved by the Steering Committee of Alzheimer's Disease Neuroimaging Initiative and Academic Committee of Jinan University. Informed consent was obtained from all subjects and/or their legal guardian(s).

3.2. Data acquisition

All MRI image data used in this study were obtained from the Alzheimer's Disease Neuroimaging Initiative (ADNI) database (http://adni.loni.usc.edu/) (Petersen et al., 2010). Founded in 2003, ADNI is a public-private partnership led by Principal Investigator Michael W.Weiner, MD. The primary goal of ADNI is to test whether serial magnetic resonance imaging (MRI), positron emission tomography (PET), other biomarkers, and clinical and neuropsychological assessments can be combined to measure the progression of mild cognitive impairment (MCI) and early Alzheimer's disease (AD). The latest information about the ANDI database can be found at (http://adni.loni.usc.edu/). The ADNI database consists of four sub-databases, including ADNI-1, ADNI-Go, ADNI-2, and ADNI-3, which are interdependent. The diagnostic labels of these medical image data are given by doctors after a series of tests.

ADNI provides several standardized datasets for researchers to study. In our research, ADNI1 Complete 2Yr 3T standardized dataset is chosen to train our model, including scans of patients taken at 6, 12, 18, and 24 months after diagnosis. The dataset contains 434 subjects, including 77 of AD, 206 of MCI, and 151 of NC. The demographic information of the dataset is shown in Table 1. We searched the ADNI1 Complete 2Yr 3T standardized dataset in the ADNI database, packaged it and downloaded. All image data in this study were stored in NifTI format. Figure 1 shows the slides samples of this dataset.

TABLE 1
www.frontiersin.org

Table 1. The structure of residual Conv module used in feature mapping layer and enhancement layer.

FIGURE 1
www.frontiersin.org

Figure 1. Partial slice samples of some ADNI datasets: (A–F) AD, (G–L) NC.

3.3. Data preprocessing

Since the data came from many different patient samples, the size of different data and the location of key parts in the image may vary to some extent, while the neural network model required that the size of each input be consistent. At the same time, due to its intensity and other attributes, MRI image is not suitable to be directly used as the input image of the model, so it needs to be transformed to some extent. In conclusion, we have to take a series of pre-processing measures for the data, so that it can be converted into appropriate input data, and is conducive to improving the performance of the model.

In this research, The primitive image size of our ADNI dataset was 256×256×160. Firstly, Since the pixel size of different medical scanned images is not the same, all pixels need to be resampled at a fixed homogeneous resolution. We resampled all MRI images to 1.5-mm isotropic voxels. Then, we scaled the intensity of the images to the range (0, 1). Since the region of interest (ROI), namely the patient's brain, was basically concentrated in the center of the image, we cropped the image based on the central region and removed some peripheral background areas. The final output after processing were 224×224×128-pixel grid resulting in 336×336×192-mm2 volume. The above data preprocessing operations were based on Python 3.8 environment, using package Monai (https://monai.io/) and package Numpy (https://numpy.org/) for processing.

3.4. Model details and training

3.4.1. Broad learning system

Although traditional deep neural networks show good accuracy in traditional recognition and classification tasks, they also expose problems such as large number of hyperparameters and long time consuming. At the same time, with the publication of more types of datasets, researchers need to seek a new method with simple structure and fast operation to deal with different requirements and tasks. In studies over the past few years, classical single-layer network structures such as Extreme Learning Machine (ELM) (Huang et al., 2006) and Random Vector Functional Link Neural Network (RVFLNN) (Pao et al., 1994) have been proposed successively.

On the basis of RVFLNN (Pao et al., 1994), Chen et al. proposed Broad Learning System (BLS) (Chen and Liu, 2017a,b). In the structure of BLS, the basic linear feature of the input was extracted by the feature mapping layer. The further feature of the former layer was extracted by the enhancement layer which contained a non-linear function. Then, the output of these layers were concat together and transferred to the output layer for classifying. Since there is only two layers of structure, BLS does not need to calculate a large number of weight parameters for multiple middle layers, which saving a lot of calculation resources and reducing the training time of the model. Previous experimental results demonstrate that BLS can still achieve excellent performance in the basic test of image recognition, which proves that BLS has good potential in the field of computer vision (Chen and Liu, 2017a,b).

For a given input sample X Rn×m, where n represents the number of samples, m represents the feature dimension of the sample. The feature mapping layer is composed of the combination of feature nodes. The feature nodes and feature mapping layer of broad learning system can be expressed as following:

di=φi(XWen+βen)    (1)
Dn=[d1,d2,......dn]    (2)

where φi is the selectable linear or non-linear activation function, Wen is the random weight, and Wen is the random bias. Wen and βen are usually optimized by sparse auto-coding algorithm. The enhancement node and enhancement layer of BLS can be denoted as:

ej=δj(DnWhm+βhm)    (3)
Em=[E1,E2,......Em]    (4)

where δj is the non-linear activation function, Whm is the random weight, and βhm is the random bias. Then, the feature mapping layer and the enhancement layer are concated and transferred to the output layer. Since Wen, Wen, Whm, and βhm remain unchanged in the training process, the objective function of BLS is:

W(||(YY)||22+λ2||W||22)    (5)

where W is the weight of the output layer of the BLS, Y is the label of X, ||(YY)||22 is used to control the minimization of training error, λ2||W||22 is used to prevent model overfitting, and λ is the regularization coefficient. Then, W can be obtained by seeking the pseudo-inverse of ridge regression:

W=G+Y    (6)
G+=limλ0 (λI+GTG)-1GT    (7)

where I is the identity matrix. Through the above steps, we constructed a complete Broad Learning System.

3.4.2. Deep-broad ensemble model

Three-dimension radiographic images contain complex pathological spatial information. However, the original broad learning system can only receive two-dimensional features as input, and the ability to extract complex image features is weak. Integrating the deep convolution module can effectively improve the feature extraction ability of the broad learning system, which can better the performance for its classification and recognition (Chen et al., 2018).

In this paper, we proposed a deep-broad ensemble model for early recognition of Alzheimer's disease based on the above ideas, which aims to maintain a considerable performance of classification, reduce the dependence of hardware and improve the efficiency of the model.

As shown in Figure 2, we used a convolution-pooling layer to initially extract the features of the original input. The size of the convolution kernel used was 7×7×7, and the size of the pooling module was 3×3×3.The backbone of the model is composed of a 3D residual convolution—feature mapping layer and a 3D residual convolution—enhancement layer. The 3D residual convolution—feature mapping layer can be divided into residual convolution module and feature mapping module. The former is composed of several 3D bottleneck convolution modules (He et al., 2016), which are used to extract shallow features of the input, and transform these features into feature vectors with a size of 256 through the global pooling, and then input into the feature mapping module for further processing; The 3D residual convolution enhancement layer also includes several 3D bottleneck convolution modules as the former to extract deeper features. The feature vectors are then input to the enhancement module through the global pooling, which is different from the input of the enhancement layer in the original BLS. Detailed parameters of the deep module used in the whole backbone model are shown in Table 1. Finally, feature mapping module and enhancement module are mapped to the output layer to produce classification results.

FIGURE 2
www.frontiersin.org

Figure 2. Structure of proposed deep-broad ensemble model.

For a Given the original MRI image input X, the output feature vector after the first convolution-pooling layer is:

Xbase=λconv-pool(X)    (8)

The feature vector of the output after the residual convolution module of the 3D residual convolution-feature mapping layer and the 3D residual convolution-enhancement layer can be denoted as:

Xd=λd(Xbase)    (9)
Xe=λe(λd(Xbase))    (10)

where λd() indicates the residual convolution module of 3D residual convolution-feature mapping layer, and λe() indicates the residual convolution module of 3D residual convolution-enhancement layer. According to formula (1)–(4), feature nodes, feature mapping modules, enhancement nodes, and enhancement mapping modules of Broad learning can be denoted as:

di=[φi(XdWen+βen)]    (11)
Dn=[d1,d2,......dn]    (12)
ej=[δj(XeWhm+βhm)]    (13)
Em=[E1,E2,......Em]    (14)

According to formula (6)–(7), the final classification output Y can be obtained. Thus, we constructed a 3D Convolution Broad Learning System based on 3D medical image input.

3.4.3. Model training

After data preprocessing, the original image was converted into a 224×224×128 matrix as the input of the model. The server we used for model training had a AMD EPYC 7543 32-core Processor, 90 GB of RAM, Nvidia GeForce RTX 3090 GPU with CUDA 11.1.

We set up two groups of experiments whose convolution module was non-pre-trained and pre-trained respectively for comparison. The dataset used for pre-training was ADNI training dataset (347 images). In the pre-training experiment, we set cross entropy as the loss function, with the learning rate of 0.0001, the optimizer of momentum-SGD, and the batch size of 4.

For the broad module in the model, the range of feature nodes was [500–4,000], the range of enhancement nodes is [100–1,000], and the range of sparsity coefficient is [0.4–0.7]. We used Pytorch 1.8 and Numpy to construct the program of the model presented in this article, and all programs and experiments were run in Python 3.8.

3.5. Model testing and analysis

For the three groups of models after training, we used the ADNI test dataset for testing. The model finally outputs the probability that an image belongs to one of these categories. The category with the highest probability was selected as the classification result. We calculated the final classification accuracy, Precision and Recall based on this result. In addition, We studied the stability of the model by modifying the hyperparameters.

3.6. Clinical interpretation of MRI

To compare the performance of our proposed model with that of an actual radiology reader, a board-certified nuclear medicine physician with several years of experience (HuanHua Wu, nuclear medicine) was invited to perform a discriminative analysis of 87 MRI images from the ADNI test dataset. In order to prevent data leakage, the reader can only obtain MRI image data and the number of the subject, and analyze them based on their professional experience. We will calculate the corresponding indicators based on this result.

4. Results

4.1. Demographics

As shown in Table 2, The dataset used in this study contained 434 MRI images from 86 patients, which contained three types of Alzheimer's symptoms: AD, patients with Alzheimer's disease; MCI, mild cognitive impairment; NC, normal person. Seventy-seven images were obtained from AD, 151 from NC, and 206 from MCI. Partial slice images of AD and NC cases in the dataset are shown in Figure 4. The average age of all patients was 75.44 years old (range from 55 to 90 years), including 73.58 years old for female (range from 55 to 90 years) and 77.20 for male (range from 57 to 89 years). The average age of AD groups was 75.32 years (range from 57 to 90 years), with the average age of 75.04 years for female (range from 64 to 90 years) and 75.71 years for male (range from 57 to 87 years). The average age of MCI groups was 74.62 years (range from 55 to 89 years), with the average age of 69.56 years for female (range from 55 to 82 years) and 77.45 years for male (range from 63 to 89 years). The average age of NC groups was 76.62 years (range from 70 to 88 years), with the average age of 76.08 years for female (range from 71 to 82 years) and 77.46 years for male (range from 70 to 88 years).

TABLE 2
www.frontiersin.org

Table 2. Demographics of ADNI dataset.

4.2. Result of training

The preprocessed dataset was divided into training set and test set in a ratio of 0.7:0.3. We trained models for AD/CN, AD/MCI, and MCI/NC tasks respectively. Accuracy (ACC), sensitivity (SEN), and F1-score were used to evaluate the performance of them, and the training time of each model was recorded.

As shown in Table 4, in the tests of AD/CN, AD/MCI, and MCI/NC, the average of accuracy for prediction were 90.97, 91.16, and 83.39%. SEN is 91, 94, 85%. F1-score is 91, 94, 84%. The above results indicated that the proposed model has good discrimination ability in AD/NC and AD/MCI tasks, but weak discrimination ability in MCI/NC tasks. The ROC curves of the proposed deep-broad ensemble model method trained on 70% of ADNI dataset were shown in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3. Receiver operating characteristic (ROC) curve of the proposed deep-broad ensemble model trained on 70% of 3YR 2T ADNI dataset and tested on the remaining 30% of ADNI dataset and independent test set. ROC curve labeled Alzheimer Disease (AD) represents the essential performance for distinguishing AD vs. all other cases. ROC curves for Mild Cognitive Impairment (MCI) and Normal Control (NC) are also reported for technical completeness.

Because the broad module of the model required different hyperparameters, We also verified the stability of the model based on different hyperparameters. Due to the huge range of hyperparameters (range of feature mapping nodes: 500–4,000, range of enhancement node: 100–1,000, range of sparsity coefficient: 0.4–0.7), we took several hyperparameters values as representative. As shown in Figures 4A–C, when the number of feature mapping nodes and the sparse coefficient increased, the model maintained good stability. When the number of enhancement nodes increases, the stability of the model is generally acceptable, excluding some unstable intervals.

FIGURE 4
www.frontiersin.org

Figure 4. The accuracy of the model under different hyperparameters (A) Feature nodes, (B) Enhancement nodes, (C) Sparsity coefficient.

In addition, We explored the influence of the pre-training for the deep module on the final classification performance of the model. As shown in Table 3, the non-pre-trained model performed better than the pre-trained model on the AD/NC and NC/MCI tasks (maximum accuracy of 92.65/90.57 and 84.68/76.44), while the two performed similarly on the AD/MCI tasks (maximum accuracy of 92.76/91.57).

TABLE 3
www.frontiersin.org

Table 3. Comparison of the proposed model and radiology readers.

Since the BLS itself had considerable ability of feature fitting, the deep module of the proposed model was mainly used to further extract complex space features of medical image, enhancing the feature extraction ability of BLS. Therefore, the pre-training of deep module was not decisive. The broad module is decisive in the fitting of image features. If the cost of pre-training process was removed, the training time of the model proposed can be further reduced, lowing the dependency of hardwares and improving the efficiency of the model.

4.3. Model interpretation: t-SNE plot

As shown in the Figure 5, We clustered the final features of the models in the three experiments respectively after dimension reduction by t-SNE. In AD/NC and AD/MCI classification experiments, the corresponding categories were almost pure, with only a small amount of mixing. In the NC/MCI experiment, the mixture of the two categories was more common. Therefore, we concluded that the proposed model is highly sensitive to AD categories, because most of the sample points were in the clustering of AD; we achieved high accuracy in both experiments.

FIGURE 5
www.frontiersin.org

Figure 5. The visualization of training set after dimension reduction with t-distributed stochastic neighbor embedding (t-SNE). Each point represents the final features of the proposed deep-broad ensemble model. (A) t-SNE for the AD/CN, (B) t-SNE for the AD/MCI, (C) t-SNE for the NC/MCI.

4.4. Comparison to previous works

We compared the proposed model with some previous works, which had developed several deep models in this task. Due to the lack of relevant hyperparameter reference, we set the number of training epochs to 40 for each work. As shown in Table 4, in the three tasks of AD/CN, AD/MCI and MCI/NC, the accuracy of the proposed model outperformed these works. The training time of the proposed model was much shorter than that of these deep models, because there was no need to update the weight parameters of the deep module of the proposed model. Therefore, compared with the previous works, the proposed model had a considerable optimization effect, less dependence on computer hardware, and was easier to deploy in the actual diagnosis process.

TABLE 4
www.frontiersin.org

Table 4. Comparison of the proposed model and previous related work.

4.5. Comparison to clinical Interpretations

As shown in Table 5, in the above tasks, the accuracy of radiology readers were 68.57, 75.00, 61.76%. Sensitivity were 38, 43, 66%. F1-score were 35, 29, 68%. Compared to radiology reader's work, the proposed model had better performance in the detection of ADNI datasets, which has statistical significance.

TABLE 5
www.frontiersin.org

Table 5. Comparison of the proposed model and radiology readers.

5. Discussion

The diagnosis and treatment of Alzheimer's disease is becoming an important medical issue for decades to come. Millions of Patients with Although Alzheimer's disease provides a rich data base for the improvement of diagnostic theories, it brings great work pressure and challenges to front-line doctors.

At present, computer science researchers had developed many detection models for radiographic images of the Alzheimer's disease. However, these current models almost consisted of single deep networks. This would lead to the problem that the models were highly dependent on hardwares, which was difficult to be popularized in non-urban areas where relevant hardware was lacking. To solve the above problems, we constructed a deep-broad ensemble model for radiographic images based on the novel BLS which has higher efficiency. Then, we trained and tested the model using the MRI dataset obtained from ADNI database, and calculated the corresponding accuracy, sensitivity and F1-score according to the results. We also compared the model with some previous work and results from radiology reader. The results demonstrates that compared with the previous work and the reader, the proposed model has better performance and greatly reduces the training time. Meanwhile, we studied the effect of the deep convolution module and the improved BLS module on the model. The results demonstrates that BLS was still the core of the model. The function of the deep module was to enhance the feature extraction capability of the BLS module, thus requiring no pre-training, which would greatly improve the performance of the model proposed in this paper. Our experiment had certain limitations. Firstly, the amount of data used in this study is still relatively small (434 images) due to the limited number of public medical image datasets for Alzheimer's disease currently available for research. Therefore, the robustness of the proposed model has not been verified on larger and more general data, which limits the application of our proposed model in real scenarios.

Second, BLS is a non-deep learning framework. Although its interpretability of it has been proven (Chen et al., 2018), BLS is not as widely used as deep neural network. Broad Learning System itself also has the limitation of relatively low accuracy, and its application in medical imaging and other fields lacks of universal reference. The hyperparameter setting of the proposed model relies on the previous research experience of machine learning researchers and lacks a better adjustment method (Gong et al., 2021).

In general, our experiment and research results demonstrate that our proposed deep-broad ensemble model method significantly reduces the training time while maintaining good detection performance. This makes our model play a referential role in practical medical image diagnosis and reduces the dependence on external hardware. With the opening of more medical image data, the model proposed in this paper can be better applied to first-line clinical diagnosis and provide reliable reference for doctors and medical image readers.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.

Ethics statement

The studies involving human participants were reviewed and approved by Academic Committee of Jinan University. The patients/participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author contributions

PM: conceptualization, methodology, validation, visualization, software, writing—original draft, review, and editing, and project administration. JW: supervision and project administration. ZZ: supervision and writing—review and editing. CC: conceptualization, resources, and supervision. Alzheimer's Disease Neuroimaging Initiative: data provision. JD: conceptualization, methodology, validation, writing—review and editing, supervision, project administration, and funding acquisition. All authors contributed to the article and approved the submitted version.

Funding

This work was supported in part by the National Key Research and Development Program of China under Grant No. 2018YFC2002500, in part by Guangdong Basic and Applied Basic Research Foundation under Grant No. 2021A1515011999, and in part by Guangdong Provincial Key Laboratory of Traditional Chinese Medicine Informatization under Grant No. 2021B1212040007. In addition, the authors thank the ADNI Committee for providing the MRI data.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Chen, C. P., and Liu, Z. (2017a). “Broad learning system: a new learning paradigm and system without going deep,” in 2017 32nd Youth Academic Annual Conference of Chinese Association of Automation (YAC) (IEEE), 1271–1276.

Google Scholar

Chen, C. P., and Liu, Z. (2017b). Broad learning system: an effective and efficient incremental learning system without the need for deep architecture. IEEE Trans. Neural Netw. Learn. Syst. 29, 10–24. doi: 10.1109/TNNLS.2017.2716952

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, C. P., Liu, Z., and Feng, S. (2018). Universal approximation capability of broad learning system and its structural variations. IEEE Trans. Neural Netw. Learn. Syst. 30, 1191–1204. doi: 10.1109/TNNLS.2018.2866622

PubMed Abstract | CrossRef Full Text | Google Scholar

Cummings, J. L., and Cole, G. (2002). Alzheimer disease. JAMA 287, 2335–2338. doi: 10.1001/jama.287.18.2335

PubMed Abstract | CrossRef Full Text | Google Scholar

Doi, K. (2007). Computer-aided diagnosis in medical imaging: historical review, current status and future potential. Comput. Med. Imaging Graph. 31, 198–211. doi: 10.1016/j.compmedimag.2007.02.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Fink, H. A., Linskens, E. J., Silverman, P. C., McCarten, J. R., Hemmy, L. S., Ouellette, J. M., et al. (2020). Accuracy of biomarker testing for neuropathologically defined Alzheimer disease in older adults with dementia: a systematic review. Ann. Internal Med. 172, 669–677. doi: 10.7326/M19-3888

PubMed Abstract | CrossRef Full Text | Google Scholar

Gong, X., Zhang, T., Chen, C. P., and Liu, Z. (2021). Research review for broad learning system: algorithms, theory, and applications. IEEE Trans. Cybernet. 52, 8922–8950. doi: 10.1109/TCYB.2021.3061094

PubMed Abstract | CrossRef Full Text | Google Scholar

Guan, Z., Kumar, R., Fung, Y. R., Wu, Y., and Fiterau, M. (2019). A comprehensive study of Alzheimer's disease classification using convolutional neural networks. arXiv preprint arXiv:1904.07950.

Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Las Vegas, NV: IEEE), 770–778.

PubMed Abstract | Google Scholar

Huang, G., Liu, Z., Van Der Maaten, L., and Weinberger, K. Q. (2017). “Densely connected convolutional networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 4700–4708.

Google Scholar

Huang, G.-B., Zhu, Q.-Y., and Siew, C.-K. (2006). Extreme learning machine: theory and applications. Neurocomputing 70, 489–501. doi: 10.1016/j.neucom.2005.12.126

CrossRef Full Text | Google Scholar

Jack, C. R., Petersen, R. C., Xu, Y. C., O'Brien, P. C., Smith, G. E., Ivnik, R. J., et al. (1999). Prediction of ad with MRI-based hippocampal volume in mild cognitive impairment. Neurology 52, 1397–1397. doi: 10.1212/WNL.52.7.1397

PubMed Abstract | CrossRef Full Text | Google Scholar

Johnson, K. A., Fox, N. C., Sperling, R. A., and Klunk, W. E. (2012). Brain imaging in Alzheimer disease. Cold Spring Harbor Perspect. Med. 2, a006213. doi: 10.1101/cshperspect.a006213

PubMed Abstract | CrossRef Full Text | Google Scholar

Korolev, S., Safiullin, A., Belyaev, M., and Dodonova, Y. (2017). “Residual and plain convolutional neural networks for 3D brain MRI classification,” in 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017) (IEEE), 835–838.

Google Scholar

Mayeux, R., and Sano, M. (1999). Treatment of Alzheimer's disease. N. Engl. J. Med. 341, 1670–1679.

Google Scholar

Mimura, M., and Yano, M. (2006). Memory impairment and awareness of memory deficits inearly-stage Alzheimer's disease. Rev. Neurosci. 17, 253–266. doi: 10.1515/REVNEURO.2006.17.1-2.253

PubMed Abstract | CrossRef Full Text | Google Scholar

Oh, K., Chung, Y.-C., Kim, K. W., Kim, W.-S., and Oh, I.-S. (2019a). Classification and visualization of Alzheimer's disease using volumetric convolutional neural network and transfer learning. Sci. Rep. 9, 1–16. doi: 10.1038/s41598-019-54548-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Pao, Y.-H., Park, G.-H., and Sobajic, D. J. (1994). Learning and generalization characteristics of the random vector functional-link net. Neurocomputing 6, 163–180. doi: 10.1212/WNL.0b013e3181cb3e25

PubMed Abstract | CrossRef Full Text | Google Scholar

Petersen, R. C., Aisen, P., Beckett, L. A., Donohue, M., Gamst, A., Harvey, D. J., et al. (2010). Alzheimer's disease neuroimaging initiative (ADNI): clinical characterization. Neurology 74, 201–209.

PubMed Abstract | Google Scholar

Prince, J. L., and Links, J. M. (2006). Medical Imaging Signals and Systems. Upper Saddle River, NJ: Pearson Prentice Hall.

Google Scholar

Rieke, J., Eitel, F., Weygandt, M., Haynes, J. -D., and Ritter, K. (2018). “Visualizing convolutional networks for MRI-based diagnosis of Alzheimer's disease,” in Machine Learning in Clinical Neuroimaging (Granada).

Google Scholar

Sanger, T. D. (1989). Optimal unsupervised learning in a single-layer linear feedforward neural network. Neural Netw. 2, 459–473.

Google Scholar

Scheltens, P., Blennow, K., Breteler, M., De Strooper, B., Frisoni, G., and Salloway, S. (2016). Van der flier WM: Alzheimer's disease. Lancet 388, 505–517. doi: 10.1016/S0140-6736(15)01124-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Simonyan, K., and Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556.

Google Scholar

Todd, S., Barr, S., Roberts, M., and Passmore, A. P. (2013). Survival in dementia and predictors of mortality: a review. Int. J. Geriatr. Psychiatry 28, 1109–1124. doi: 10.1002/gps.3946

PubMed Abstract | CrossRef Full Text | Google Scholar

Weiner, M. W., Veitch, D. P., Aisen, P. S., Beckett, L. A., Cairns, N. J., Green, R. C., et al. (2013). The Alzheimer's disease neuroimaging initiative: a review of papers published since its inception. Alzheimers Dement. 9, e111–e194. doi: 10.1016/j.jalz.2013.05.1769

PubMed Abstract | CrossRef Full Text | Google Scholar

Xie, S., Girshick, R., Dollár, P., Tu, Z., and He, K. (2017). “Aggregated residual transformations for deep neural networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Honolulu, HI: IEEE), 1492–1500.

Google Scholar

Zhang, H., Wu, C., Zhang, Z., Zhu, Y., Lin, H., Zhang, Z., et al. (2022). “ResNest: split-attention networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (New Orleans, LA: IEEE), 2736–2746.

Google Scholar

Keywords: deep-broad ensemble model, Alzheimer's disease, early detection, MRI, validation, efficiency

Citation: Ma P, Wang J, Zhou Z, Chen CLP, the Alzheimer's Disease Neuroimaging Initiative and Duan J (2023) Development and validation of a deep-broad ensemble model for early detection of Alzheimer's disease. Front. Neurosci. 17:1137557. doi: 10.3389/fnins.2023.1137557

Received: 04 January 2023; Accepted: 26 June 2023;
Published: 11 July 2023.

Edited by:

Xi Jiang, University of Electronic Science and Technology of China, China

Reviewed by:

Saneera Hemantha Kulathilake, Rajarata University of Sri Lanka, Sri Lanka
Ling Xing, Southwest University of Science and Technology, Luoyang, China

Copyright © 2023 Ma, Wang, Zhou, Chen, the Alzheimer's Disease Neuroimaging Initiative and Duan. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Zhiguo Zhou, emhpZ3VvemhvdSYjeDAwMDQwO2JpdC5lZHUuY24=; Junwei Duan, andkdWFuJiN4MDAwNDA7am51LmVkdS5jbg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.