- 1Department of Radiology, Gongli Hospital, School of Medicine, Shanghai University, Shanghai, China
- 2School of Life Sciences, Institute of Biomedical Engineering, Shanghai University, Shanghai, China
- 3School of Communication and Information Engineering, Shanghai University, Shanghai, China
Objectives We proposed a novel deep learning radiomics (DLR) method to distinguish cognitively normal adults at risk of Alzheimer’s disease (AD) from normal control based on T1-weighted structural MRI images.
Methods: In this study, we selected MRI data from the Alzheimer’s Disease Neuroimaging Initiative Database (ADNI), which included 417 cognitively normal adults. These subjects were divided into 181 individuals at risk of Alzheimer’s disease (preAD group) and 236 normal control individuals (NC group) according to standard uptake ratio >1.18 calculated by amyloid Positron Emission Tomography (PET). We further divided the preaAD group into APOE+ and APOE− subgroups according to whether APOE ε4 was positive or not. All data sets were divided into one training/validation group and one independent test group. The proposed DLR method included three steps: (1) the pre-training of basic deep learning (DL) models, (2) the extraction, selection and fusion of DLR features, and (3) classification. The support vector machine (SVM) was used as the classifier. In the comparative experiments, we compared our proposed DLR method with three existing models: hippocampal model, clinical model, and traditional radiomics model. Ten-fold cross-validation was performed with 100 time repetitions.
Results: The DLR method achieved the best classification performance between preAD and NC than other models with an accuracy of 89.85% ± 1.12%. In comparison, the accuracies of the other three models were 72.44% ± 1.37%, 82.00% ± 4.09% and 79.65% ± 2.21%. In addition, the DLR model also showed the best classification performance (85.45% ± 9.04% and 92.80% ± 2.61%) in the subgroup experiment.
Conclusion: The results showed that the DLR method provided a potentially clinical value to distinguish preAD from NC.
Introduction
Alzheimer’s disease (AD) is a neurodegenerative disease characterized by progressive cognitive decline (1). Due to the irreversibility of AD, it is critical to identify AD patients at an ultra-early stage. According to the latest A-T-N diagnosis criteria (2–4), individuals who showed obvious brain amyloid beta (Aβ+) deposition have entered the Alzheimer’s continuum and represented a high-risk of AD. This population could be defined as the preclinical AD group (PreAD) (5, 6).
So far, structural resonance imaging (MRI) have been widely used in the diagnosis of AD (7–11). For instance, previous studies have shown that patients with mild cognitive impairment(MCI) had increased hippocampal atrophy compared to normal control (NC) subjects (12). The atrophy of hippocampal and entorhinal cortex could also be used as an index to predict the conversion from MCI to AD (13).
Currently, artificial intelligence (AI) techniques based on MRI have frequently been used in the early diagnosis of AD. One typical AI application is radiomics. For example, Zhao et al. investigated hippocampal texture radiomics features as effective MRI biomarkers for AD and achieved an accuracy of 87.4% to distinguish AD and normal controls (NC) (14). Zhou and Shu et al. utilized MRI radiomics features to predict development of MCI to AD and achieved the accuracy of 78.4 and 80.7%, respectively (15, 16). Notably, Li et al. conducted an exploratory study to diagnosis preAD from NC based on radiomics multi-parameter MRI and obtained an average accuracy of 83.7% [T. (17)]. Although the feasibility of traditional radiomics methods has been proven, these methods could not be widely applied because of obvious shortcomings, such as manual extraction of regions of interest (ROIs) and hand-coding, which usually require complex manual operations. Therefore, an alternative method is required.
The deep learning radiomics (DLR) method may be the alternative (18, 19). This technique was able to mine the high dimension features of medical images automatically, and effectively address the shortage of hand-coding by radiomics. Recently, DLR has been used in brain tumor-related research and AD diagnosis (20, 21). For example, previous studies achieved good predictive performance of preoperative meningioma with an accuracy of 92.6% (22). Wang et al. extracted MRI-based DLR features to predict the prognosis of high-grade glioma (23). For AD diagonosis, early DLR-based methods always focused on pre-determined regions of interest prior to deep training, which may hamper diagnostic performance. For example, Khvostikov et al. and Li et al. trained DLR models based on pre-extracted hippocampal regions of MRI and other multimodal neuroimaging data (24, 25). Apart from the above, Basaia et al. used a single cross-sectional MRI scan and deep neural networks to automatically classify AD and MCI, with high accuracies of 98.2% between AD and NC, and of 74.9% from MCI to AD progression (26). Lee et al. also used DLR method for AD classification and achieved the accuracies of 95.35% and 98.74% on different datasets (27). However, there is no existing DLR model for preAD detection.
Therefore, in this study we hypothesized that the DLR method was useful in the diagnosis of PreAD. Considering the hippocampus volume has not been shrunk in AD early stage, we used MRI images of the whole brain for DLR classification. In addition, we also hypothesized that the DLR technique could achieve high classification accuracy in detecting subgroups of preAD from NC, such as APOE ε4+ individuals.
Materials and Methods
Figure 1 showed the overall framework of this study, which consisted of six steps: (1) enrolled subjects; (2) imaging preprocessing, including segmentation, normalization and smoothing; (3) basic deep learning (DL) model pre-training, in this step several DL models were pre-trained in order to get the best one for DLR feature extraction; (4) feature extraction and fusion; (5) classification; (6) comparative experiments.
Subjects
The data used in this study were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database.1 ADNI is a longitudinal, multicenter study to develop clinical, imaging, genetic and biochemical biomarkers for early detection and tracking of AD. The latest information is available at http://adni.loni.usc.edu/about/.
In this study, we collected 236 NC and 181 preAD data. Demographic data included age, sex, gender, education, neuropsychological assessment tests [Dementia Rating Scale (CDRSB) and Mini-Mental State Examination (MMSE)], Apolipoprotein E (APOE) ε4 and imaging information. T1-MRI and amyloid positron emission tomography (PET) images were selected for all subjects. The preAD group was defined as who standard uptake value ratio (SUVR) value of amyloid PET was >1.18 in whole cerebral cortex (17, 28). Whole cerebellum was used for reference when deriving SUVR. In addition, to validate our proposed DLR model, we enrolled 12 preAD individuals who converted into the MCI state. We selected MRI images in both baseline and MCI stages.
All subjects were divided into two groups, one training/validation group and one independent test group. The training/validation group was from ADNI 1, ADNI 2 and ADNI 3, including 212 NC and 162 preAD subjects. The test group was from ADNI Go, including 24 NC subjects and 19 preAD subjects.
Images Acquisition and Preprocessing
The image acquisition process was described in the ADNI website at http://adni.loni.usc.edu/about/. All MRI data have been evaluated by quality control (QC) at the Mayo Clinic Aging and Dementia Imaging Research Laboratory. The SUVR values of Amyloid PET were downloaded from the ADNI website directly.
The preprocessing of MRI images was performed by statistical parametric mapping (SPM12) software2 on MATLAB 2016b platform.3 First, MRI images were segmented into probabilistic gray matter (GM), white matter (WM), and cerebrospinal fluid (CSF); Then, each GM image was normalized into the Montreal Neurological Institute (MNI) space by diffeomorphic anatomical registration via exponentiated lie algebra, and smoothed using an 8-mm Gaussian-smoothing kernel. As a result, each image has a spatial resolution of 91 × 109 × 91 with a voxel size of 2 mm × 2 mm × 2 mm; Finally, in order to adapt and speed up the training of the deep learning model, 3D images were sliced from the axial direction into 91 single-channel images with the size of 91 × 109 to tile 2D images, and then resized into 224 × 224 for normalizing. Each 3D MRI image was tiled into a group of 2D images and resized into 224 × 224 pixels for subsequent DL model training.
The Proposed Deep Learning Radiomics Method
Figure 2 illustrates our proposed DLR method. The method consisted of three parts: (1) Basic DL model pre-training. We used six Convolutional Neural Networks (CNN) networks as candidate DL models and pre-trained them, respectively. After training, we selected one as the final DL model to obtain the DLR features according to the classification results. (2) Feature fusion. To obtain DLR features, we obtained DLR feature maps from the last convolutional layer of the final selected DL model, and extracted the maximum value of each feature map through global max pooling. These extracted features were defined as DLR features and combined with clinical features (sex, education, etc.) as input data for classification. (3) Classification. Based on the above features, the support vector machine (SVM) was used as the classifier to distinguish preAD from NC.
Training for Candidate Deep Learning Models
Six CNN models, including AlexNet, ZFNet, ResNet18, ResNet34, InceptionV3, and Xception, were applied in the training step to define the best training model.
• AlexNet: it is the first CNN network architecture that uses ReLU as the activation function, and uses interleaving pooling technology in CNN (29).
• ZF-Net: it is fine-tuned on the basis of AlexNet. It uses deconvolution to visually analyze the intermediate feature map of CNN and improves model performance by analyzing feature behavior (30).
• Inceptionv3: it improves the CNN model by using convolution decomposition and regularization (31).
• Xception: it improves Inception V3 by using depth wise separable convolution to replace the Inception module (32, 33).
• ResNet: it introduces new network features based on the previous traditional CNN network (34). Several ResNet subtypes were proposed according to different numbers of hidden layers, such as ResNet18, ResNet34, ResNet101, and so on.
As an example, Figure 3 showed the network structure of the ResNet34 model.
Figure 3. (A) The network structure of the ResNet34 model. “7 × 7” represents the size of the convolution kernel, “conv” represents convolution, “avg pool” represents average pooling, and “fc” represents fully connected layer. “64” means the number of channels, and “/2” means stride of 2. (B) Residual learning: a building block. x represents direct identity mapping, F(x) represents residual mapping, and F(x)+x is output.
During the raining step, the selected six models were trained in the training/validation group and tested in the test group. Guided by the test results, we optimized the DL model by tuning hyper parameters.
Classifier
We combined DLR features and clinical information (gender, education, age, etc.) as input data for classification. SVM was used as the classifier. As a classic supervised learning method, SVM has been widely used in statistical classification and regression analysis due to its ability to map vectors to a higher dimensional space that creates a maximum margin hyperplane to achieve high classification performance (35). In this study, we used the linear kernel function in SVM to detect classification reliability and generalization ability.
Comparative Experiments
To demonstrate the superiority of our proposed DLR method, we compared our model and three existing models in comparative experiments, including: (1) Clinical model: clinical characteristics included demographic data, neuropsychological cognitive assessment results, and APOE ε4 genotyping characteristics of all subjects. (2) Hippocampal model: the hippocampal volumes were used as inputs for the classification; (3) Traditional radiomics model: traditional radiomics features of were extracted for the classification. In this experiment, we extracted features by using the radiomics tool developed by Vallieres et al.4 We used brain DMN regions as ROIs and performed texture analysis on each input ROI using the “Texture Toolbox” in the Radiomics Toolbox. Feature extraction steps included wavelet bandpass filtering, isotropic resampling, Lloyd–Max quantization and feature calculation. The detailed extraction process of the radiomics features were described in the previous studies (36, 37).
Three comparative experiments were employed in this study: (1) NC vs. preAD; (2) NC vs. preAD APOE+; and (3) NC vs. preAD APOE−. Ten-fold cross-validation was performed with 100 time repetitions. We calculated accuracy, sensitivity, and specificity to evaluate the classification results. The mathematical expressions of the three indicators were as follows:
Longitudinal Study
The 12 individuals with longitudinal data were used to validate the proposed DLR model. Firstly, we calculated the probability value of SVM classifier, and defined it as the decision score; then we compared the decision scores in both baseline and MCI states in 12 individuals.
Statistical Analysis
In this study, we used two-sample t-tests or chi-square tests to compare demographic and clinical characteristics between the NC and preAD groups and between the APOE+ and APOE− subgroups. All statistical analyses were performed using SPSS version 22.0 software (SPSS Inc., Chicago, IL, United States) and performed in Matlab2019b (Mathworks Inc., Sherborn, MA, United States). A p-value < 0.05 was considered to be significantly different.
Results
Demographic Information
The results of demographic data were shown in Table 1. There was a significant difference in age and years of education between the preAD group and the NC group in the training/validation group (p < 0.001), and there was a difference in CDRSB and MMSE (CDRSB: p = 0.006, MMSE: p = 0.003), while no difference in gender between the two groups. There was no significant difference in gender, education level, CDRSB and MMSE between the preAD group and the NC group in the test group, whereas there was a difference in age (p = 0.03).
Pre-training for Candidate Deep Learning Models
Table 2 summarized the classification performance of six candidate DL models, including classification accuracy, sensitivity, and specificity. By comparing the results of the two groups, ResNet34 was selected to be the best model. Therefore, we chose the pre-training ResNet34 model and extracted DLR features for the next step.
Comparative Experiments
Normal Control vs. Preclinical Alzheimer’s Disease Group
Table 3 showed the classification results of the four models between NC and preAD groups. Among the four models, the DLR model showed the best classification performance in the test group, with the accuracy of 89.85% ± 1.12%, sensitivity of 94.74% ± 0.1%, and specificity of 85.98% ± 2.01%. The performance of the hippocampal model, traditional radiomics model, and clinical model were all significantly lower than DLR model, with the accuracies of 72.44% ± 1.37%, 82.00% ± 4.09% and 79.65% ± 2.21%, sensitivities of 42.68% ± 2.93%, 68.59% ± 8.35% and 82.754% ± 4.24%, specificities of 96.09% ± 1.31%, 92.62% ± 4.58% and 77.20% ± 2.61%, respectively.
Figure 4 presented the ROC curves of the four models. The mean AUCs (± SD) for the hippocampal model, traditional radiomics model, clinical model, and DLR model in were 0.691 ± 0.012, 0.806 ± 0.013, 0.800 ± 0.021 and 0.904 ± 0.014, respectively.
Normal Control vs. Preclinical Alzheimer’s Disease Subgroups
Table 4 showed the classification results between NC and preAD APOE+ groups. The accuracy, sensitivity and specificity of the DLR model in the test group were 92.80% ± 2.61%, 88.89% ± 0.01%, and 94.47% ± 3.72%. The performance of the hippocampal model, traditional radiomics model, and clinical model were all significantly lower than our proposed model, with the accuracies of 72.44%.
Table 5 showed the classification results between NC and preAD APOE− groups. The accuracy, sensitivity and specificity of the DLR model in the test group were 85.45 ± 9.04%, 90.40% ± 9.47%, and 83.10% ± 11.66%. The performance of the hippocampal model, traditional radiomics model, and clinical model were all significantly lower than our proposed model, with the accuracies of 63.36% ± 7.42%, 83.87% ± 3.04%, and 70.10% ± 3.50%. In Tables 3, 4, the bold values represented the classification performance of the our proposed method.
Figure 5 showed the ROC curves of the four models. The mean AUCs (± SD) for the hippocampal model, traditional radiomics model, clinical model and the best DLR model between NC and preAD APOE+ were 0.638 ± 0.061, 0.728 ± 0.024, 0.645 ± 0.041 and 0.917 ± 0.010, and between NC and preAD APOE− were 0.634 ± 0.075, 0.823 ± 0.041, 0.679 ± 0.042, and 0.868 ± 0.011, respectively.
Figure 5. ROC curves of the four models between NC and preAD APOE+ groups (left) and between NC and preAD APOE– groups (right).
Longitudinal Study
Figure 6 showed the results of the longitudinal study. The decision scores had a slight upward trend from the PreAD baseline to the MCI stage. The results showed that our model also had a great prediction performance.
Discussion
Currently, DLR is the hot spot and focus of current imaging development. In view of its superiority in disease diagnosis, DLR methods have been successfully applied in tumor genotype prediction, preoperative analysis, prognosis evaluation, and cancer diagnosis, etc., but DLR research for neurological diseases remained lacking. In this study, we proposed a DLR model to distinguish cognitively normal adults at risk of Alzheimer’s disease from normal control based on T1-weighted structural MRI images. Compared with other traditional models, such as hippocampal model, clinical model or traditional radiomics model, our proposed DLR model achieved best classification results.
In the comparative experiments, the DLR method achieved the highest accuracy in both training/validation group (99.40% ± 3.23%) and separate test group (89.85% ± 1.12%). Therefore, we proved the robustness of the DLR model.
Currently, several studies have investigated the classification between preAD and NC by using machine learning or traditional quantitative methods. For example, Ding et al. distinguished preAD from NC by investigating the coupling relationship between glucose and oxygen metabolism from hybrid PET/MRI, with an AUC of 0.787 (38). Li et al. used a voxel-based SSM/PCA method to analyze fluorodeoxyglucose-PET (FDG-PET) images with AUC of 0.815 (39), Li et al. conducted an exploratory study for identifying preAD based on radiomics analysis of MRI and obtained an average accuracy of 83.7% [T. (17)]. In comparison to previous studies, our DLR model achieved the best classification results. The reason can be explained as following: (1) the DLR method can directly extract high-throughput image features from CNN. Since it does not involve additional feature extraction operations, it will not bring additional errors; (2) the results of traditional methods were usually influenced by individual factors and imaging machine parameters; while the DLR method combined DLR image features and clinical information, which partly solved the problems of individual heterogeneity.
To demonstrate the robustness of the proposed DLR model, we performed experiments in the APOE ε4 subgroup analysis. Notably, cerebral amyloid deposition is also affected by the ApoE ε4 genotype (40). Higher levels of amyloid accumulation were observed in SCD subjects with ApoE ε4 carriers than noncarriers (41, 42). Therefore, we proposed to add ApoE ε4 genotype features to further validate the accuracy of the model. Notably, the DLR model achieved better classification results between NC vs. preAD APOE+ (92.80% ± 2.61%) than the two other experiments (89.85% ± 1.12% and 85.45% ± 9.04%). The high sensitivity (88.89% ± 0.01%) and specificity (94.47% ± 3.72%) results also showed that the DLR model was very powerful in identifying cognitively normal adults at risk of Alzheimer’s disease.
Although the DLR method could distinguish preAD from NC, it still had some limitations. First, more data was still needed to verify the generality and robustness of the proposed method. In this study, subjects were collected only from the ADNI database. Whether our model was powerful for other racial populations need further exploration. Secondly, we only compared six DL models. Although the Resnet34 model achieved good classification performance, it was unknown whether other DL models beyond the six were more suitable. In addition, we used the whole brain MRI image to train the DLR models in this study. However, future studies were required to explore whether DLR models based on the hippocampus or entorhinal cortex instead were more effective. Furthermore, in this study, 2D DLR models were employed. However, whether 3D DLR models could achieve better classification performances need further exploration. Finally, the proposed DLR model was based on T1-MRI images. It may be possible to improve the classification performance of DLR by combining other imaging modals, such as FDG PET, amyloid PET and tau PET images.
Conclusion
We proposed a DLR method based on T1-MRI images to discriminate preAD and NC. The results demonstrated that our proposed DLR method could improve diagnostic performance. The DLR method had potentials for clinical applications in the future.
Data Availability Statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Ethics Statement
Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.
Author Contributions
JJ conceived and designed the experiments, analyzed and interpreted the data, and wrote the manuscript. JZ performed the experiments, interpreted the data, and wrote the manuscript. ZL and LL performed the experiments and wrote the manuscript. BH conceived and designed the experiments, provided research funding, analyzed and interpreted the data, and reviewed the manuscript. All authors read and approved the final version of the article for publication.
Funding
This article was supported by grants from the Shanghai Pudong New Area Health System Leading Personnel Training Program (PWRl2017-04) and Discipline Construction of Pudong New Area Health Committee (PWGw2020-01).
Conflict of Interest
The authors declare that this study was conducted without any commercial or financial relationships that could be construed as potential conflicts of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
Data collection and sharing for this project were funded by the Alzheimer’s Disease Neuroimaging Initiative (ADNI; National Institutes of Health Grant U01 AG024904) and DODADNI (Department of Defense, award number W81XWH-12–2-0012). ADNI is funded by the National Institute of Aging and the National Institute of Biomedical Imaging and Bioengineering and through generous contributions from the following: AbbVie, Alzheimer’s Association; Alzheimer’s Drug Discovery Foundation; Araclon Biotech; BioClinica Inc; Biogen; Bristol-Myers Squibb Company; CereSpir Inc.; Eisai Inc.; Elan Pharmaceuticals Inc; Eli Lilly and Company; EuroImmun; F.Hoffmann-La Roche Ltd. and its affiliated company Genentech Inc; Fujirebio; GE Healthcare; IXICO Ltd.; Janssen Alzheimer Immunotherapy Research & Development, LLC.; Johnson & Johnson Pharmaceutical Research & Development LLC.; Lumosity; Lundbeck; Merck & Co., Inc.; Meso Scale Diagnostics, LLC; NeuroRx Research; Neu-rotrack Technologies; Novartis Pharmaceuticals Corporation; Pfizer Inc; Piramal Imaging; Servier; Takeda Pharmaceutical Company; and Transition Therapeutics. The Canadian Institutes of Health Research is providing funds to support ADNI clinical sites in Canada.
Footnotes
- ^ https://adni.loni.usc.edu/wp-content/uploads/how_to_apply/ADNI_Acknowledgement_List.pdf
- ^ https://www.fil.ion.ucl.ac.uk/spm/software/spm12/
- ^ https://www.mathworks.com/products/matlab.html
- ^ https://github.com/mvallieres/radiomics
References
1. Laurent C, Buée L, Blum D. Tau and neuroinflammation: what impact for alzheimer’s disease and tauopathies? Biomed J. (2018) 41:21–33. doi: 10.1016/j.bj.2018.01.003
2. Dubois B, Hampel H, Feldman H, Scheltens P, Aisen P, Andrieu S, et al. Preclinical alzheimer’s disease: definition, natural history, and diagnostic criteria. Alzheimers Dement. (2016) 12:292–323. doi: 10.1016/j.jalz.2016.02.002
3. Jack C, Bennett D, Blennow K, Carrillo M, Dunn B, Haeberlein S, et al. NIA-AA Research Framework: Toward a biological definition of Alzheimer’s disease. Alzheimers & Dementia (2018) 14:535–62. doi: 10.1016/j.jalz.2018.02.018
4. Li T-R, Yang Q, Hu X, Han Y. Biomarkers and tools for predicting alzheimer’s disease at the preclinical stage. Curr Neuropharmacol. (2022) 20:713–37. doi: 10.2174/1570159X19666210524153901
5. Alzheimer’s Association. 2017 Alzheimer’s disease facts and figures. Alzheimers Dement. (2017) 13:325–73. doi: 10.1016/j.jalz.2017.02.001
6. Jessen F, Amariglio RE, Buckley RF, van der Flier WM, Han Y, Molinuevo JL, et al. The characterisation of subjective cognitive decline. Lancet Neurol. (2020) 19:271–8. doi: 10.1016/S1474-4422(19)30368-0
7. De Santi S, de Leon M, Rusinek H, Convit A, Tarshish C, Roche A, et al. Hippocampal formation glucose metabolism and volume losses in MCI and AD. Neurobiol Aging. (2001) 22:529–39. doi: 10.1016/S0197-4580(01)00230-5
8. Gyasi Y, Pang Y, Li X, Gu J, Cheng X, Liu J, et al. Biological applications of near infrared fluorescence dye probes in monitoring alzheimer’s disease. Eur J Med Chem. (2020) 187:111982. doi: 10.1016/j.ejmech.2019.111982
9. Jagust W. Imaging the evolution and pathophysiology of alzheimer disease. Nat Rev Neurosci. (2018) 19:687–700. doi: 10.1038/s41583-018-0067-3
10. Johnson K, Fox N, Sperling R, Klunk W. Brain imaging in alzheimer disease. Cold Spring Harb Perspect Med. (2012) 2:a006213. doi: 10.1101/cshperspect.a006213
11. Suk H, Lee S, Shen D, Alzheimer’s Disease Neuroimaging Initiative. Hierarchical feature representation and multimodal fusion with deep learning for AD/MCI diagnosis. Neuroimage. (2014) 101:569–82. doi: 10.1016/j.neuroimage.2014.06.077
12. Dickerson B, Salat D, Greve D, Chua E, Rand-Giovannetti E, Rentz D, et al. Increased hippocampal activation in mild cognitive impairment compared to normal aging and AD. Neurology. (2005) 65:404–11. doi: 10.1212/01.wnl.0000171450.97464.49
13. Devanand D, Pradhaban G, Liu X, Khandji A, De Santi S, Segal S, et al. Hippocampal and entorhinal atrophy in mild cognitive impairment—prediction of alzheimer disease. Neurology. (2007) 68:828–36. doi: 10.1212/01.wnl.0000256697.20968.d7
14. Zhao K, Ding Y, Wang P, Dou X, Zhou B, Yao H, et al. Early classification of alzheimer’s disease using hippocampal texture from structural MRI. In: A Krol, B Gimi editors. Conference of the SPIE Medical Imaging. (Vol. 10137), Orlando, FA: (2017). doi: 10.1117/12.2254198
15. Shu Z-Y, Mao D-W, Xu Y, Shao Y, Pang P-P, Gong X-Y. Prediction of the progression from mild cognitive impairment to Alzheimer’s disease using a radiomics-integrated model. Ther Adv Neurol Disord. (2021) 14:17562864211029552. doi: 10.1177/17562864211029551
16. Zhou H, Jiang J, Lu J, Wang M, Zhang H, Zuo C, et al. Dual-model radiomic biomarkers predict development of mild cognitive impairment progression to alzheimer’s disease. Front Neurosci. (2019) 12:1045. doi: 10.3389/fnins.2018.01045
17. Li T, Wu Y, Jiang J, Lin H, Han C, Jiang J, et al. Radiomics analysis of magnetic resonance imaging facilitates the identification of preclinical alzheimer’s disease: an exploratory study. Front Cell Dev Biol. (2020) 8:605734. doi: 10.3389/fcell.2020.605734
18. Liu Z, Wang S, Dong D, Wei J, Fang C, Zhou X, et al. The applications of radiomics in precision diagnosis and treatment of oncology: opportunities and challenges. Theranostics. (2019) 9:1303–22. doi: 10.7150/thno.30309
19. Suzuki K. Overview of deep learning in medical imaging. Radiol Phys Technol. (2017) 10:257–73. doi: 10.1007/s12194-017-0406-5
20. Li Y, Wei D, Liu X, Fan X, Wang K, Li S, et al. Molecular subtyping of diffuse gliomas using magnetic resonance imaging: comparison and correlation between radiomics and deep learning. Eur Radiol. (2022) 32:747–58. doi: 10.1007/s00330-021-08237-6
21. Park J, Kickingereder P, Kim H. Radiomics and deep learning from research to clinical workflow: neuro-oncologic imaging. Korean J Radiol. (2020) 21:1126–37. doi: 10.3348/kjr.2019.0847
22. Yang L, Xu P, Zhang Y, Cui N, Wang M, Peng M, et al. A deep learning radiomics model may help to improve the prediction performance of preoperative grading in meningioma. Neuroradiology. (2022). doi: 10.1007/s00234-022-02894-0
23. Wang Y, Shao Q, Luo S, Fu R. Development of a nomograph integrating radiomics and deep features based on MRI to predict the prognosis of high grade Gliomas. Mathe Biosci Eng. (2021) 18:8084–95. doi: 10.3934/mbe.2021401
24. Khvostikov A, Aderghal K, Benois-Pineau J, Krylov A, Catheline G. 3D CNN-based classification using sMRI and MD-DTI images for alzheimer disease studies. arXiv [Preprint] (2018). doi: 10.48550/ARXIV.1801.05968
25. Li H, Habes M, Fan Y. Deep ordinal ranking for multi-category diagnosis of alzheimer’s disease using hippocampal MRI data. arXiv [Preprint] (2017). doi: 10.48550/ARXIV.1709.01599
26. Basaia S, Agosta F, Wagner L, Canu E, Magnani G, Santangelo R, et al. Automated classification of alzheimer’s disease and mild cognitive impairment using a single MRI and deep neural networks. Neuroimage Clin. (2019) 21:101645. doi: 10.1016/j.nicl.2018.101645
27. Lee B, Ellahi W, Choi J. Using deep CNN with data permutation scheme for classification of alzheimer’s disease in structural magnetic resonance imaging (sMRI). Ieice Trans Inform Syst. (2019) E102D:1384–95. doi: 10.1587/transinf.2018EDP7393
28. Fakhry-Darian D, Patel NH, Khan S, Barwick T, Win Z. Optimisation and usefulness of quantitative analysis of 18 F-florbetapir pet. Br J Radiol. (2019) 92:20181020. doi: 10.1259/bjr.20181020
29. Krizhevsky A, Sutskever I, Hinton G. ImageNet classification with deep convolutional neural networks. Commun Acm. (2017) 60:84–90. doi: 10.1145/3065386
30. Zeiler M, Fergus R. Visualizing and understanding convolutional networks. In: D Fleet, T Pajdla, B Schiele, T Tuytelaars editors. Proceedings of the Computer Vision, ECCV 2014 - 13th European Conference. (Vol. 8689), Cham: (2014). p. 818–33. doi: 10.1007/978-3-319-10590-1_53
31. Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z, Ieee. Rethinking the inception architecture for computer vision. Conference of the Computer Vision and Pattern Recognition 2016. Las Vegas, NV: (2016). p. 2818–26. doi: 10.1109/CVPR.2016.308
32. Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. In: F Bach, D Blei editors. Proceedings of the 32nd International Conference on International Conference on Machine Learning. (Vol. Vol. 37), Stroudsburg, PA: (2015). p. 448–56.
33. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, et al. Going deeper with convolutions. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Boston, MA: (2015). p. 1–9. doi: 10.1109/cvpr.2015.7298594
34. He K, Zhang X, Ren S, Sun J, Ieee. Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas, NV: (2016). p. 770–8. doi: 10.1109/CVPR.2016.90
35. Amari S, Wu S. Improving support vector machine classifiers by modifying kernel functions. Neural Netw. (1999) 12:783–9. doi: 10.1016/S0893-6080(99)00032-5
36. Gillies R, Kinahan P, Hricak H. Radiomics: images are more than pictures. they are data. Radiology. (2016) 278:563–77. doi: 10.1148/radiol.2015151169
37. Kumar V, Gu Y, Basu S, Berglund A, Eschrich S, Schabath M, et al. Radiomics: the process and the challenges. Magn Reson Imaging. (2012) 30:1234–48. doi: 10.1016/j.mri.2012.06.010
38. Ding C, Du W, Zhang Q, Wang L, Han Y, Jiang J. Coupling relationship between glucose and oxygen metabolisms to differentiate preclinical alzheimer’s disease and normal individuals. Hum Brain Mapp. (2021) 42:5051–62. doi: 10.1002/hbm.25599
39. Li T, Dong Q, Jiang X, Kang G, Li X, Xie Y, et al. Exploring brain glucose metabolic patterns in cognitively normal adults at risk of alzheimer’s disease: a cross-validation study with Chinese and ADNI cohorts. Neuroimage-Clin. (2022) 33:102900. doi: 10.1016/j.nicl.2021.102900
40. Moreno-Grau S, Rodriguez-Gomez O, Sanabria A, Perez-Cordon A, Sanchez-Ruiz D, Abdelnour C, et al. Exploring APOE genotype effects on alzheimer’s disease risk and amyloid beta burden in individuals with subjective cognitive decline: The fundacioace healthy brain initiative (FACEHBI) study baseline results. Alzheimers Dement. (2018) 14:634–43. doi: 10.1016/j.jalz.2017.10.005
41. Risacher S, Kim S, Nho K, Foroud T, Shen L, Petersen R, et al. APOE effect on alzheimer’s disease biomarkers in older adults with significant memory concern. Alzheimers Dement. (2015) 11:1417–29. doi: 10.1016/j.jalz.2015.03.003
Keywords: deep learning radiomic, Alzheimer’s disease, magnetic resonance imaging, support vector machine, artificial intelligence
Citation: Jiang J, Zhang J, Li Z, Li L, Huang B and Alzheimer’s Disease Neuroimaging Initiative (2022) Using Deep Learning Radiomics to Distinguish Cognitively Normal Adults at Risk of Alzheimer’s Disease From Normal Control: An Exploratory Study Based on Structural MRI. Front. Med. 9:894726. doi: 10.3389/fmed.2022.894726
Received: 12 March 2022; Accepted: 28 March 2022;
Published: 21 April 2022.
Edited by:
Md.Mohaimenul Islam, Aesop Technology, TaiwanCopyright © 2022 Jiang, Zhang, Li, Li, Huang and Alzheimer’s Disease Neuroimaging Initiative. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Bingcang Huang, aGJjOTIwOUBzaW5hLmNvbQ==
†Data used in preparation of this manuscript were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu). As such, the investigators within the ADNI contributed to the design and implementation of ADNI and/or provided data but did not participate in the analysis or writing of this report. A complete listing of ADNI investigators can be found at: https://adni.loni.usc.edu/wp-content/uploads/how_to_apply/ADNI_Acknowledgement_List.pdf