- 1School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, China
- 2School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, China
- 3Key Laboratory for NeuroInformation of the Ministry of Education, School of Life Sciences and Technology, University of Electronic Science and Technology of China, Chengdu, China
- 4School of Physics, University of Electronic Science and Technology of China, Chengdu, China
- 5School of Electrical Engineering and Electronic Information, Xihua University, Chengdu, China
- 6Center for Language and Brain, Shenzhen Institute of Neuroscience, Shenzhen, China
Autism spectrum disorder (ASD) is a range of neurodevelopmental disorders with behavioral and cognitive impairment and brings huge burdens to the patients’ families and the society. To accurately identify patients with ASD from typical controls is important for early detection and early intervention. However, almost all the current existing classification methods for ASD based on structural MRI (sMRI) mainly utilize the independent local morphological features and do not consider the covariance patterns of these features between regions. In this study, by combining the convolutional neural network (CNN) and individual structural covariance network, we proposed a new framework to classify ASD patients with sMRI data from the ABIDE consortium. Moreover, gradient-weighted class activation mapping (Grad-CAM) was applied to characterize the weight of features contributing to the classification. The experimental results showed that our proposed method outperforms the currently used methods for classifying ASD patients with the ABIDE data and achieves a high classification accuracy of 71.8% across different sites. Furthermore, the discriminative features were found to be mainly located in the prefrontal cortex and cerebellum, which may be the early biomarkers for the diagnosis of ASD. Our study demonstrated that CNN is an effective tool to build the framework for the diagnosis of ASD with individual structural covariance brain network.
Introduction
Autism spectrum disorder (ASD) encompasses a range of neurodevelopmental disorders. The core symptoms of ASD comprise abnormal emotional regulation and social interactions, restricted interest, repetitive behaviors, and hypo-/hyperreactivity to sensory stimuli (Guze, 1995). Many individuals with autism spectrum disorder usually exhibit impairments in learning, development, control, and interaction, as well as some daily life skills. ASD causes heavy economic burden for the patients’ families and the society. It is urgent to establish an early and accurate diagnosis framework to identify ASD patients from typical controls (TC). Recently, noninvasive and in vivo neuroimaging techniques have become an area of intense investigation to explore the auxiliary diagnostic for ASD.
A variety of neuroimaging modalities, such as structural MRI (sMRI), diffusion MRI, functional MRI, magnetoencephalography, electroencephalography, and electrocorticography, are widely adopted to uncover patterns in both brain anatomical structure and function. Structural MRI provides abundant measures to delineate the structural properties of the brain (Wang et al., 2017, 2018, 2019; Wu et al., 2017; Xu et al., 2019). Although there were many controversial findings regarding brain structural changes in individuals with autism (Chen et al., 2011), to identify ASD based on sMRI still received a great deal of attention from researchers. Wang et al. (2016) linearly projected gray matter (GM) and white matter (WM) features extracted from sMRI onto a canonical space to maximize their correlations. With the projected GM and WM as features, they achieved the classification accuracy of 75.4% with the data scanned at New York University (NYU) Langone Medical Center from ABIDE. Recently, the classification accuracy for ASD is up to 90.39%, which is obtained with sMRI constrained to the same single-site dataset (Kong et al., 2019). Although previous studies achieved high classification accuracy, almost all these results were obtained from a single site, and reproducibility and generation in a multisite remain uncertain (Button et al., 2013).
It is well accepted that multisite datasets can represent greater variance of disease and control samples to establish more stable generalization models for replication across different sites, participants, imaging parameters, and analysis methods (Nielsen et al., 2013). Thus, there are some studies focused on larger sample sizes from a multisite. However, the classification accuracy drops significantly due to the complexity and heterogeneity of ASD (Nielsen et al., 2013). In a study by Zheng et al. (2018), the elastic network was utilized to quantify corticocortical similarity based on seven morphological features on 132 selected subjects from four independent sites of the ABIDE dataset. Although the classification performance of this study was 78.63%, the samples used in this study only cover a small portion of the ABIDE dataset. Additionally, there are few other studies that use larger datasets from ABIDE but with low accuracy. For instance, Haar et al. (2016) used linear and quadratic (nonlinear) discriminant analyses to classify ASD and control subjects based on structural measures of the quality-controlled samples from ABIDE, but the accuracies were only 56 and 60% when using subcortical volumes or cortical thickness measures. Katuwal et al. (2015) used random forest (RF), support vector machine (SVM), and gradient boosting machine (GBM) to classify 373 ASD from 361 TC male subjects from the ABIDE database and obtained 60% classification accuracy. The accuracy was further improved to 67% when IQ and age information were added to morphometric features. Although the existing approaches can obtain high classification accuracy based on sMRI measures in a single site or with a small number of subjects, an acceptable method to achieve high classification accuracy across different sites with different scanning paradigms is still needed.
The low classification accuracy for different sites may mainly result from the following reasons. First, the data collected from different sites expand the variances of structural measures, which increases the difficulty in learning high accuracy classifiers with such data. Second, the brain is an integrative and dynamic system for information processing between brain regions (Vértes and Bullmore, 2014), yet most of the existing methods only extract independent local morphological features of different brain areas with sMRI and did not consider the interregional morphological covariant relationships. Third, although some deep neural network classifiers were used in ASD/TC classification by transforming the features to a one-dimensional vector followed by features selection algorithm, the classification results are hard to interpret in the absence of the contributions of the classification features leading to lack of clinical significance. In view of the abovementioned problems, to further explore an efficient classification method is essential to establish the ASD diagnosis model. Here, we combine a deep learning classifier and gradient-weighted class activation mapping (Grad-CAM) (Selvaraju et al., 2020) based on morphological covariance brain networks to identify ASD patients from TC with all the ABIDE dataset. We first constructed the individual-level morphological covariance brain networks, and the interregional morphological covariance values were used as the input feature for the classifier. Next, the convolutional neural network (CNN) classifier with Grad-CAM is applied to differentiate ASD from TC and to identify features contributing the largest for the classification in our framework.
Materials and Methods
The ABIDE Dataset
Data used in this study are accessed from a large open access data repository, Autism Brain Imaging Data Exchange I (ABIDE1), which came from 17 international sites with no prior coordination (Di Martino et al., 2014). ABIDE includes structural MRI, corresponding rs-fMRI, and phenotype information for individuals with ASD and TC and allows for replication, secondary analyses, and discovery efforts. Although all data in ABIDE were collected with 3 T scanners, the sequence parameters as well as the type of scanner varied across sites. In this paper, we used the structural MR images of 518 ASD patients and 567 age-matched normal controls (ages 7–64 years, median 14.7 years across groups) aggregated from all 17 international sites. The key phenotypical information is summarized in Table 1. As seen from Table 1, the variation in age range across samples varied greatly, and most of the ASD subjects are male with 25% of the sites excluding females by design.
Table 1. Demographic information of subjects with autism spectrum disorder (ASD) and typical controls (TC).
Data Preprocessing
All structural MR images used in our work were preprocessed by the deformable medical image registration toolbox—DRAMMS (Ou et al., 2011). DRAMMS is a software package designed for 2D-to-2D and 3D-to-3D deformable medical image registration tasks. The preprocessing procedures in our study include cross-subject registration, motion correction, intensity normalization, and skull stripping. Especially, T1W MRI images in different sites were registered to the SRI24 atlas (Rohlfing et al., 2010) for the morphological covariance brain network mapping.
Individual-Level Morphological Covariance Brain Networks
The morphological features of the human brain have long been characterized by structural MRI. In our study, the individual-level morphological covariance brain network (Wang et al., 2016) is used to estimate interregional structural connectivity to characterize the interregional morphological relationship. The detailed construction procedures were described below. First, after preprocessing, the structural T1 images were segmented into cerebrospinal fluid (CSF), WM, and GM by the multiplicative intrinsic component optimization (MICO) method (Li et al., 2014). Next, a GM volume map was obtained for each participant in template space. Second, the large-scale morphological covariance brain network for each participant was constructed based on their GM volume images according to a previous study (Wang et al., 2016). A brain network is usually comprised of a collection of nodes and edges, wherein the network nodes are defined as different regions in SRI24 atlases and the edge is defined as the interregional similarity in the distribution of the regional GM volume. The SRI24 atlas (Rohlfing et al., 2010) parcellates the whole brain into 116 subregions and 58 subregions in each hemisphere. Because of low signal-to-noise ratio and blank values of gray matter volume in Vermis during network analysis, we excluded eight areas in the Vermis (the cerebellar Vermis labeled from 109 to 116) to ensure the reliability of our study. Finally, a 108 × 108 matrix was obtained for each subject for further analyses. To be specific, the edge of the individual network is calculated as follows: the kernel density estimation (KDE) (Rosenblatt, 1956) is firstly used to estimate the probability density function (PDF) of the extracted GM volume values as Eq. (1).
where P(i) is the PDF of the i-th brain area, N is the total number of regions, and fh(i) is the kernel density of the i-th area defined as (Wang et al., 2016):
where K(.) is a non-negative function that integrates to one and has mean zero, h > 0 is a smoothing parameter called the bandwidth, and v(i) is the GM volume value of the i-th ROI.
The variation of the KL divergence (KLD) is calculated subsequently from the above PDFs as Eq. (3):
where P and Q are the PDFs of different ROI. The network edge is formally defined as the structural connectivity between two regions and is quantified by a KL divergence-based similarity (KLS) measure (Kong et al., 2014) with the calculated variation of KLD. Thus, the similarity matrix can be defined as:
For more details of the calculation process, refer to the previous literature (Wang et al., 2016).
Convolutional Neural Network Classifier
Deep CNNs have led to a series of breakthroughs for image classification (Liu et al., 2017, 2018). The deep residual networks (ResNet) (He et al., 2016) have recently achieved state-of-the-art on challenging computer vision tasks, which consists of an ensemble of basic residual unit. According to the study of He et al. (2016), ResNet tries to learn both local and global features via skip connections combining different levels to overcome the incapability of integrating different level features found in plain networks, as seen in Figure 1A. Here, X is the input of the residual unit, and F(X) denotes the residue mapping of the stacked convolution layers. The formulation of F(X) + X can be realized by feedforward neural networks with shortcut connections. Thus, a simple identity mapping directly connects the input and output layers by using ResNet with the skip connection. In our work, five bottlenecks are used to perform the classification, and the detailed architecture of ResNet used in our work is shown in Figure 1B. The binary cross-entropy (BCE) cost function is used as the loss function of our work in the residual learning, which is defined as follows:
Figure 1. Residual learning and architectures for our work. (A) The formulation of F(X) + X can be realized by feedforward neural networks with shortcut connections. (B) Architectures for our work. The downsampling is performed by Conv_1, Conv_2, Conv_3, Conv_4, and Conv_5 with a stride of 2.
where y is the ground-truth label, and is the prediction result of our work.
Grad-CAM
The Grad-CAM is able to produce “visual explanations” for decisions from a large class of CNN-based models and makes them more transparent (Selvaraju et al., 2017, 2020). It uses the gradient information flowing into the last convolutional layer of CNN to assign importance values to each neuron for a particular decision of interest to avoid the model structure modification and refrainment to keep both interpretability and accuracy (Selvaraju et al., 2020). Also, the produced localization map highlights the important regions in the image for predicting the concept. It is an important tool for users to evaluate and place trust in classification systems (Selvaraju et al., 2020). In this work, we only focus on the explaining output layer decisions by identifying the contributions of the classification features to help researchers focus on the highlighted regions and trust the model. The gradient of the score yc for class c is computed with respect to feature map activations Ak of a convolutional layer. In the study by Selvaraju et al. (2020), the neuron importance weights are defined as:
where is the feature map activations Ak indexed by i and j.
Then, a ReLU function is applied to a weighted combination of forward activation maps to obtain the Grad-CAM (Selvaraju et al., 2020):
Finally, the heat map highlighting the regions with a positive influence on ASD/TC classification is obtained by upsampling .
Implementation
An overview of the proposed ASD/TC classification framework in our work is shown in Figure 2. We first constructed the individual-level morphological covariance brain network according to the SRI24 atlas as the input images for the classification. Subsequently, the ResNet network is used to perform the classification. Meanwhile, the importance value to each neuron is obtained from Grad-CAM to explain model decisions. Specially, our implementation for ASD/TC classification based on the morphological covariance brain networks performs the following practice. The size of input image for ResNet is 108 × 108, and the crop is not needed to be sampled from an image or its horizontal flip. We adopt batch normalization (Ioffe and Szegedy, 2015) right after each convolution and before activation based on a previous study (Ioffe and Szegedy, 2015). Moreover, we do not use dropout (Hinton et al., 2012) which is the practice of a previous study (Ioffe and Szegedy, 2015). We initialized the weights randomly and trained all plain/residual nets from scratch. We use the Adam optimization method (Zhang et al., 2016) with a minibatch size of 32 to minimize the BCE in this study. Also, the learning rate is settled as 1e-5 in our ResNet work. Furthermore, we use a 10-fold cross-validation strategy in specific implementation processes and repeat 20 times to evaluate our proposed method. Specifically, all subjects used in our work are randomly equally partitioned into 10 groups defined as {S1, S2, S3,…, S10} in each classification process. Group S10 is usually set as the testing set and the other nine groups are further randomly equally repartitioned into 10 subgroups, one of which is selected as the validation set and the other nine subgroups used as the training data.
Figure 2. The overall flow chart of our study. Briefly, the individual-level morphological covariance brain network is first constructed according to the SRI24 atlas and gray matter volume map of each subject, as shown in the dashed boxes. The above morphological covariance brain network is used as the input feature for the classifier. Subsequently, the ResNet network is used to perform the ASD/TC classification, as shown in the middle of this figure. Meanwhile, the Grad-CAM is combined with the ResNet-based architecture based on the rectified convolution feature maps from the ResNet to obtain the heat map for explaining model decisions, as shown at the bottom of this figure.
Results
Four parameters, namely accuracy (ACC), sensitivity (SEN), specificity (SPE), and F1 score, are calculated to evaluate the performance of our proposed ASD/TC classification framework. The deep convolutional neural network used in our work achieved a mean classification accuracy of 71.8%, mean sensitivity value of 81.25%, specificity value of 68.75%, and F1 score of 0.687 from cross-validation. Our results improved the mean classification accuracy of the state-of-the-art from 70 to 71.8% in the ABIDE data, and the former accuracy is obtained by DNN based on fMRI in ABIDE (Heinsfeld et al., 2018). To evaluate our results obtained with the deep convolutional neural network, the performance of our model is compared with the results of classifiers trained using RF (Vapnik, 1998), SVM (Ho, 1995), XgBoost (XGB) (Chen and Guestrin, 2016), and autoencoder (AE). With the purpose of using 2D input data for subject classification by these conventional machine learning methods, a vector of features is firstly retrieved by flattening the 2D morphological covariance brain network (i.e., collapse it in a one dimension vector). The number of resultant features is 11,664, which is computed by 108 × 108. Evaluation of all the models is based on a 10-fold cross-validation schema, which mixes data from all 17 sites while keeping the proportions between different sites. The results of comparing these methods are reported in Table 2. Furthermore, the performance of these classifiers was assessed by the area under the curve (AUC) values shown in Figure 3. Our proposed framework has the best performances in classifying ASD from TC with the highest ACC, SEN, F1 score, and AUC values compared with the other methods. Furthermore, the permutation test with 5,000 times is used to evaluate the significance of the prediction accuracy. During the permutation testing, 20% of the labels of the samples are changed randomly in each time. The histogram of accuracy of the permutation test is shown in Figure 4. The accuracy of our method (0.718) is indicated by the red dotted line. As shown in Figure 4, the 71.8% accuracy of our method is higher than 95% of the permutated accuracy values.
Figure 3. Comparisons between our method and other methods for classification. The area under the curve (AUC) values were used to assess the classification performances for our method, AE, RF, SVM, and XgBoost.
Figure 4. The histogram of accuracy of the permutation test. The permutation test with 5,000 times was used to evaluate the significance of our method. The accuracy of our method (0.718) is indicated by the red dotted line. The classification accuracy is higher than 95% of the permutated accuracy values.
To determine the weight values of the features contributing to classification, the Grad-CAM method was adopted, and the weight of each connectivity was obtained. The individual and fused connectivities supporting the correct classification of ASD patients using Grad-CAM visualizations for our ResNet framework are shown in Figure 5A. In order to make the fusion of Grad-CAM more transparent and explainable, we selected the covariance connectivities with the largest contribution to classification. The largest contributions of the connectivities were determined by identifying the weights above the mean + 3SD. Finally, 63 connectivities between 12 different regions were found using the fused Grad-CAM approach. The top 12 regions correspond to the bilateral precentral gyrus (left and right), superior frontal gyrus, orbital part of the superior frontal gyrus, and cerebellum 8–10 (see Figure 5B).
Figure 5. Weights of the features for classification. (A) The individual and fused connectivities supporting the correct classification of ASD patients were mapped using the Grad-CAM visualizations for our ResNet framework. The red regions correspond to high score for ASD class. (B) The top 12 regions and the corresponding connectivities which have the largest contribution correctly classifying ASD patients were identified.
Discussion
As far as we know, the great majority of machine learning or deep learning methods used to identify ASD patients from controls are mainly based on the resting-state fMRI. Although fMRI constructs individual brain networks by estimating interregional functional connectivity, those networks are made by the graph theory analytical method, which are only efficient for imaging data with 4D time series. However, given the great individual variability of fMRI, sMRI and its derived measures with high reproducibility have been widely used for disease classification. Although some previous studies extracted local conventional morphological features, such as gray matter volume, thickness, or volume of different regions from sMRI for machine learning, the relationship between structural properties of different regions has not been explored since the coordinated patterns of the local morphological features between regions are important for cognitive development (Bullmore and Bassett, 2010; Vértes and Bullmore, 2014). Thus, the construction of a structural covariance network with sMRI to explore individual brain topological organization and to investigate its alterations or abnormalities under both healthy and pathological conditions has attracted increasing attention. The whole-brain morphological network at the individual level based on sMRI characterizes the topological organizations at both the global and nodal levels (Wang et al., 2016). Thus, the individual-level morphological brain networks can better reflect individual behavior differences in both typical and atypical populations than the group-level morphological network. Moreover, compared with the local regional measures with sMRI, the individual morphological covariance network can provide more features to meet the requirements for the number of features during deep learning training. Thus, combining the morphological covariance network and deep learning will open a new avenue for future studies with sMRI.
In recent works, deep learning algorithms improve the classification accuracy in the identification of ASD versus TC. However, they are usually treated as “black-box” methods because of the lack of understanding of their internal functions (Lipton, 2016). The black-box methods cannot explain their predictions in a way that humans can understand. To fix this problem, the Grad-CAM technique uses gradients of the target concept (identification of ASD in a classification network for our work) and produces a coarse localization map to identify the weight of each feature of the image during classification or prediction, which produces “visual explanations” for decisions from the CNN-based models and makes them to be more transparent and explainable (Selvaraju et al., 2020). Compared with other visualization techniques, Grad-CAM can highlight the important connectivities in the morphological covariance brain networks for discriminating ASD without model architectural changes or retraining. Thus, Grad-CAM combined with our classification model provides a reference for future study to determine the important features in deep learning framework.
Using the Grad-CAM method, we found that the morphological covariance between frontal and cerebellar areas has the largest contribution for classification. The frontal areas include the precentral gyrus, superior frontal gyrus, and orbital part of the superior frontal gyrus. For the cerebellum, cerebellum 8, 9, and 10 were found to contribute greatly for classification. The precentral gyrus and cerebellum have been widely demonstrated to be associated with motor processing and integration of sensory information. Thus, the structural covariance between the precentral gyrus and cerebellum suggested that they may be related to the rigid, stereotyped, and repetitive behaviors in ASD (Mei et al., 2020). Furthermore, we also found the structural covariance connectivities among the superior frontal gyrus, orbital part of the superior frontal gyrus, and cerebellum contributing largely to the classification. The cerebellum participates not only in motor functions but also in emotion, memory, language, and social cognition processing (Strick et al., 2009; Buckner, 2013). The superior frontal gyrus has been demonstrated to be involved in social cognition (Monk et al., 2009; Li et al., 2013). Thus, the superior frontal gyrus and its orbital part and the cerebellum may be related to social cognition processing in ASD. Given that rigid, stereotyped, and repetitive behaviors and impaired social cognition are the core symptoms of ASD, our findings further demonstrated the reliability and feasibility of our proposed method for ASD classification. Also, our proposed method may advance establishing the framework for early diagnosis of ASD.
Conclusion
In this study, we proposed a convolutional neural network framework based on the individual-level morphological covariance brain network for ASD diagnosis. We found that our proposed method outperformed other classification methods for the classification of ASD in a multisite. Moreover, using Grad-CAM, we can identify the weight of each feature for classification, which solves the black-box problems of deep learning. Our study proposes a new paradigm for ASD classification that has a good performance in multisite datasets and will facilitate establishing the diagnosis framework for ASD.
Data Availability Statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics Statement
The studies involving human participants were reviewed and approved by the St. James’s Hospital/AMNCH (ref: 2010/09/07) and the Linn Dara CAMHS Ethics Committees (ref: 2010/12/07). Written informed consent to participate in this study was provided by the participants’ legal guardian/next of kin.
Author Contributions
JW, SC, and JG designed this study. JG, MC, YG, and YuL performed the experiments. JG, YaL, and JW wrote the manuscript. All the authors discussed and edited this manuscript.
Funding
This work was funded in part by the Sichuan Science and Technology Program (2021YJ0186 and 2019YJ0193), the National Natural Science Foundation of China (61701078 and 71572152), Sichuan Key Research and Development Plan (21ZDYF3062), and the Science Promotion Programme of UESTC, China (Y03111023901014006).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Footnotes
References
Buckner, R. L. (2013). The cerebellum and cognitive function: 25 years of insight from anatomy and neuroimaging. Neuron 80, 807–815. doi: 10.1016/j.neuron.2013.10.044
Bullmore, E., and Bassett, D. (2010). Brain graphs: graphical models of the human brain connectome. Annu. Rev. Clin. Psychol. 7, 113–140. doi: 10.1146/annurev-clinpsy-040510-143934
Button, K. S., Ioannidis, J. P., Mokrysz, C., Nosek, B. A., Flint, J., Robinson, E. S., et al. (2013). Power failure: why small sample size undermines the reliability of neuroscience. Nat. Rev. Neurosci. 14, 365–376. doi: 10.1038/nrn3475
Chen, R., Jiao, Y., and Herskovits, E. H. (2011). Structural MRI in autism spectrum disorder. Pediatric Res. 69(5 Pt 2), 63R–68R.
Chen, T., and Guestrin, C. (2016). XGBoost: a scalable tree boosting system. arXiv [Preprint] doi: 10.1145/2939672.2939785
Di Martino, A., Yan, C. G., Li, Q., Denio, E., Castellanos, F. X., Alaerts, K., et al. (2014). The autism brain imaging data exchange: towards a large-scale evaluation of the intrinsic brain architecture in autism. Mol. Psychiatry 19, 659–667.
Guze, S. B. (1995). Diagnostic and statistical manual of mental disorders, 4th ed. (DSM-IV). Am. J. Psychiatry 152, 1228–1228.
Haar, S., Berman, S., Behrmann, M., and Dinstein, I. (2016). Anatomical abnormalities in autism? Cereb. Cortex 26, 1440–1452. doi: 10.1093/cercor/bhu242
He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Las Vegas, NV), 770–778.
Heinsfeld, A. S., Franco, A. R., Craddock, R. C., Buchweitz, A., and Meneguzzi, F. (2018). Identification of Autism spectrum disorder using deep learning and the ABIDE dataset. Neuroimage Clin. 17, 16–23. doi: 10.1016/j.nicl.2017.08.017
Hinton, G., Srivastava, N., Krizhevsky, A., Sutskever, I., and Salakhutdinov, R. (2012). Improving neural networks by preventing co-adaptation of feature detectors. arXiv [Preprint] arXiv.1207.0580.
Ho, T. K. (1995). “Random decision forests,” in Proceedings of 3rd International Conference on Document Analysis & Recognition (Montreal QC).
Ioffe, S., and Szegedy, C. (2015). “Batch normalization: accelerating deep network training by reducing internal covariate shift,” in Proceedings of the 32nd International Conference on Machine Learning (France).
Katuwal, G. J., Cahill, N. D., Baum, S. A., and Michael, A. M. (2015). “The predictive power of structural MRI in Autism diagnosis,” in Proceedings of the 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society IEEE Engineering in Medicine and Biology Society Annual International Conference, Vol. 2015 (Milan), 4270–4273.
Kong, X. Z., Wang, X., Huang, L., Pu, Y., Yang, Z., Dang, X., et al. (2014). Measuring individual morphological relationship of cortical regions. J. Neurosci. Methods 237, 103–107. doi: 10.1016/j.jneumeth.2014.09.003
Kong, Y., Gao, J., Xu, Y., Pan, Y., Wang, J., and Liu, J. (2019). Classification of autism spectrum disorder by combining brain connectivity and deep neural network classifier. Neurocomputing 324, 63–68. doi: 10.1016/j.neucom.2018.04.080
Li, C., Gore, J. C., and Davatzikos, C. (2014). Multiplicative intrinsic component optimization (MICO) for MRI bias field estimation and tissue segmentation. Magn. Reson. Imaging 32, 913–923. doi: 10.1016/j.mri.2014.03.010
Li, W., Qin, W., Liu, H., Fan, L., Wang, J., Jiang, T., et al. (2013). Subregions of the human superior frontal gyrus and their connections. Neuroimage 78, 46–58. doi: 10.1016/j.neuroimage.2013.04.011
Liu, Y., Chen, X., Peng, H., and Wang, Z. (2017). Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 36, 191–207. doi: 10.1016/j.inffus.2016.12.001
Liu, Y., Chen, X., Wang, Z., Wang, Z. J., Ward, R. K., and Wang, X. (2018). Deep learning for pixel-level image fusion: recent advances and future prospects. Inf. Fusion 42, 158–173. doi: 10.1016/j.inffus.2017.10.007
Mei, T., Llera, A., Floris, D. L., Forde, N. J., Tillmann, J., Durston, S., et al. (2020). Gray matter covariations and core symptoms of autism: the EU-AIMS longitudinal European Autism project. Mol. Autism 11:86.
Monk, C. S., Peltier, S. J., Wiggins, J. L., Weng, S. J., Carrasco, M., Risi, S., et al. (2009). Abnormalities of intrinsic functional connectivity in autism spectrum disorders. Neuroimage 47, 764–772. doi: 10.1016/j.neuroimage.2009.04.069
Nielsen, J. A., Zielinski, B. A., Fletcher, P. T., Alexander, A. L., Lange, N., Bigler, E. D., et al. (2013). Multisite functional connectivity MRI classification of autism: ABIDE results. Front. Hum. Neurosci. 7:599. doi: 10.3389/fnhum.2013.00599
Ou, Y., Sotiras, A., Paragios, N., and Davatzikos, C. (2011). DRAMMS: deformable registration via attribute matching and mutual-saliency weighting. Med. Image Anal. 15, 622–639. doi: 10.1016/j.media.2010.07.002
Rohlfing, T., Zahr, N. M., Sullivan, E. V., and Pfefferbaum, A. (2010). The SRI24 multichannel atlas of normal adult human brain structure. Hum. Brain Mapp. 31, 798–819. doi: 10.1002/hbm.20906
Rosenblatt, M. (1956). Remarks on some nonparametric estimates of a density function. Ann. Math. Stat. 27, 832–837. doi: 10.1214/aoms/1177728190
Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017). “Grad-CAM: visual explanations from deep networks via gradient-based localization,” in Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV) (Venice), 618–626.
Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2020). Grad-CAM: visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vis. 128, 336–359. doi: 10.1007/s11263-019-01228-7
Strick, P. L., Dum, R. P., and Fiez, J. A. (2009). Cerebellum and nonmotor function. Annu. Rev. Neurosci. 32, 413–434. doi: 10.1146/annurev.neuro.31.060407.125606
Vapnik, V. (1998). “The support vector method of function estimation,” in Nonlinear Modeling: Advanced Black-Box Techniques, eds J. A. K. Suykens and J. Vandewalle (Boston, MA: Springer US), 55–85. doi: 10.1007/978-1-4615-5703-6_3
Vértes, P., and Bullmore, E. (2014). Annual research review: growth connectomics – the organization and reorganization of brain networks during normal and abnormal development. J. Child Psychol. Psychiatry Allied Dis. 56, 299–320. doi: 10.1111/jcpp.12365
Wang, H., Jin, X., Zhang, Y., and Wang, J. (2016). Single-subject morphological brain networks: connectivity mapping, topological characterization and test-retest reliability. Brain Behav/ 6:e00448.
Wang, J., Becker, B., Wang, L., Li, H., Zhao, X., and Jiang, T. (2019). Corresponding anatomical and coactivation architecture of the human precuneus showing similar connectivity patterns with macaques. Neuroimage 200, 562–574.
Wang, J., Feng, X., Wu, J., Xie, S., Li, L., Xu, L., et al. (2018). Alterations of gray matter volume and white matter integrity in maternal deprivation monkeys. Neuroscience 384, 14–20. doi: 10.1016/j.neuroscience.2018.05.020
Wang, J., Wei, Q., Bai, T., Zhou, X., Sun, H., Becker, B., et al. (2017). Electroconvulsive therapy selectively enhanced feedforward connectivity from fusiform face area to amygdala in major depressive disorder. Soc. Cogn. Affect. Neurosci. 12, 1983–1992. doi: 10.1093/scan/nsx100
Wu, H., Sun, H., Wang, C., Yu, L., Li, Y., Peng, H., et al. (2017). Abnormalities in the structural covariance of emotion regulation networks in major depressive disorder. J. Psychiatric Res. 84, 237–242. doi: 10.1016/j.jpsychires.2016.10.001
Xu, J., Wang, J., Bai, T., Zhang, X., Li, T., Hu, Q., et al. (2019). Electroconvulsive therapy induces cortical morphological alterations in major depressive disorder revealed with surface-based morphometry analysis. Int. J. Neural. Syst. 29:1950005. doi: 10.1142/s0129065719500059
Zhang, K., Zuo, W., Chen, Y., Meng, D., and Zhang, L. (2016). Beyond a Gaussian Denoiser: residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 26, 3142–3155. doi: 10.1109/tip.2017.2662206
Keywords: autism spectrum disorder, individual morphological covariance brain network, convolutional neural network, gradient-weighted class activation mapping, structural MRI
Citation: Gao J, Chen M, Li Y, Gao Y, Li Y, Cai S and Wang J (2021) Multisite Autism Spectrum Disorder Classification Using Convolutional Neural Network Classifier and Individual Morphological Brain Networks. Front. Neurosci. 14:629630. doi: 10.3389/fnins.2020.629630
Received: 15 November 2020; Accepted: 29 December 2020;
Published: 28 January 2021.
Edited by:
Kai Yuan, Xidian University, ChinaReviewed by:
Xun Chen, University of Science and Technology of China, ChinaZaixu Cui, University of Pennsylvania, United States
Copyright © 2021 Gao, Chen, Li, Gao, Li, Cai and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Shimin Cai, shimin.cai81@gmail.com; Jiaojian Wang, jiaojianwang@uestc.edu.cn; Jingjing Gao, jingjing.gao@uestc.edu.cn
†These authors have contributed equally to this work