- 1Department of Electrical and Computer Engineering, University of Nevada Las Vegas, Las Vegas, NV, United States
- 2Kirk Kerkorian School of Medicine, University of Nevada Las Vegas, Las Vegas, NV, United States
- 3Medical Physics Program, Department of Health Physics, University of Nevada Las Vegas, Las Vegas, NV, United States
- 4Department of Radiology, Feinberg School of Medicine, Northwestern University, Chicago, IL, United States
- 5Department of Neurological Surgery, Feinberg School of Medicine, Northwestern University, Chicago, IL, United States
- 6Malnati Brain Tumor Institute of the Lurie Comprehensive Cancer Center, Feinberg School of Medicine, Northwestern University, Chicago, IL, United States
Radiological imaging techniques, including magnetic resonance imaging (MRI) and positron emission tomography (PET), are the standard-of-care non-invasive diagnostic approaches widely applied in neuro-oncology. Unfortunately, accurate interpretation of radiological imaging data is constantly challenged by the indistinguishable radiological image features shared by different pathological changes associated with tumor progression and/or various therapeutic interventions. In recent years, machine learning (ML)-based artificial intelligence (AI) technology has been widely applied in medical image processing and bioinformatics due to its advantages in implicit image feature extraction and integrative data analysis. Despite its recent rapid development, ML technology still faces many hurdles for its broader applications in neuro-oncological radiomic analysis, such as lack of large accessible standardized real patient radiomic brain tumor data of all kinds and reliable predictions on tumor response upon various treatments. Therefore, understanding ML-based AI technologies is critically important to help us address the skyrocketing demands of neuro-oncology clinical deployments. Here, we provide an overview on the latest advancements in ML techniques for brain tumor radiomic analysis, emphasizing proprietary and public dataset preparation and state-of-the-art ML models for brain tumor diagnosis, classifications (e.g., primary and secondary tumors), discriminations between treatment effects (pseudoprogression, radiation necrosis) and true progression, survival prediction, inflammation, and identification of brain tumor biomarkers. We also compare the key features of ML models in the realm of neuroradiology with ML models employed in other medical imaging fields and discuss open research challenges and directions for future work in this nascent precision medicine area.
Introduction
Glioblastoma (GBM, WHO grade 4 glioma, IDH-wildtype) is the most aggressive primary brain tumor in adults with a dismal median overall survival (OS) of only 12 to 18 months and a 5-year OS rate of 6.8% (1, 2). Approximately 13,000 GBM cases are diagnosed in the United States each year, with an incidence rate of 3.2 per 100,000 members of the population (3, 4). Despite standard-of-care therapy including aggressive surgical resection followed by radiation therapy and chemotherapy, more than 90% of glioblastomas recur (4). To date, magnetic resonance imaging (MRI) remains the standard approach in the diagnosis, prognosis, and therapeutic monitoring of GBM patients because it is non-invasive, accessible, and cost efficient. However, interpretation of radiological imaging data can be subjective, challenging, and time-consuming, mainly because histologic findings are often radiologically occult. For example, therapy-induced treatment effect (i.e., pseudoprogression (PsP) or radiation necrosis) and true tumor progression manifest with identical MRI appearances, and differentiation between these entities remains an unsolved conundrum in current neuro-radio-oncology, particularly with novel therapies such as immune checkpoint inhibitors (5).
Radiomics (6, 7) in neuro-oncology seeks to improve the understanding of the biology and effects of treatment on the imaging appearance of brain tumors. Radiomics can promote precision medicine by extracting quantitative features from clinical imaging arrays and using methods from the field of artificial intelligence (AI) to make the radiological diagnosis more objective, accurate, and automatic. Rather than designing hard-coded step-by-step algorithms based on prior knowledge in biology or medicine, or design specific “learning” approaches to mimic human cognitive functions, machine learning (ML) as a subfield of AI can create a computational model and train it with a number of datasets to statistically solve problems without being explicitly programmed (8). Generally, ML includes supervised learning, unsupervised learning, and reinforcement learning. Supervised learning trains an ML model to predict a target variable from a set of predictive variables (i.e., data samples) with the help of labels/annotations (i.e., ground truth of the target variables) and the loss function, also known as cost function, which is a computational difference between predicted target variable values and the label/annotation values (9). It should be noted that, although labeling and annotation share the same meaning in ML, they slightly differ in neuro-oncology radiomic analysis. In the context of this manuscript, labeling is related with classification problems (e.g., the ground truth of tissue is histological, including different classes of brain tumors, treatment effect versus tumor growth, and others) whereas annotation refers to segmentation problems (partitioning an image into multiple regions/objects, such as enhancing tumor, necrosis, and unenhancing tumor and edema). Unsupervised learning infers the inherent structure from the input data without labels/annotations (10). In reinforcement learning, intelligent agents learn to take actions in an environment in order to maximize the notion of cumulative reward (11). Currently, most AI techniques applied in brain tumor radiomic studies belong to the supervised ML, for both classifications and segmentations. Unsupervised ML is mainly employed for image segmentations while reinforcement learning has not been explored in this area. Therefore, in this paper, we mainly focus on the supervised ML techniques for most GBM radiomic analyses. Thanks to the rapid development of radiology and computational hardware, researchers can now take advantage of many radiological data to train various ML models, such as decision trees (DTs), logistic regression, artificial neural networks (ANNs), support vector machines (SVMs), and k-nearest neighbors (k-NN) for brain tumor radiomic analysis. The different techniques applied in AI (mainly ML algorithms) are the technical core for the analysis of large amounts of multidimensional radiologic and clinical data (12), which directly determine the quality of radiomic analysis results.
In the past decade, ML has been widely exploited in many data-driven applications, e.g., imaging and computer vision (13), bioinformatics (14), online advertising (15), and natural language processing (16). The dataflow of a general supervised ML-based GBM radiomic analysis can be divided into four steps as shown Figure 1A: 1) Data Acquisition. MRIs are performed on patients with a brain tumor. These raw MRI data are further preprocessed (e.g., data cleaning, co-registration, bias correction, normalization), and then they are labeled/annotated by radiologists to define the regions for the ML training and validation process. The labeled/annotated imaging data are deposited into customized/private datasets that are owned and maintained by medical research institutions. Some imaging data are also uploaded into public datasets for the purpose of open access to all researchers. AI-assisted radiomic analysis can acquire imaging data from both types of datasets. It is of note, however, that private datasets usually contain a fairly large amount of raw data, hundreds of samples for each institution if applicable, whereas public datasets usually contain limited amount of less well-labeled/annotated, non-standardized imaging data (17); 2) Data Augmentation and Preprocessing for ML Models. The acquired data and its labels/annotations are usually first subject to augmentation, in which image data are processed in pair-wise format (i.e., each pair contains a data sample and the corresponding labels/annotations) to increase the sample variety, hence improving the generality of the data. Multiple approaches are utilized in augmentation, such as geometric transformation, color augmentation, and synthesis of similar-appearing imaging data. Then, the augmented imaging data can be preprocessed (e.g., through feature extraction) to simplify and/or to improve the effectiveness and efficiency of the subsequent ML training process (18); 3) ML Model Training and Validation: The augmented and preprocessed data are subsequently fed to ML models to train the model parameters in order to minimize the “cost function,” while the implicit data feature will be extracted statistically. The ML models are also validated during the training process to prevent overfitting, which is when the trained ANN model only predicts accurately to the training dataset but loses the generalizability to new samples (19); 4) AI-Assisted Clinical Diagnosis/Deployment. Once the trained models meet the accuracy requirement, they can be deployed to the application to perform predictions such as classification and segmentation. As aforementioned, no biologic hypothesis or knowledge is required to build an ML model. However, inclusion of this information and/or other forms of data (e.g., clinical data, genomic data) may help with the overall ML model performance by reducing the data size or improving the data quality during the preprocessing step.
Figure 1 AI/ML in GBM radiomic analysis: (A) Overall workflow of AI-assisted GBM analysis: 1) Data Acquisition. Raw radiological image data are acquired by MRI scanning of GBM patients. Images are collected into public or private data sets. Before analysis, images are preprocessed (e.g., data cleaning, co-registration, normalization) and standardized (e.g., format, resolutions, voxel sizes). Then, radiologists annotate the images, color-coding different parts of the tumor habitat. 2) Data Augmentation and Preprocessing for ML Models. Imaging dataset and its annotations from step 1 are further “augmented” via geometric transformations, photometric transformation, and/or synthetic data (e.g., GAN) to improve the data generalizability, followed by the optional preprocessing for ML modeling, a process that includes feature extraction to filter out “useless” data and extract explicit features (e.g., biological and/or geometry) in the images. 3) ML modeling and training. Augmented and preprocessed data are fed into various ML models (e.g., SVM, RF, CNN) for GBM radiomic analysis training and validation. Advanced techniques such as transfer learning and multimodal data fusion (e.g., clinical and genomic data) can be employed to improve the training accuracy as well as generality. 4) AI-Assisted Clinical Diagnosis/Deployment. Predictions from the ML models for various medical demands, such as differential diagnosis and survival estimation. (B) Current major challenges (left panel: 1, 2, 3) and perspectives for corresponding solutions (right panel: 1*, 2*, 3*) in AI/ML-assisted GBM radiomic analysis: 1→1* Current GBM radiological image datasets are limited in low numbers, insufficient annotations, and poor organization. Enrichment and standardization of current GBM radiological datasets are urgently needed, while incorporation of clinical and/or genomic data (red circle) can further enhance the performance of ML prediction models; 2→2* develop more comprehensive ML models to further improve the prediction accuracy and address the relatively low generalizability of current models; 3→3* further strengthen collaborations among clinicians, biomedical researchers, and computer scientists to overcome the lack of efficient communications between these parties for the highly multidisciplinary research.
Traditional ML methods such as SVMs (20–22) and random forests (RFs) (23–26), an ensemble combination of decision trees, are commonly used for pattern classification in tumor studies. Recently, ANN, especially convolutional neural network (CNN)-based deep learning (DL), is gaining popularity because of its improved scalability and the capability of exploiting deep layers to extract implicit local and global features in neuro-oncology images. They can achieve state-of-the-art performance in object detection and tracking (27, 28), image classification, and semantic segmentation (25, 29–32). In ML, each “neuron” is referring to a computational unit in the ANN rather than a biological neural. With more complicated ML models and structures (e.g., more neuron network layers, more neurons in each layer) and a larger number of parameters introduced into the neural network, the training process intends to extract more features but may suffer extensive computational performance degradation and overfitting of the trained model. Current state-of-the-art ML models can achieve an accuracy as high as 0.97 (i.e., 97%) in brain tumor radiomic analysis (33). However, these results are based on a limited number of datasets and from retrospective studies, which may still not be generalizable for patients from different geographic locations. Therefore, current clinical brain tumor radiomic analysis cannot entirely rely on the ML-based techniques and still needs manual verification. In summary, existing ML techniques can only partially fulfill the need for automatic detection and analysis of GBM characteristics for both clinical and preclinical studies (34–37).
In addition to the ML techniques, the quality of radiologic images that are used for ML training dramatically affects the outcomes of radiomic analysis. Radiological images can be acquired from different imaging modalities, such as MRI, computed tomography (CT), and positron emission tomography (PET). Among these, MRI image data are currently employed as an essential data type in radiomic neuro-oncology applications because 1) they provide exquisite detail of brain, spinal cord, and vascular anatomy through excellent tissue contrast in any imaging plane; 2) different MRI sequences are able to capture key components of tumor biology with high sensitivity, such as blood–brain barrier breakdown, necrosis, edema, non-enhancing tumor infiltration, blood flow, and cellular density, and can distinguish tumoral sub-compartments that are likely to reflect local cellular phenotypes and genotypes; and 3) they can non-invasively and non-destructively interrogate tumors repeatedly to assess response to treatment and thus they can be integrated into therapeutic strategies. Understanding these image-based features is critical as they not only represent a key data resource in radiomic analysis (6) but also help improve the accuracy and other performance criteria of ML models.
In this review, we provide an overview of the latest advancements and in-depth discussions on the most urgent and challenging questions of AI-assisted GBM radiomic analysis. Given the exponential increase of AI-based radiomic studies led by researchers from various backgrounds, such as oncology, radiology, computer science, and engineering, our review article briefly explains the key concepts of ML techniques instead of delving into the technical details. This article is structured with emphases on the deployments of various ML techniques in meeting specific GBM radiomic clinical needs, e.g., differentiating GBM from other brain tumors or non-tumors, predicting overall survival (OS), and correlating with other biomarkers. First, as ML technology in radiomics is radiological imaging data-driven, we start with the discussion on imaging data preparations that are commonly employed in current GBM radiomic analyses. We briefly introduce the acquisition pipeline for private or customized imaging datasets and summarize public radiologic datasets that are currently available for researchers to train their ML models for various applications. We also describe general methods for data augmentation and preprocessing for ML models, which are critical for training, validation, and testing of ML algorithms. Next, we overview the ML techniques that have been employed in radiomic analysis for GBM diagnosis and treatment. Advantages and limitations of existing ML models including both algorithms and architectures are discussed in the context of various GBM-associated medical applications. Finally, we bring up our perspectives on the strategies for overcoming challenges regarding AI/ML applications in GBM radiomic analysis, including 1) the most challenging issues affecting the generalizability and accuracy of AI-assisted radiomic GBM analysis; 2) promising strategies to enhance performance of AI models in GBM radiomic analysis; 3) outlook on the collaborative teamwork between computer scientists, engineers, physicians, and biomedical researchers. By elaborating current research developments and challenges in the state-of-the-art ML-assisted GBM analysis, we hope to inspire researchers from different fields for the development of the next generation of AI-assisted radiomic tools that can significantly improve early detection, treatment efficacy, and life quality of patients with GBM.
Radiomic data preparation
As ML is an intensive data-driven algorithm/process, the quality of the training data can significantly influence the trainable parameters in ML models, hence affecting the accuracy and generalizability of the network outcomes (38). Thus, it is essential to review the key components of data preparation in ML-based GBM radiomic analysis, including radiological imaging data resources, the image acquisition pipeline, imaging datasets (private/customized datasets vs. publicly accessible datasets), data augmentation, and preprocessing techniques for the subsequent ML model training.
Radiomic image resources
MRI is the most frequently used radiological modality for brain tumor imaging. MRI provides better contrast resolution than CT, with better tissue characterization. It can also detect blood vessels, vascular malformations, and demyelinating disease (39). It does not involve X-rays or the use of ionizing radiopharmaceuticals, either. Therefore, MRI is particularly suitable to image gliomas. Yet, MRI may be perceived as less comfortable by patients (e.g., overweight or fear of enclosed spaces) and cannot be performed if the patient has ferromagnetic implants. In addition, MRI does not show ossified or calcified structures as well as CT (such as the calvarium) and therefore may not show the effects of tumors on the calvarium in comparison to CT (39, 40).
For GBM MRI, T1-weighted (T1), T1-contrast-enhanced (T1c or T1-ce), T2-weighted (T2), and fluid-attenuated inversion recovery (FLAIR) are the most commonly used MRI sequences, because they can provide different yet complementary information in characterizing tissue such as gray matter, white matter, fat, blood, fluid, and lesions (41, 42). MRI is based on radiofrequency pulses within a magnetic field in which time of repetition (TR) and time of echo (TE) are calculated. T1 and T1-ce are produced through short TR and TE times; T2 is produced by larger TR and TE times; and FLAIR is produced through very large TR and TE times. In addition, diffusion-weighted imaging (DWI) can detect the restriction of random movements of water molecules that makes DWI extremely sensitive to detect acute stroke and increased cellularity as in GBMs, lymphoma, and metastases (43); magnetic resonance angiography (MRA) and venography (MRV) can generate pictures of the arteries and veins to evaluate for stenosis or aneurysms; magnetic resonance spectroscopy (MRS) is used to measure the levels of different metabolites and biochemical changes in the brain (44), providing information on tumor metabolism (45, 46); perfusion-weighted imaging (PWI) shows the perfusion of tissues by blood, such as the cerebral blood volume of a tumor relative to normal-appearing white matter of the brain (47); and functional MRI (fMRI) detects the increase in blood oxygen level when blood flow increases to a brain area involved in the performance of an assigned task (e.g., finger tapping, lip pursing, thinking of words, thinking of answers to questions after hearing a story) (48) and depicts where eloquent brain areas are in relation to the tumor as the surgeon or radiation oncologist plans a surgical approach to biopsy or resection or radiation therapy. The novel amide proton transfer (APT) imaging can detect amide protons of endogenous mobile proteins and peptides in tissue based on chemical exchange–dependent saturation transfer (CEST) MRI (49, 50).
Another useful radiological technology is the PET scan, which takes advantage of a slightly radioactive substance (e.g., C-11 methionine (MET), F-18 fluorothymidine (FLT), F-18 fluoroethyl-L-tyrosine (FET)) that functionally is preferentially taken up by tumor cells (51). PET is especially helpful for fast-growing (high-grade) tumors and for distinguishing between tumors and non-tumor (e.g., scar, inflammation) tissue (52, 53). Therefore, the combined use of MRI and PET can provide complementary information to achieve more accurate brain tumor diagnosis (54).
Radiological image acquisition pipeline
The image acquisition pipelines remain very similar between radiological scanners (50, 55). Subjects undergoing MRI (i.e., patients and control subjects) are usually examined on a clinical 1.5T or 3T scanner with a multichannel receive-only head coil array under various scanning parameters (e.g., TR, TE, field of view, matrix/voxel size). A sequence of 2D and 3D radiologic images (e.g., various MRI, CT, PET) is obtained. These images are cleaned, normalized, and co-registered (i.e., image preprocessing, Figure 1A). It is worth noting that different image intensity normalization schemes may influence not only the registration (56) and segmentation process (57, 58) but also the implicit texture features hidden in the different modal MR images and thus affect the subsequent feature selection and ML training outcomes for various GBM classification applications (59). Shinohara et al. (60) introduced a set of seven statistical principles of image normalization (SPIN). In addition to the common mean-maximum (or standard deviation) normalizations (59) and histogram-based normalizations (57, 61, 62), Shinohara et al. (60) also proposed a hybrid multimodal normalization method to match the natural balance of tissue intensities with physical interpretation. On the other hand, some data may be standardized (e.g., voxel sizes, resolution) while others may not. Then, the images are labeled and annotated into various categories of tissues and/or lesions by experienced radiologists using a variety of software (e.g., ITK-SNAP (63), 3D Slicer (64)) to produce a labeled/annotated imaging data set (17).
Major public datasets
Since not all ML researchers can directly access private/customized high-quality labeled/annotated brain tumor datasets, which are usually owned and protected by medical institutions, public datasets are essential and provide an equal platform to these researchers to train and compare the outcomes of their ML models. In neuro-oncology, one of the most commonly used public online image datasets is from the Brain Tumor Segmentation (BraTS) challenges, organized by the Medical Image Computing and Computer Assisted Interventions (MICCAI) and other professional organizations (34, 65–69) since 2012. As of March 2022, the latest BraTS 2021 consists of a total of 2,040 brain tumor cases/patients, and it is divided into three subsets: training (1,251 cases), validation (219 cases), and testing (570 cases). Only training and validation subsets are open to the public research access, and these two subsets include a set of multimodal 3D MRI scans (i.e., T1, T1c, T2, FLAIR) for each case. The training dataset also includes a 3D annotation model (i.e., GD-enhancing tumor, peritumoral edema/non-enhancing infiltrative tumor, necrotic tumor core (NCR), and normal) for each case (17). In addition, BraTS 2021 includes O6-methylguanine-DNA-methyltransferase (MGMT) biomarkers for 585 patients (out of a total of 1,251 cases) for the training datasets. It should be noted that, unlike BraTS 2020, BraTS 2021 does not include survival information any longer. BraTS 2020 included the survival information for 265 patients (out of a total of 460 cases) in the training and validation datasets. Other widely used datasets include The Cancer Imaging Archive (TCIA) (70) and The Whole Brain Atlas by Harvard Medical School (71). TCIA has a collection of 13 brain tumor sub-datasets, including the Ivy Glioblastoma Atlas Project (IvyGAP) (72), The Cancer Genome Atlas (TCGA)-GBM (73), GLIS-RT (74), and CPTAC-GBM (75). These sub-datasets mainly focus on high-grade glioma (HGG/GBM) and lower-grade glioma (LGG). Some of the data in TCIA are also included and standardized in the most recent BraTS 2021 dataset. The Harvard brain atlas consists of the radiology data (e.g., MRI, CT, PET) for about 40 subsets of normal brain and various brain disease states. However, none of the datasets in the TCIA or Harvard brain atlas are pixel-wisely annotated for the ML segmentation tasks. In addition, the data format for each patient in TCIA collection varies in terms of pulse sequences (e.g., T1, T1c, T2, FLAIR, PWI, DWI) and the resolution (i.e., with matrices varying from 128*128 to 896*896), even in the same sub-dataset (e.g., TCGA-GBM, IvyGAP). A summary of these datasets is depicted in Table 1.
Data augmentation and preprocessing for the ML models
Data augmentation is a commonly used technique in ML for the purpose of promoting the accuracy and generalizability of the ML algorithms. Data augmentation can be attained by 1) adding slightly modified copies and/or 2) creating new synthetic data from already existing data. The former usually employs geometric transformation and photometric transformation including flipping, pixel-level augmentation, cropping, rotating, noise injection, and random erasing (76, 77), while the latter may make use of generative adversarial networks (GANs) (78, 79) to create new synthetic images that resemble the original dataset. It should be noted that GANs also belong to ML-based networks that require abundant training data to generate resembled data. Data augmentation acts as a regularizer and helps reduce class imbalance and overfitting (76), so as to improve both the accuracy and the generalizability of the ML outcomes.
Current GBM radiomic studies are often hindered by limited and unbalanced data samples; therefore, using ML models alone may not achieve statistically significant outcomes (80–82). In this regard, the preprocessing plays a vital role by enhancing and extracting some image features, especially the biological/medical meaningful ones in the regions of interests (ROIs), and/or filters out some “useless” image data from the datasets, before performing the ML training and analysis (18). General data preprocessing approaches for ML models include feature extraction and feature selection (18, 81). In GBM radiomic datasets, MRIs contain various features, such as image texture (23), local histograms (24), structure tensor eigenvalues (25), gray-level co-occurrence matrix (GLCM) (83), and local binary pattern (LBP) (41). Yet some of these features might be correlated in that the total number of effective features can be further reduced, by employing feature selection algorithms such as principal component analysis (PCA) (84), least absolute shrinkage and selection operator (LASSO) (85), linear discriminant analysis (LDA) (86), t-tests (87), analysis of variance (88), and information gain based methods (89), or based on certain evaluation criteria, such as probability of error (POE) and average correlation coefficient (ACC) (59, 90). Injecting feature extraction and selection can significantly help to reduce the computational complexity of ML models and speed up the training process and possibly improve the accuracy of ML models for brain tumor classification and segmentations (91, 92). Nevertheless, one should note that such preprocessing should be treated with care so that the ML model is not overfitted to particular features, which could lose generalizability to a different dataset.
Discussion
We have illustrated various radiomic data sources and data preparation techniques that are commonly employed in ML-based GBM radiomic analysis. One prominent issue in current GBM radiomic data preparation lies in the lack of standardized image acquisition specifications (e.g., repetition time, echo time, voxel sizes, image resolutions) between different radiological equipment and medical institutions (i.e., multicenter multi-vendor, McMv, datasets), which may 1) bias the image data (e.g., intensity of pixels, actual voxel size); 2) require additional image data preparation (e.g., cropping, up/down-sampling) to train ML models with different datasets; and 3) impede the development and cross-validation of more general/robust and accurate ML models for McMv datasets. Although BraTS has made a huge effort and progress in standardizing radiology data for over 2,000 GBM patients/cases, it is not yet sufficient for various GBM analysis applications. The second important issue is that most existing datasets have limited types of brain tumors (e.g., GBM/HGG, metastasis, and LGG), while the Harvard Atlas is limited by the number of subjects/patients. Scarcity of brain tumor/disease types and lack of data impede the application of ML to accurately distinguish various brain tumors and diseases. Lastly, most datasets include only MRI data while only a few datasets consist of other modality radiology data such as CT and PET. Most of them do not include other biological information (e.g., survival time, histopathological data, biomarkers) either. There is also lack of longitudinal radiology data to show disease evolution for patients receiving various treatments. With more complementary data and medical/treatment history (including radiomic data) that help comprehensively describe the brain tumor/disease status, an improvement in the accuracy of radiomic analysis and prediction can be expected.
Application of AI/ML in GBM diagnosis and therapeutic monitoring
The early brain tumor radiomic studies often relied on conventional radiomic feature-based ML methods that extract relatively explicit image texture features (e.g., shape, GLCM, LBP) to train traditional ML models such as SVMs (20–22) and RFs (23–26) in order to differentiate brain tumor versus non-tumors (or different types of brain tumors), predictions of overall survival, etc. Recently, by taking advantage of deep neural networks (DNNs) that include more neurons and layers to statistically recognize global, deep, and implicit imaging features, DL techniques can achieve state-of-the-art performance for automatic analysis of brain tumors on multimodality imaging and clinical data (32). Additionally, deep feature-based ML techniques build statistically/biologically meaningful models or utilize DNNs to extract deep implicit features from the radiology images and then apply traditional ML models for classifications (93, 94). Despite their differences, the above ML models all exploit prior biomedical and image features knowledge to 1) preprocess the radiological imaging data to extract imaging and/or biological meaningful features and 2) optimize the ML structure/algorithm for specific classification/segmentation tasks. Examples include the stacked denoising autoencoders (95) and the Convolutional Restricted Boltzman Machine (96). All these ML models have been applied in radiomic analysis to address unmet needs in GBM diagnosis, therapeutic monitoring, and/or prognosis (e.g., brain tumor classifications, survival predictions and biomarker identifications), while at this point, CNN-based DL enjoys the most generalizability and highest accuracy. Details are further discussed under the context of individual study case as follows.
GBM diagnosis and classification
One of the major ML applications in GBM radiomic analysis is to facilitate the differentiation between GBM and other histopathological processes. More specifically, such applications mainly fall into three categories: 1) distinguish brain tumor from other non-cancerous pathologies; 2) distinguish GBM from other brain tumors; and 3) differentiate between true progression and treatment effect (PsP or radiation necrosis). We hereby provide an overview of a few of these classification problems in brain tumor diagnosis.
1) Differentiating tumor from non-tumor
One of the most critical radiomic functions in brain tumors is to distinguish between malignant brain tumor and non-tumor pathologies, which include tumefactive demyelination, infection, inflammation (e.g., paraneoplastic syndromes and autoimmune disease), cortical dysplasia, and stroke. However, due to insufficient data available for ML training of each specific non-tumor type, existing studies mostly classify all data into two major categories: tumor (e.g., GBM/HGG, metastases, LGG) and non-tumor (i.e., control/normal and non-cancerous pathologies such as inflammation). Some studies tested their ML models on their own private data, while others took advantage of public datasets (e.g., BraTS, TCIA, Harvard brain atlas) or a combination of private and public data to expand the model’s generalizability. As discussed in Section 2.4, data preprocessing (e.g., filtering and feature extractions) are also often used to denoise and enhance the lesion region in the input MRI slides, with a hope to speed up the ML training process and improve the accuracy of distinguishing between tumors and non-tumors.
For those using public datasets, Ari et al. (97) proposed a three-phase extreme learning machine local receptive field (ELM-LRF) method for tumor classifications: removal of the noise using local and non-local methods, segmentation of benign or malignant tumor using ELM-LRF, and then use of a CNN classification. As a result, they achieved an effective classification accuracy of 0.97. Mohsen et al. (84) took advantage of a discrete wavelet transform (DWT) for feature extraction and principal component analysis (PCA) for reduction, together with a fuzzy C-means DNN to classify a dataset of 66 brain MRIs from Harvard Brain Atlas into four classes, i.e., normal, glioblastoma, brain sarcoma, and brain metastatic bronchogenic carcinoma tumors. An accuracy of 0.97 was achieved, and an area under the curve (AUC) approximated 0.984.
Alves et al. (83) quantified the gray-level pattern, pixel interrelationships, and the spectral properties of an image and achieved two fundamental features from an MRI sample, i.e., GLCM and gray-level run-length (GLRL). By combining this texture analysis with ML models (e.g., SVM, RF), they differentiated brain tumors from inflammatory lesions in their local MRI dataset and achieved a high accuracy of 0.83 and AUC of 0.906. Citak-Er et al. (93) applied a multiregional and multiparametric recursive feature elimination method, which was based on the Mann–Whitney ranking score, and then they employed the SVM-based multilayer perceptron (MLP) classification model to achieve a tumor detection accuracy of 0.93.
Amin et al. (41) mixed the public datasets and local datasets to differentiate tumors and non-tumors. They employed a Weiner filter to denoise and enhance the lesion region in the input MRI slides and used potential field (PF) clustering to identify the tumor region. Gabor wavelet transform (GWT) and LBP features were fused with various ML models (i.e., SVM, DT, k-NN, and naïve Bayes) to further improve the classification accuracy. The approach yielded an accuracy greater than 0.93 and an AUC of 0.96. Zhou et al. (98) treated holistic 3D MRI samples as sequences of 2D slices to extract some 3D features on brain tumors. They introduced a recursive structure, i.e., the long short-term memory (LSTM), to a deep CNN model (i.e., DenseNet) to handle such sequential data classification, and this DenseNet-LSTM model achieved an outstanding accuracy of 0.92 using the BraTS dataset (99).
Both Banerjee et al. (100) and Xu et al. (101) introduced transfer learning (TL) to improve the accuracy of the DNN-based ML classifier with non-brain-tumor images. They first pretrained the ML classifier with the large general image dataset, ImageNet (102), and then they fixed the pretrained parameters in the CNN hidden layers and fine-tuned the parameters in the output layers with neuro-oncology MRI images. Xu et al. (101) even embedded an SVM with the CNN to distinguish between GBM and LGG. By doing so, the two studies achieved classification accuracies of 0.97 and 0.975, respectively.
2) Differentiating primary from secondary brain tumors
Secondary/metastatic brain tumors have as high as fivefold incidence of that of primary brain tumors and manifest a rapid growth, causing significant brain tissue damages. Patients typically present with multiple metastatic tumors throughout the brain (103). A traditional non-ML-based approach to distinguish multifocal GBM from metastases on α[11C]-methyl-L-tryptophan (AMT)-PET images is to examine the tumoral standardized uptake values (SUVs), mean tumor/cortex SUV ratio, and tumor/cortex volume of distribution (VD)-ratio (104). Compared to GBM, metastases had lower values of all three parameters. However, this approach can only achieve an accuracy of 0.72.
Many studies implemented a combination of various feature extractions and regular ML models to find the best performance for their applications. Zacharaki et al. (86) introduced a Gabor texture filter with feature ranking to extract tumor features, and derived feature ranking scores, and then applied three ML models to distinguish GBM from metastases: SVM with recursive feature elimination (SVM-RFE), linear discriminant analysis (LDA, also known as Fisher linear discriminant) with Fisher’s discriminant rule (105), and k-NN. Among the three models, SVM-RFE achieved the highest mean accuracy and AUC of 0.91 and 0.936, respectively. Chen et al. (85) compared 30 diagnostic models that were built based on five feature selection models and six classification algorithms for distinguishing GBM and metastases. The five feature selection models included distance correlation, RF, least absolute shrinkage and selection operator (LASSO), eXtreme gradient boosting (XGBoost), and gradient boosting decision tree (GBDT), while the six classification algorithms included LDA, SVM, RF, k-NN, Gaussian naïve bayes (GaussianNB), and logistic regression (LR). The results showed that the combinational model of distance correlation and LR outperformed all other combinations in terms of testing accuracy (0.79) and AUC (0.80), although some other combinations achieved similar results as well.
Priya et al. (106) analyzed 60 GBM and 60 metastases cases with 12 regression or ML-based classifier models and four feature reduction/selection strategies—45 combinations in total. According to their results, the mean performance of various models was slightly better with FLAIR images than multiparametric sequences in terms of AUC, while the combination of full feature and LASSO achieved the highest AUC of 0.953, although full features with other models, such as ElasticNet (107) and RF, achieved similar results. de Causans et al. (55) trained on T1 MRI data with 71 GBM and 72 metastasis cases using 100 extracted features, based on the Image Biomarker Standardization Initiative (IBSI) (108). With these selected features, a total of 144 models combining nine feature scaling methods and 16 classifiers (regression and ML-based) were compared. All 144 classifiers of the 21 GBM and 16 metastases cases achieved a mean accuracy and AUC of 0.8 and 0.85, respectively.
3) Differentiating GBM from primary central nervous system lymphoma
GBM and primary central nervous system lymphoma (PCNSL) are not only common intracranial malignancies but also often share similarities in radiological appearance. However, the management for each disease is quite different (109). Recently, multiple ML-based predictive analytics have arisen to help differentiate GBM from PCNSL radiologically with a relatively high sensitivity and specificity. Outcomes were assessed based on test characteristics such as accuracy, sensitivity, specificity, and AUC.
Due to the binary classification nature of distinguishing between GBM and PCNSL, SVM used to be the most commonly employed model for its computational simplicity (110–115). Other commonly used ML models such as k-NN and RF were also exploited and compared (114, 116). Even though the training datasets were relatively small, approximately 110 or fewer samples with about two-thirds of the entire dataset containing GBM and one-third of PCNSL, it turned out that most ML-based models were able to achieve an accuracy between 0.9 and 0.96, a sensitivity of 0.84 or higher, a specificity of 0.89 or higher, and AUC of 0.92 or higher. More notably, Kang et al. (114) and Suh et al. (116) even compared the prediction outcomes of their ML models with the prediction from human radiologists, and the results showed the superiority of ML models over human radiologists in all four criteria, especially in accuracy, sensitivity, and AUC. It is unknown whether the combination of the ML model and human radiologist read would have attained even higher accuracy, sensitivity, and AUC.
Recently, more sophisticated DL models have been employed. Priya et al. (117) examined five different ML approaches (i.e., LASSO, SVM, RF, Ridge, and MLP) to distinguish between 97 GBM and 46 PCNSL cases, with all five approaches sharing similar results. Yet, LASSO had the best performance (0.88 in accuracy and 0.92 in AUC) when using features from the whole tumor, while MLP had the best performance (0.86 in accuracy and 0.91 in AUC) when only using the features from the single largest slice. For an even larger dataset (i.e., 160 GBM and 160 PCNSL), McAvoy et al. (118) applied a CNN variant, EfficientNet (119), and by using TL based on ImageNet, they achieved an accuracy of 0.93 and AUC of 0.94.
4) Differentiating treatment effects versus true disease progression
Pseudoprogression (PsP) is the apparent growth of a lesion or development of new lesions on imaging that represents inflammatory treatment-related changes but looks just like viable tumor growth on MRI. PsP is most common between 3 and 6 months after the completion of radiation therapy, and the corresponding imaging findings will subside on their own over time (120). PsP is more likely in MGMT promoter-methylated tumors treated with temozolomide. The increased contrast enhancement on MRI may be caused by the increased vascular permeability from cytotoxic therapies including radiotherapy and chemotherapies such as temozolomide, which may benefit patients receiving immunotherapy and temozolomide but often leads to premature discontinuation of treatment owing to the false judgment of progression of disease (121). Radiation necrosis is another treatment effect that can occur any time after radiation therapy but is most common 1–2 years after radiation. It should be differentiated from true progression of viable tumor before treatment changes are contemplated. As a result, accurate differentiation between treatment effect (i.e., pseudoprogression or radiation necrosis) and true tumor progression is critical in the treatment decision. PsP may be associated with a survival advantage. A key radiology tool in differentiating pseudoprogression or radiation necrosis from true progression of disease is dynamic susceptibility contrast (DSC) MR perfusion-weighted imaging (PWI). Elevated corrected relative cerebral blood volume (crCBV) relative to normal-appearing white matter is more common in a viable tumor than in treatment effects (122). However, PWI is unreliable in patients treated with immunotherapy such as immune checkpoint inhibitors, such that, per immunotherapy response assessment in neuro-oncology (iRANO), the patient is followed for 3 months and then a determination is made of whether the initial increase in size of the lesion represented treatment effects or a viable tumor (123–125).
Booth et al. (126) first analyzed the tumor heterogeneity in T2 MRI using topological descriptors called Minkowski functionals (MFs). Then they utilized an SVM model, together with image features such as MFs, size, and signal intensity, to distinguish between pseudoprogression and true progression, and achieved an accuracy of 0.88, slightly higher than using RF for feature selection and LASSO for classification (0.86). Hu et al. (127) took advantage of T1 MRI and other eight-dimensional feature vectors, including T2, FLAIR, proton density, ADC, PWI, derived relative cerebral blood volume (rCBV), relative cerebral blood flow (rCBF), and mean transit time maps, to train an SVM model, and achieved an AUC of 0.94 in distinguishing between pseudoprogression and true progression. The ADC map derived from DWI and rCBV and rCBF derived from PWI were found to make a greater contribution to the discrimination than the conventional radiology images do.
Due to the time correlation embedded in true progression and PsP radiology data, Lee et al. (128) and Jang et al. (129) exploited recursive LSTM-CNN structures on MRI to distinguish between the two occurrences. In comparison to Lee’s multimodal MRI data (i.e., T1, T2, FLAIR), Jang et al. (129) combined/fused T1 MRI data with clinical features to develop an LSTM-CNN clinic-feature-fused model and achieved an AUC of 0.87 and F1 score of 0.74, outperforming the model trained with MRI data only and the RF-based model.
Akbari et al. (130) employed TL with a CNN pretrained on ImageNet and feature extraction based on four structural MRIs (i.e., T1, T1-ce, T2, FLAIR), diffusion tensor imaging (DTI), and PWI (rCBV, peak height (PH), percentage signal recovery (PSR)) and achieved an accuracy of 0.84 and AUC of 0.83. Ismail et al. (131) extracted 30 global and local shape features from T1-ce, T2, and FLAIR images and used an SVM classifier to achieve an accuracy of 0.90 in distinguishing PsP from true tumor progression.
In addition to the ambiguity between PsP and true tumor progression, immunotherapies in GBM also suffer from the lack of reliable evaluation methods on the radiological imaging manifestation regarding the alteration of the tumor immune microenvironment (TME, e.g., tumor immune cell infiltration, functional characterization of immune effector/suppressive cells, gene expression profile of immunostimulatory/immunosuppressive cells), a crucial parameter for assessment of intratumoral immune responses (5). In their pioneering work (132), Narang et al. utilized T1-weighted post-contrast and T2-FLAIR images in combination with T-cell surface marker CD3D/E/G mRNA expression data from 78 GBM patient-derived TCGA data to extract six imaging features that are associated with intra-GBM CD3 activity. These imaging features were further trained and tested using an internal dataset from 69 GBM patients that has immunohistochemically (IHC) validated intratumoral CD3 counts. The image-based intra-GBM CD3+ T-cell infiltration model reaches an accuracy of 97.1% and AUC of 0.993 for the training set, with an accuracy of 76.5% and AUC of 0.847 in the test group. A similar study has been reported recently in lower-grade gliomas (LGG) with an expansion from CD3 expression data to multiple immune gene expression profiles, including major histocompatibility complex (MHC)-related molecules, immune checkpoint molecules, and effector/suppressor immune cells (94). In this study, radiomic features extracted by a deep learning neural network-based model have been demonstrated to predict the TME-associated signature immunophenotype mRNAs with an AUC of 0.821 in the test group. Unfortunately, there is no IHC validation on expression of signature immune genes in the test group specimens.
To date, although numerous ML models for differentiating PsP from true tumor progression have been proposed and tested, none of them have been prospectively validated, reflecting the lack of confidence in clinicians to apply these radiomic approaches in their clinic practice. Multiple factors can lead to this significant issue, such as difficulty to applying small sample size-derived prediction models to a large population cohort, poor reproducibility, and lack of consistency between various ML models and/or datasets (further discussed in Section 4). One of the important and challenging factors is that currently there is no clear objective histological definition of pseudoprogression. In a representative study by Melguizo-Gavilanes et al. (133), MRI images and surgical resection-derived histological data from 34 patients with GBM were retrospectively reviewed. Only one-third of the cohort (11/34) demonstrated a concordance for PsP between radiological interpretation and histological diagnosis, whereas the majority of the patients had a histologically “mixed” pattern with tumor and treatment effect, indicating that even histology might not be applied as a gold standard to differentiate PsP and tumor true progression.
Overall survival prediction
Overall survival prediction of GBM patients provides useful information for surgical and treatment planning. Conventional survival prediction based on clinical information is subjective and could be inaccurate. Radiomic analysis, on the other hand, provides a variety of MRI features to predict disease prognosis, thus providing beneficial information for personalized treatment. Nevertheless, manual feature engineering is still time consuming, laborious, and subjective and may not be able to effectively encode other predictive but implicit information hidden in the multimodal neuroimages (134). Thus, an accurate, generalized yet automated OS prediction is desired.
Macyszyn et al. (135) extracted about 60 features from 105 GBM patients to train an SVM-based predictive model for patient survival and molecular subtype. The predictors were evaluated in 29 new patients and achieved a three-way (long/medium/short survival for longer than 18 months, between 6 and 18 months, and shorter than 6 months) accuracy of about 0.80. Another classifier was trained to discriminate among each of the various GBM molecular subtypes and achieved an accuracy of about 0.76. Sanghazni et al. (136) derived texture features (e.g., first-order texture features, GLCM), tumor shape and volumetric features, and patient ages from 173 patients’ multimodal MRI data (e.g., T1-ce, T2, and Flair) and used an SVM-RFE-based ML model to perform binary (i.e., short and long‘s threshold upon 400 days) and multiclass (i.e., <10, 10~15, and >15 months) OS prediction. Prediction accuracies of 0.987 and 0.89 were achieved for binary and multiclass predictions, respectively.
Choi et al. (137) collected 250 radiomic features extracted from 296 LGG cases from institutional and TCGA/TCIA datasets. They trained three random survival forest (RSF, i.e., a variant of RF) models with 1) these radiomic features; 2) non-imaging prognostic factors including age, resection extent, WHO grade, and IDH status; and 3) combination of 1 and 2 on the institutional dataset and validation of the model on the TCGA/TCIA dataset. When applying radiomic features or non-imaging features alone, the two RSF models achieved an AUC of 0.620 and 0.627, respectively. When applying radiomic features together with non-imaging prognostic parameters, the AUC was improved to 0.709. Similarly, in a GBM hypoxia-associated radiomic study, Beig et al. (138) also revealed that when combining clinical features (age, gender, and Karnofsky Performance Score (KPS)) with 270 radiomic features, the concordance index for survival prediction rises to 0.83 in comparison to 0.74 when using radiomic features alone (138). Grist et al. (139) examined various analysis techniques on survival predictions through perfusion and MRI data, especially DWI, collected from 69 pediatric patients. Approaches included conventional regressions and Bayesian analysis on apparent diffusion coefficient (ADC) maps, uncorrected and corrected cerebral blood volume (uCBV and cCBV) maps, and K2 maps (140) and achieved an AUC between 0.63 and 0.82. Supervised (i.e., SVM, RF, and a single-layer neural network) and unsupervised (i.e., k-means clustering) ML analyses achieved an accuracy between 0.90 and 0.98 in distinguishing between high- and low-risk clusters, with distinct differences in survival. In addition to the above models, the Tiwari group has developed a radiomic risk score in which the extracted GBM radiomic features were trained by various Cox regression-based algorithms for survival stratification with an overall concordance index at 0.7 to 0.8 (141–143).
Nie et al. (134) proposed a two-stage learning-based method to predict the OS of HGG patients. Specifically, in the first stage, they adopted a CNN to extract implicit features from multiparametric maps that are computed by multimodal multichannel MRI (i.e., T1-ce, DTI, and rs-fMRI) from 68 HGG patients. Then, those radiomic features along with the demographic and tumor-related features (e.g., age, tumor size, and histological type) were trained in an SVM to model OS prediction (i.e., long or short overall survival time, with a threshold of 650 days). The experimental results demonstrated an accuracy of up to 0.91.
Identifying biomarkers of brain tumors
Radiogenomics uses radiomics techniques to predict the genetic makeup of tumors. This promotes precision medicine by identifying patients with tumor molecular markers that can be targeted by particular drugs and by predicting how aggressive a tumor will behave, with implications for survival and treatment choice. Via exploring the implicit correlation between radiological images and genomic data such as DNA microarrays, microRNA, RNA-Seq, ML techniques can help improve the effectiveness and efficiency in identifying the biomarkers of brain tumors (144).
Isocitrate dehydrogenase (IDH) mutation
Since the initial reworking of the WHO CNS Tumor Classification System in 2016, genetic biomarkers have become increasingly important in the classification of brain tumors. Isocitrate dehydrogenase is an enzyme in the Krebs cycle, and its mutated gene (IDH) is an oncogene. The mutant IDH enzyme produces an oncometabolite 2-hydroxyglutarate (2HG) (145), which promotes the growth of various cancers throughout the body. In brain tumors, IDH-mutated tumors are less aggressive than IDH wild-type tumors, yet they can convert to the latter. In the 2021 WHO CNS Tumor Classification System, only IDH wild-type tumors are classified as GBMs. It is of utmost importance for therapeutic planning to differentiate between the IDH mutation and IDH wild type, and it would greatly benefit patients if this determination could be done non-invasively and obviate biopsy or resection. Yogananda et al. (33) developed a 3D Dense-UNet network using (a) T2 images only (T2-net) and (b) a combination of T1-ce, T2, and FLAIR images (TS-net) from TCIA and TCGA to non-invasively predict IDH mutation. The T2-net demonstrated a mean cross-validation accuracy of 0.97 (sensitivity 0.97, specificity 0.98, AUC 0.98), and TS-net demonstrated a mean cross-validation accuracy of 0.97 (sensitivity 0.98, specificity 0.97, AUC 0.99). In addition, this model automatically segmented the tumor to show areas with either IDH mutation or IDH wild type. Dice scores were 0.85 for T2-net and 0.89 for TS-net. The benefit of being able to use only T2-weighted images is that gadolinium-based contrast material, which deposits in the brain to unknown effect, does not have to be administered and T2-weighted images can be quickly acquired and are less sensitive to motion artifact.
MGMT promoter methylation
MGMT promoter methylation predicts less aggressive glioma behavior for both IDH-mutated and IDH-wild-type gliomas. When its promoter is methylated, the MGMT gene, which is involved in DNA repair, is hindered and the tumor has greater difficulty overcoming the damage caused by chemotherapy such as temozolomide. Yogananda et al. (146) used a 3D-dense UNet on only T2 images to simultaneously segment the tumor and predict the presence of MGMT promoter methylation with a mean three-fold cross-validation accuracy of 0.95 (sensitivity 0.96, specificity 0.92, AUC 0.93, Dice score 0.82).
H3K27M alterations
In 2016, the WHO released a new histological diagnosis in the classification of CNS malignancies: diffuse midline glioma (DMG), H3K27M-mutant. It was renamed as H3K27M-altered in 2021 because there are multiple mechanisms involved. These WHO grade 4 tumors are found in or near the midline in the brainstem, thalamus, spinal cord, pineal region, hypothalamus, and cerebellum and exhibit aggressive clinical behavior (147, 148). H3K27M is the most frequent mutation in brainstem gliomas (BSGs) (149). Su et al. (150) extracted radiomics features from FLAIR images from 40 patients with H3K27M mutations and 60 wild-type patients, all with midline gliomas. The Tree-based Pipeline Optimization Tool (TPOT) was applied to optimize the ML pipeline and select important radiomics features. A total of 10 independent TPOT ML models were compared and tested on 22 independent cohorts of patients, achieving an accuracy ranging from 0.6 to 0.84, and the AUC from 0.73 to 0.90. Pan et al. (149) included a total of 151 patients with newly diagnosed BSGs. A total of 1,697 features, including six clinical parameters and 1,691 imaging features (e.g., GLCM, LBP), were extracted from pre- and post-contrast T1 and T2 images. Spearman’s correlation and relief algorithm were applied for feature selection. Thirty-six MRI features and three clinical features remained and were fed to an RF model to predict H3K27M mutations. For comparison, a least-square estimation method-based ML model was developed by utilization of the KPS at diagnosis, symptom duration at diagnosis, and edge sharpness on T2, which achieved an accuracy of 0.80 and AUC of 0.79 in the test cohort if using MRI features alone but can be improved to 0.84 and an AUC of 0.83 if integrated with clinical parameters. The simplified model achieved an AUC of 0.78. Zhuo et al. (151) studied 81 BSG patients with APT imaging at 3T MR and known H3K27M status. APTw values (i.e., mean, median, and max) and radiomic features within manually delineated 3D tumor masks were extracted. H3K27M-mutant prediction using APTw-derived radiomics was conducted using various models, such as SVM, AdaBoost, autoencoder, LASSO regression, and RF, which achieved an accuracy of 0.86 and an AUC of 0.93 as validated by a prospective cohort of 29 BSG patients.
Discussion
Despite that numerous ML studies have been conducted in GBM radiomic analysis, comparing the results from individual articles is not a trivial task due to the use of different data sets. The accuracies, AUCs, and Dice scores in different studies may vary from 0.7 to 0.98: most state-of-the-art studies using public datasets (e.g., BraTS) achieve an accuracy of 0.84–0.94, but some studies with certain private data can reach 0.98. Meanwhile, current major public datasets also lack sub-categories for brain tumor classification and segmentations, which restricts the development of a more powerful and comprehensive ML-model to distinguish more brain tumor types. Without sufficiently large datasets, ML models with too many parameters (i.e., neurons in each layer and the number of layers) are easily overfitting to a specific dataset, losing the generalizability of the model to other patient groups.
Challenges and perspectives on future AI/ML techniques
Overview of current challenges in ML-based radiomic neuro-oncology studies
As ML is a data-driven statistical approach to extract common features within different data samples, sufficient imaging datasets are required to train advanced ML models and to fairly evaluate their performances (e.g., accuracy, Dice score, AUC) in the field of neuro-oncology. Currently, only a limited number of brain tumor sub-categories have been analyzed with ML studies while many other brain tumor/disease types have not, due to the lack of labeled/annotated data for training. Examples include differentiating dysembryoplastic neuroepithelial tumor (DNET), ganglioglioma, pleomorphic xanthoastrocytoma (PXA), and multinodular and vacuolating neuronal tumor (MVNT).
However, establishing standardized radiological imaging datasets or standardizing McMv datasets for extensive and generalized ML-based GBM analysis can be manpower and time consuming because most of these datasets require highly accurate manual labeling/annotation to serve as the “ground truth” for the ML model training and validation. In addition, these datasets should be generalizable for various neuro-oncology analyses and patient groups and should be carefully labeled by various disease categories. Additional information (e.g., survival time, related biotest results, related clinical/medical history) may also be necessary for more sophisticated and comprehensive analyses. This requires a continuous update of the datasets, leading to a significant cost of data management (Figure 1B).
Another challenge is that current mathematic mechanisms in the ML model are based on statistics, which means there may not be a “deterministic optimal” algorithm or architecture for an ML model to achieve the “ideal/optimal” outcomes. The initial values of the trainable parameters in ML models and the slight differences in structure may affect the training outcome significantly. Even when using public datasets (e.g., BraTS), similar ML networks may yet achieve varying results (152, 153). Thus, many researchers intend to simply add more layers in CNNs to improve the accuracy, potentially causing extensive yet unnecessary computational complexity during the training process but overlooking the biological connections and meaning behind those data. On the other hand, too much engineering (i.e., strong feature extraction, data restriction/collection) in data preprocessing may also lead to overfitting of the ML network to the training data and lose the generalizability of trained ML models for larger populations with more diversity (Figure 1B).
Promising strategies enhancing performance of AI models in GBM radiomic analysis
Aside from using genuine radiological brain tumor images alone to train ML models, three other trends are gaining popularity to improve the model performance in accuracy, Dice score, AUC, and generalizability. The first trend is to use TL (100, 101), which takes advantage of other larger non-neuro-oncology or even non-medical image datasets to pretrain the ML model. Then by keeping the pretrained parameters in the low-level hidden layers (i.e., closer to the input layer) and fine-tuning the ones in the high-level layers and output layers with brain tumor training image datasets, the pretrained ML model can be adopted for brain tumor analysis. Typical image datasets for pretraining ML models includes ImageNet (102), the modified National Institute of Standards and Technology (MNIST) database (154), and International Symposium on Biomedical Imaging (ISBI) (155). However, if the pretraining dataset is drastically unsimilar to the target dataset, the pretraining effect is limited. Therefore, a standardized radiomic medical dataset with various categories is preferred, benefitting not only neuro-oncology studies but also other medical and biomedical studies.
The second trend is to use GANs to generate synthetic data for augmentation (78, 79, 155). However, as discussed in Section 2.4, this approach itself requires a large set of genuine images to train the discriminator network in the GAN, before it can synthesize accurate-appearing brain tumor images to train other ML models for brain tumor analysis.
The third trend is to fuse multimodal data for a more comprehensive analysis. Examples include multimodal MRI (156, 157), combinations of MRI and PET (54), image genomics (i.e., radiogenomics), and clinical data to study the association between imaging biomarkers and genomic characteristics (144, 149–151, 158, 159). Especially for radiogenomics, some studies (144, 158, 159) have identified associations between quantitative image features and gene expression profiles of glioblastoma (e.g., H3K27M, TP53, EGFR, NF1, and IDH1) and its molecular subtypes (e.g., classical, mesenchymal, proneural, and neural). Additional studies indicate that quantitative MR imaging features derived from entire tumor volumes can be used to identify glioblastoma subtypes with distinct molecular pathways (160, 161). With the help of additional complementary correlated features from different types of radiomic images and/or genomic information, or simply just the medical history of the patients, ML can take advantage of data to achieve more accurate predictions (Figure 1B).
Outlook on teamwork among computer scientists/engineers, physicians, and biomedical researchers
As aforementioned, high-quality clinical data and labels/annotations are critical to ML algorithms for both accuracy and generalizability, and biological knowledge can help extract certain features to improve the accuracy as well as the training efficiency. Therefore, strong collaborations should be established among computer scientists, engineers, physicians, and biomedical researchers to facilitate the standardization and enrichment of neuro-oncology radiomic datasets and the development of innovative and more advanced AI/ML models (Figure 1B). In addition, with larger amounts of data to track patients’ treatment process and the outcomes, it is even possible to develop ML/AI techniques to determine more suitable plans for their treatment, to improve the patients’ survival time as well as their quality of life.
Conclusion
With the urgent needs for highly accurate and automatic analysis of brain tumors and the rapid growth of clinical imaging data, image-based ML/AI techniques are playing an increasingly important role. Various combinations of feature extraction algorithms and ML models have been implemented and have achieved comparable or even better performance than manual analysis. However, challenges remain for exploring cancer heterogeneity, higher prediction accuracy, and generalizability for larger, more diverse patient groups. We believe that, by improving dataset quality, employing multimodal data fusion, developing more advanced ML models, and further enhancing collaborations between computer scientists, engineers, physicians, and biomedical researchers, AI techniques will accelerate quantitative cancer imaging analysis for clinical applications with great improvements in patient care.
Author contributions
MZ, YK, VH, LZ, and SZ formulated the conception of this review. MZ, SL, and SZ performed initial screening of the citations, datasets, and related AI/ML techniques. MZ, SL, LZ, and VH wrote and reviewed the manuscript and prepared the figures. AH provided critical comments for the manuscript. All authors approved the manuscript for publication and declare no competing financial interest.
Funding
This study was supported by the Nevada Governor’s Office of Economic Development (GOED) AWD-02-00001154 to SZ.
Acknowledgments
Thanks for UNLV University Libraries Open Article Fund to support this open-access journal article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Chang K, Zhang B, Guo X, Zong M, Rahman R, Sanchez D, et al. Multimodal imaging patterns predict survival in recurrent glioblastoma patients treated with bevacizumab. Neuro-Oncology (2016) 18:1680–7. doi: 10.1093/neuonc/now086
2. About GBM - GBM awareness day | national brain tumor society . Available at: https://braintumor.org/take-action/about-gbm/ (Accessed April 3, 2022).
3. Hanif F, Muzaffar K, Perveen K, Malhi SM, Simjee SU. Glioblastoma multiforme: A review of its epidemiology and pathogenesis through clinical presentation and treatment. Asian Pac J Cancer Prev (2017) 18:3. doi: 10.22034/APJCP.2017.18.1.3
4. Ostrom QT, Patil N, Cioffi G, Waite K, Kruchko C, Barnholtz-Sloan JS. CBTRUS statistical report: Primary brain and other central nervous system tumors diagnosed in the united states in 2013–2017. Neuro-Oncology (2020) 22:iv1–iv96. doi: 10.1093/neuonc/noaa200
5. Bagley SJ, Desai AS, Nasrallah MP, O’Rourke DM. Immunotherapy and response assessment in malignant glioma: Neuro-oncology perspective. Topics Magnet Resonance Imaging (2020) 29:95–102. doi: 10.1097/RMR.0000000000000233
6. Gillies RJ, Kinahan PE, Hricak H. Radiomics: Images are more than pictures, they are data. Radiology (2016) 278:563–77. doi: 10.1148/radiol.2015151169
7. Kumar V, Gu Y, Basu S, Berglund A, Eschrich SA, Schabath MB, et al. Radiomics: the process and the challenges. Magnet Resonance Imaging (2012) 30:1234–48. doi: 10.1016/j.mri.2012.06.010
8. Samuel AL. “Some studies in machine learning using the game of checkers. II–recent progress.,”. In: Levy DNL, editor. Computer games I. New York, NY: Springer (1988). p. 366–400. doi: 10.1007/978-1-4613-8716-9_15
9. Caruana R, Niculescu-Mizil A. An empirical comparison of supervised learning algorithms. proceedings of the 23rd international conference on machine learning. ICML ‘06. New York, NY, USA: Association for Computing Machinery (2006) p. 161–8. doi: 10.1145/1143844.1143865
10. Barlow HB. Unsupervised learning. Neural Comput (1989) 1:295–311. doi: 10.1162/neco.1989.1.3.295
11. Kaelbling LP, Littman ML, Moore AW. Reinforcement learning: A survey. J Artif Intell Res (1996) 4:237–85. doi: 10.1613/jair.301
12. Rajpurkar P, Chen E, Banerjee O, Topol EJ. AI In health and medicine. Nat Med (2022) 28:31–8. doi: 10.1038/s41591-021-01614-0
13. Voulodimos A, Doulamis N, Doulamis A, Protopapadakis E. Deep learning for computer vision: A brief review. Comput Intell Neurosci (2018) 2018:e7068349. doi: 10.1155/2018/7068349
14. Larrañaga P, Calvo B, Santana R, Bielza C, Galdiano J, Inza I, et al. Machine learning in bioinformatics. Briefings Bioinf (2006) 7:86–112. doi: 10.1093/bib/bbk007
15. Perlich C, Dalessandro B, Raeder T, Stitelman O, Provost F. Machine learning for targeted display advertising: transfer learning in action. Mach Learn (2014) 95:103–27. doi: 10.1007/s10994-013-5375-2
16. Brownlee J. Deep learning for natural language processing: develop deep learning models for your natural language problems. Mach Learn Mastery (2017).
17. Baid U, Ghodasara S, Mohan S, Bilello M, Calabrese E, Colak E, et al. The RSNA-ASNR-MICCAI BraTS 2021 benchmark on brain tumor segmentation and radiogenomic classification (2021). Available at: http://arxiv.org/abs/2107.02314 (Accessed March 14, 2022).
18. Kotsiantis SB, Kanellopoulos D, Pintelas PE. Data preprocessing for supervised leaning. Int J Comput Sci (2006) 1:111–7. doi: 10.5281/zenodo.1082415
19. Alpaydin E. Introduction to machine learning. 4th ed. Cambridge, MA, USA: MIT Press (2020). 712 p.
20. Ruan S, Lebonvallet S, Merabet A, Constans J. Tumor segmentation from a multispectral MRI images by using support vector machine classification, in: 2007 4th IEEE International Symposium on Biomedical Imaging: From Nano to Macro, Washington DC: IEEE (2007). pp. 1236–9. doi: 10.1109/ISBI.2007.357082
21. Li H, Song M, Fan Y. Segmentation of brain tumors in multi-parametric MR images via robust statistic information propagation. In: Kimmel R, Klette R, Sugimoto A, editors. Computer vision – ACCV 2010. lecture notes in computer science. Berlin, Heidelberg: Springer (2011). p. 606–17. doi: 10.1007/978-3-642-19282-1_48
22. Li H, Fan Y. Label propagation with robust initialization for brain tumor segmentation, in: 2012 9th IEEE International Symposium on Biomedical Imaging (ISBI), Barcelona, Spain: IEEE. (2012) pp. 1715–8. doi: 10.1109/ISBI.2012.6235910
23. Reza S, Iftekharuddin KM. Improved brain tumor tissue segmentation using texture features. In: MICCAI BraTS (Brain tumor segnmentation challenge). Boston, MA (2014). p. 27–30.
24. Götz M, Weber C, Blöcher J, Stieltjes B, Meiner H-P, Maier-Hein KH. Extremely randomized trees based brain tumor segmentation. In: MICCAI BraTS challenge. Boston, MA, USA: Zenodo (2014). p. 6–11. doi: 10.5281/zenodo.3666272
25. Kleesiek J, Biller A, Urban G, Hamprecht FA. Ilastik for multi-modal brain tumor segmentation. In: MICCAI BraTS (Brain tumor segmentation challenge). Boston, MA, USA:CiteSeerx (2014). p. 12–7.
26. Meier R, Bauer S, Slotboom J, Wiest R, Reyes M. Appearance-and context-sensitive features for brain tumor segmentation. In: MICCAI BraTS challenge. Boston, MA, USA (2014). p. 20–6. doi: 10.13140/2.1.3766.7846
27. Girshick R, Donahue J, Darrell T, Malik J. Rich feature hierarchies for accurate object detection and semantic segmentation, in: 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH: IEEE. (2014) pp. 580–7. doi: 10.1109/CVPR.2014.81
28. Zhao Z-Q, Zheng P, Xu S-T, Wu X. Object detection with deep learning: A review. IEEE Trans Neural Networks Learn Syst (2019) 30:3212–32. doi: 10.1109/TNNLS.2018.2876865
29. Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. IEEE Comput Soc (2015) 3431–40. doi: 10.1109/CVPR.2015.7298965
30. Zheng S, Jayasumana S, Romera-Paredes B, Vineet V, Su Z, Du D, et al. Conditional random fields as recurrent neural networks, in: 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile: IEEE. (2015) pp. 1529–37. doi: 10.1109/ICCV.2015.179
31. Liu Z, Li X, Luo P, Loy C-C, Tang X. Semantic image segmentation via deep parsing network, in: 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile: IEEE (2015). pp. 1377–85. doi: 10.1109/ICCV.2015.162
32. Zhao X, Wu Y, Song G, Li Z, Zhang Y, Fan Y. A deep learning model integrating FCNNs and CRFs for brain tumor segmentation. Med Image Anal (2018) 43:98–111. doi: 10.1016/j.media.2017.10.002
33. Bangalore Yogananda CG, Shah BR, Vejdani-Jahromi M, Nalawade SS, Murugesan GK, Yu FF, et al. A novel fully automated MRI-based deep-learning method for classification of IDH mutation status in brain gliomas. Neuro-Oncology (2020) 22:402–11. doi: 10.1093/neuonc/noz199
34. Menze BH, Jakab A, Bauer S, Kalpathy-Cramer J, Farahani K, Kirby J, et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans Med Imaging (2015) 34:1993–2024. doi: 10.1109/TMI.2014.2377694
35. Clementi G, Salvati D, Scorzelli G, Paoluzzi A, Pascucci V. Progressive extraction of neural models from high-resolution 3D images of brain, in: Proceedings of CAD’16. CAD Solutions LLC, Vancouver, Canada: CAD Solutions, LLC (2016). pp. 348–51. doi: 10.14733/cadconfP.2016.348-351
36. Talo M, Baloglu UB, Yıldırım Ö, Rajendra Acharya U. Application of deep transfer learning for automated brain abnormality classification using MR images. Cogn Syst Res (2019) 54:176–88. doi: 10.1016/j.cogsys.2018.12.007
37. Emblem KE, Zoellner FG, Tennoe B, Nedregaard B, Nome T, Due-Tonnessen P, et al. Predictive modeling in glioma grading from MR perfusion images using support vector machines. Magnet Resonance Med (2008) 60:945–52. doi: 10.1002/mrm.21736
38. Sessions V, Valtorta M. The effects of data quality on machine learning algorithms. ICIQ (2006) 485–98.
39. Tests for brain and spinal cord tumors in adults . Available at: https://www.cancer.org/cancer/brain-spinal-cord-tumors-adults/detection-diagnosis-staging/how-diagnosed.html (Accessed February 8, 2022).
40. Thust SC, Yousry T. Imaging of skull base tumours. Rep Pract Oncol Radiother (2016) 21:304. doi: 10.1016/j.rpor.2015.12.008
41. Amin J, Sharif M, Raza M, Saba T, Anjum MA. Brain tumor detection using statistical and machine learning method. Comput Methods Programs Biomed (2019) 177:69–79. doi: 10.1016/j.cmpb.2019.05.015
42. Tandel GS, Biswas M, Kakde OG, Tiwari A, Suri HS, Turk M, et al. A review on a deep learning perspective in brain cancer classification. Cancers (2019) 11:111. doi: 10.3390/cancers11010111
43. Hou DJ, Tong KA, Ashwal S, Joo E, Shutter L, Obenaus A. 478 diffusion-weighted imaging improves outcome prediction in adults with diffuse axonal injury. J Invest Med (2006) 54:S161–1. doi: 10.2310/6650.2005.X0004.477
44. Landheer K, Schulte RF, Treacy MS, Swanberg KM, Juchem C. Theoretical description of modern 1H in vivo magnetic resonance spectroscopic pulse sequences. J Magnet Resonance Imaging (2020) 51:1008–29. doi: 10.1002/jmri.26846
45. Golder W. Magnetic resonance spectroscopy in clinical oncology. ORT (2004) 27:304–9. doi: 10.1159/000077983
46. Rosen Y, Lenkinski RE. Recent advances in magnetic resonance neurospectroscopy. Neurotherapeutics (2007) 4:330–45. doi: 10.1016/j.nurt.2007.04.009
47. Chen F, Ni Y-C. Magnetic resonance diffusion-perfusion mismatch in acute ischemic stroke: An update. World J Radiol (2012) 4:63–74. doi: 10.4329/wjr.v4.i3.63
48. Yacoub E, Harel N, Uğurbil K. High-field fMRI unveils orientation columns in humans. PNAS (2008) 105:10607–12. doi: 10.1073/pnas.0804110105
49. Zhou J, Blakeley JO, Hua J, Kim M, Laterra J, Pomper MG, et al. Practical data acquisition method for human brain tumor amide proton transfer (APT) imaging. Magnet Resonance Med (2008) 60:842–9. doi: 10.1002/mrm.21712
50. Sartoretti E, Sartoretti T, Wyss M, Reischauer C, van Smoorenburg L, Binkert CA, et al. Amide proton transfer weighted (APTw) imaging based radiomics allows for the differentiation of gliomas from metastases. Sci Rep (2021) 11:5506. doi: 10.1038/s41598-021-85168-8
51. Harã T, Kosaka N, Shinoura N, Kondo T. PET imaging of brain tumor with [methyl-11C] choline. J Nucl Med (1997) 38:842–7.
52. Langleben DD, Segall GM. PET in differentiation of recurrent brain tumor from radiation injury*. J Nucl Med (2000) 41:1861–7.
53. Chao ST, Suh JH, Raja S, Lee S-Y, Barnett G. The sensitivity and specificity of FDG PET in distinguishing recurrent brain tumor from radionecrosis in patients treated with stereotactic radiosurgery. Int J Cancer (2001) 96:191–7. doi: 10.1002/ijc.1016
54. Galldiks N, Law I, Pope WB, Arbizu J, Langen K-J. The use of amino acid PET and conventional MRI for monitoring of brain tumor therapy. NeuroImage: Clin (2017) 13:386–94. doi: 10.1016/j.nicl.2016.12.020
55. de Causans A, Carré A, Roux A, Tauziède-Espariat A, Ammari S, Dezamis E, et al. Development of a machine learning classifier based on radiomic features extracted from post-contrast 3D T1-weighted MR images to distinguish glioblastoma from solitary brain metastasis. Front Oncol (2021) 11:638262. doi: 10.3389/fonc.2021.638262
56. Hellier P. Consistent intensity correction of MR images, in: Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429), Catalonia, Spain: IEEE (2003). pp. I–1109. doi: 10.1109/ICIP.2003.1247161
57. Shah M, Xiao Y, Subbanna N, Francis S, Arnold DL, Collins DL, et al. Evaluating intensity normalization on MRIs of human brain with multiple sclerosis. Med Image Anal (2011) 15:267–82. doi: 10.1016/j.media.2010.12.003
58. Sweeney EM, Shinohara RT, Shea CD, Reich DS, Crainiceanu CM. Automatic lesion incidence estimation and detection in multiple sclerosis using multisequence longitudinal MRI. AJNR Am J Neuroradiol (2013) 34:68–73. doi: 10.3174/ajnr.A3172
59. Collewet G, Strzelecki M, Mariette F. Influence of MRI acquisition protocols and image intensity normalization methods on texture classification. Magnet Resonance Imaging (2004) 22:81–91. doi: 10.1016/j.mri.2003.09.001
60. Shinohara RT, Sweeney EM, Goldsmith J, Shiee N, Mateen FJ, Calabresi PA, et al. Statistical normalization techniques for magnetic resonance imaging. NeuroImage: Clin (2014) 6:9–19. doi: 10.1016/j.nicl.2014.08.008
61. Nyúl LG, Udupa JK. On standardizing the MR image intensity scale. Magnet Resonance Med (1999) 42:1072–81. doi: 10.1002/(SICI)1522-2594(199912)42:6<1072::AID-MRM11>3.0.CO;2-M
62. Sun X, Shi L, Luo Y, Yang W, Li H, Liang P, et al. Histogram-based normalization technique on human brain magnetic resonance images from different acquisitions. Biomed Eng Online (2015) 14:73. doi: 10.1186/s12938-015-0064-y
63. Yushkevich PA, Piven J, Hazlett HC, Smith RG, Ho S, Gee JC, et al. User-guided 3D active contour segmentation of anatomical structures: Significantly improved efficiency and reliability. NeuroImage (2006) 31:1116–28. doi: 10.1016/j.neuroimage.2006.01.015
64. 3D slicer image computing platform, in: 3D slicer. Available at: https://slicer.org/ (Accessed April 11, 2022).
65. Ayache N, Delingette H, Golland P, Mori K eds. Medical image computing and computer-assisted intervention – MICCAI 2012: 15th international conference, nice, France, October 1-5, 2012, proceedings. Berlin, Heidelberg: Springer (2012). doi: 10.1007/978-3-642-33454-2
66. Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby JS, et al. Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Sci Data (2017) 4:170117. doi: 10.1038/sdata.2017.117
67. Bakas S, Reyes M, Jakab A, Bauer S, Rempfler M, Crimi A, et al. Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the BRATS challenge. (2019). arXiv:181102629. doi: 10.48550/arXiv.1811.02629
68. Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby J, et al. Segmentation labels and radiomic features for the pre-operative scans of the TCGA-GBM collection. (2017). doi: 10.7937/K9/TCIA.2017.KLXWJJ1Q
69. Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby J, et al. Segmentation labels and radiomic features for the pre-operative scans of the TCGA-LGG collection. (2017). doi: 10.7937/K9/TCIA.2017.GJQ7R0EF
70. Clark K, Vendt B, Smith K, Freymann J, Kirby J, Koppel P, et al. The cancer imaging archive (TCIA): Maintaining and operating a public information repository. J Digit Imaging (2013) 26:1045–57. doi: 10.1007/s10278-013-9622-7
71. Johnson KA, Becker JA. Harvard Medical school – the whole brain atlas . Available at: http://www.med.harvard.edu/aanlib/home.html.
72. Puchalski RB, Shah N, Miller J, Dalley R, Nomura SR, Yoon J-G, et al. An anatomic transcriptional atlas of human glioblastoma. Science (2018) 360(6389):660–3. doi: 10.1126/science.aaf2666
73. Scarpace L, Mikkelsen T, Cha S, Rao S, Tekchandani S, Gutman T, et al. Radiology data from the cancer genome atlas glioblastoma multiforme [TCGA-GBM] collection. (2016). doi: 10.7937/K9/TCIA.2016.RNYFUYE9
74. Shusharina N, Bortfeld T, Cardenas C, De B, Diao K, Hernandez S, et al. Cross-modality brain structures image segmentation for the radiotherapy target definition and plan optimization. In: Shusharina N, Heinrich MP, Huang R, editors. Segmentation, classification, and registration of multi-modality medical imaging data. Cham: Springer International Publishing (2021). p. 3–15. Lecture Notes in Computer Science. doi: 10.1007/978-3-030-71827-5_1
75. National Cancer Institute Clinical Proteomic Tumor Analysis Consortium (CPTAC). Radiology data from the clinical proteomic tumor analysis consortium glioblastoma multiforme [CPTAC-GBM] collection. (2018). doi: 10.7937/k9/tcia.2018.3rje41q1
76. Shorten C, Khoshgoftaar TM. A survey on image data augmentation for deep learning. J Big Data (2019) 6:60. doi: 10.1186/s40537-019-0197-0
77. Nalepa J, Marcinkiewicz M, Kawulok M. Data augmentation for brain-tumor segmentation: A review. Front Comput Neurosci (2019) 13:83. doi: 10.3389/fncom.2019.00083
78. Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial nets. proceedings of the 27th international conference on neural information processing systems - volume 2. NIPS’14. Cambridge, MA, USA: MIT Press (2014) p. 2672–80.
79. Ghassemi N, Shoeibi A, Rouhani M. Deep neural network with generative adversarial networks pre-training for brain tumor classification based on MR images. Biomed Signal Process Control (2020) 57:101678. doi: 10.1016/j.bspc.2019.101678
80. Pal KK, Sudeep KS. (2016). Preprocessing for image classification by convolutional neural networks, in: 2016 IEEE International Conference on Recent Trends in Electronics, Information Communication Technology (RTEICT), Bangalore, India. pp. 1778–81. doi: 10.1109/RTEICT.2016.7808140
81. Huang J, Li Y-F, Xie M. An empirical analysis of data preprocessing for machine learning-based software cost estimation. Inf Softw Technol (2015) 67:108–27. doi: 10.1016/j.infsof.2015.07.004
82. Zhang S, Zhang C, Yang Q. Data preparation for data mining. Appl Artif Intell (2003) 17:375–81. doi: 10.1080/713827180
83. Alves AFF, Miranda JR de A, Reis F, de Souza SAS, Alves LLR, Feitoza L deM, et al. Inflammatory lesions and brain tumors: is it possible to differentiate them based on texture features in magnetic resonance imaging? J Venom Anim Toxins incl Trop Dis (2020) 26:1–10. doi: 10.1590/1678-9199-JVATITD-2020-0011
84. Mohsen H, El-Dahshan E-SA, El-Horbaty E-SM, Salem A-BM. Classification using deep learning neural networks for brain tumors. Future Computing Inf J (2018) 3:68–71. doi: 10.1016/j.fcij.2017.12.001
85. Chen C, Ou X, Wang J, Guo W, Ma X. Radiomics-based machine learning in differentiation between glioblastoma and metastatic brain tumors. Front Oncol (2019) 9:806. doi: 10.3389/fonc.2019.00806
86. Zacharaki EI, Wang S, Chawla S, Soo Yoo D, Wolf R, Melhem ER, et al. Classification of brain tumor type and grade using MRI texture and shape in a machine learning scheme. Magnet Resonance Med (2009) 62:1609–18. doi: 10.1002/mrm.22147
87. Mitchell TM, Hutchinson R, Niculescu RS, Pereira F, Wang X, Just M, et al. Learning to decode cognitive states from brain images. Mach Learn (2004) 57:145–75. doi: 10.1023/B:MACH.0000035475.85309.1b
88. Cox DD, Savoy RL. Functional magnetic resonance imaging (fMRI) “brain reading”: detecting and classifying distributed patterns of fMRI activity in human visual cortex. NeuroImage (2003) 19:261–70. doi: 10.1016/S1053-8119(03)00049-1
89. Peng H, Long F, Ding C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell (2005) 27:1226–38. doi: 10.1109/TPAMI.2005.159
90. Dash M, Liu H. Feature selection for classification. Intelligent Data Anal (1997) 1:131–56. doi: 10.1016/S1088-467X(97)00008-5
91. Chen Z, Menzies T, Port D, Boehm B. Feature subset selection can improve software cost estimation accuracy. In: Proceedings of the 2005 workshop on predictor models in software engineering. PROMISE ‘05. New York, NY, USA: Association for Computing Machinery (2005). p. 1–6. doi: 10.1145/1083165.1083171
92. Das S. Filters, wrappers and a boosting-based hybrid for feature selection. In: Proceedings of the eighteenth international conference on machine learning. ICML ‘01. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc (2001). p. 74–81.
93. Citak-Er F, Firat Z, Kovanlikaya I, Ture U, Ozturk-Isik E. Machine-learning in grading of gliomas based on multi-parametric magnetic resonance imaging at 3T. Comput Biol Med (2018) 99:154–60. doi: 10.1016/j.compbiomed.2018.06.009
94. Li Z, Liu P, An T, Yang H, Zhang W, Wang J. Construction of a prognostic immune signature for lower grade glioma that can be recognized by MRI radiomics features to predict survival in LGG patients. Trans Oncol (2021) 14:101065. doi: 10.1016/j.tranon.2021.101065
95. Vaidhya K, Thirunavukkarasu S, Alex V, Krishnamurthi G. Multi-modal brain tumor segmentation using stacked denoising autoencoders. In: A C, Menze B, Maier O, Reyes M, Handels H, editors. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries (BrainLes 2015). Cham: Springer International Publishing (2015). p. 181–94. doi: 10.1007/978-3-319-30858-6_16
96. Agn M, Puonti O, Rosenschöld PM, Law I, Van Leemput K. Brain tumor segmentation using a generative model with an RBM prior on tumor shape. In: Crimi A, Menze B, Maier O, Reyes M, Handels H, editors. Cham: Springer International Publishing (2016). p. 168–80. doi: 10.1007/978-3-319-30858-6_15
97. Ari A, Hanbay D. Deep learning based brain tumor classification and detection system. Turkish J Electrical Eng Comput Sci (2018) 26:2275–86. doi: 10.3906/elk-1801-8
98. Zhou Y, Li Z, Zhu H, Chen C, Gao M, Xu K, et al. “Holistic brain tumor screening and classification based on DenseNet and recurrent neural network.”. In: Crimi A, Bakas S, Kuijf H, Keyvan F, Reyes M, van Walsum T, editors. Brainlesion: Glioma, multiple sclerosis, stroke and traumatic brain injuries. Cham: Springer International Publishing (2018). p. 208–17. doi: 10.1007/978-3-030-11723-8_21
99. Cheng J, Huang W, Cao S, Yang R, Yang W, Yun Z, et al. Enhanced performance of brain tumor classification via tumor region augmentation and partition. PloS One (2015) 10:e0140381. doi: 10.1371/journal.pone.0140381
100. Banerjee S, Mitra S, Masulli F, Rovetta S. Brain tumor detection and classification from multi-sequence MRI: Study using ConvNets. In: Crimi A, Bakas S, Kuijf H, Keyvan F, Reyes M, van Walsum T, editors. International MICCAI brainlesion workshop: Brainlesion: Glioma, multiple sclerosis, stroke and traumatic brain injuries (BrainLes 2018). Cham: Springer International Publishing (2018). p. 170–9. doi: 10.1007/978-3-030-11723-8_17
101. Xu Y, Jia Z, Ai Y, Zhang F, Lai M, Chang EI-C. Deep convolutional activation features for large scale brain tumor histopathology image classification and segmentation, in: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, Australia: IEEE (2015). pp. 947–51. doi: 10.1109/ICASSP.2015.7178109
102. Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L. ImageNet: A large-scale hierarchical image database, in: 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL: IEEE (2009). pp. 248–55. doi: 10.1109/CVPR.2009.5206848
103. Patcheil RA. Metastatic brain tumors. Neurol Clinics (1995) 13:915–25. doi: 10.1016/S0733-8619(18)30025-2
104. Kamson DO, Mittal S, Buth A, Muzik O, Kupsky WJ, Robinette NL, et al. Differentiation of glioblastomas from metastatic brain tumors by tryptophan uptake and kinetic analysis: a pet study with MRI comparison. Mol Imaging (2013) 12:327–37. doi: 10.2310/7290.2013.00048
105. Huberty CJ. Discriminant analysis. Rev Educ Res (1975) 45:543–98. doi: 10.3102/00346543045004543
106. Priya S, Liu Y, Ward C, Le NH, Soni N, Pillenahalli Maheshwarappa R, et al. Machine learning based differentiation of glioblastoma from brain metastasis using MRI derived radiomics. Sci Rep (2021) 11:10478. doi: 10.1038/s41598-021-90032-w
107. Zou H, Hastie T. Regularization and variable selection via the elastic net. J R Stat Soc (2005) 67:301–20. doi: 10.1111/j.1467-9868.2005.00503.x
108. Zwanenburg A, Vallières M, Abdalah MA, Aerts HJWL, Andrearczyk V, Apte A, et al. The image biomarker standardization initiative: Standardized quantitative radiomics for high-throughput image-based phenotyping. Radiology (2020) 295:328–38. doi: 10.1148/radiol.2020191145
109. Nguyen AV, Blears EE, Ross E, Lall RR, Ortega-Barnett J. Machine learning applications for the differentiation of primary central nervous system lymphoma from glioblastoma on imaging: a systematic review and meta-analysis. Neurosurg Focus (2018) 45(5):1–10. doi: 10.3171/2018.8.FOCUS18325
110. Liu Y-H, Maftah M, Das T, Bai L, Robson K, Auer D. Classificatioo of MR tumor images based on gabor wavelet analysis. J Med Biol Eng (2012) 32:22–8. doi: 10.5405/jmbe.813
111. Yamasaki T, Chen T, Hirai T, Murakami R. Classification of cerebral lymphomas and glioblastomas featuring luminance distribution analysis. Comput Math Methods Med (2013) 2013:e619658. doi: 10.1155/2013/619658
112. Yang Z, Feng P, Wen T, Wan M, Hong X. Differentiation of glioblastoma and lymphoma using feature extraction and support vector machine. CNS Neurol Disord - Drug Targets- CNS Neurol Disorders) (2017) 16:160–8. doi: 10.2174/1871527315666161018122909
113. Alcaide-Leon P, Dufort P, Geraldo AF, Alshafai L, Maralani PJ, Spears J, et al. Differentiation of enhancing glioma and primary central nervous system lymphoma by texture-based machine learning. Am J Neuroradiol (2017) 38:1145–50. doi: 10.3174/ajnr.A5173
114. Kang D, Park JE, Kim Y-H, Kim JH, Oh JY, Kim J, et al. Diffusion radiomics as a diagnostic model for atypical manifestation of primary central nervous system lymphoma: development and multicenter external validation. Neuro-Oncology (2018) 20:1251–61. doi: 10.1093/neuonc/noy021
115. Chen Y, Li Z, Wu G, Yu J, Wang Y, Lv X, et al. Primary central nervous system lymphoma and glioblastoma differentiation based on conventional magnetic resonance imaging by high-throughput SIFT features. Int J Neurosci (2018) 128:608–18. doi: 10.1080/00207454.2017.1408613
116. Suh HB, Choi YS, Bae S, Ahn SS, Chang JH, Kang S-G, et al. Primary central nervous system lymphoma and atypical glioblastoma: Differentiation using radiomics approach. Eur Radiol (2018) 28:3832–9. doi: 10.1007/s00330-018-5368-4
117. Priya S, Ward C, Locke T, Soni N, Maheshwarappa RP, Monga V, et al. Glioblastoma and primary central nervous system lymphoma: differentiation using MRI derived first-order texture analysis – a machine learning study. Neuroradiol J (2021) 34:320–8. doi: 10.1177/1971400921998979
118. McAvoy M, Prieto PC, Kaczmarzyk JR, Fernández IS, McNulty J, Smith T, et al. Classification of glioblastoma versus primary central nervous system lymphoma using convolutional neural networks. Sci Rep (2021) 11:15219. doi: 10.1038/s41598-021-94733-0
119. Tan M, Le Q. EfficientNet: Rethinking model scaling for convolutional neural networks, in: Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA PMLR (2019). pp. 6105–14.
120. Ellingson BM, Chung C, Pope WB, Boxerman JL, Kaufmann TJ. Pseudoprogression, radionecrosis, inflammation or true tumor progression? challenges associated with glioblastoma response assessment in an evolving therapeutic landscape. J Neurooncol (2017) 134:495–504. doi: 10.1007/s11060-017-2375-2
121. Jia W, Gao Q, Han A, Zhu H, Yu J. The potential mechanism, recognition and clinical significance of tumor pseudoprogression after immunotherapy. Cancer Biol Med (2019) 16:655. doi: 10.20892/j.issn.2095-3941.2019.0144
122. Le Fèvre C, Constans J-M, Chambrelant I, Antoni D, Bund C, Leroy-Freschini B, et al. Pseudoprogression versus true progression in glioblastoma patients: A multiapproach literature review. part 2 – radiological features and metric markers. Crit Rev Oncol Hematol (2021) 159:103230. doi: 10.1016/j.critrevonc.2021.103230
123. Okada H, Weller M, Huang R, Finocchiaro G, Gilbert MR, Wick W, et al. Immunotherapy response assessment in neuro-oncology: a report of the RANO working group. Lancet Oncol (2015) 16:e534–42. doi: 10.1016/S1470-2045(15)00088-1
124. Song J, Kadaba P, Kravitz A, Hormigo A, Friedman J, Belani P, et al. Multiparametric MRI for early identification of therapeutic response in recurrent glioblastoma treated with immune checkpoint inhibitors. Neuro-Oncology (2020) 22:1658–66. doi: 10.1093/neuonc/noaa066
125. Chawla S, Shehu V, Gupta PK, Nath K, Poptani H. Physiological imaging methods for evaluating response to immunotherapies in glioblastomas. Int J Mol Sci (2021) 22:3867. doi: 10.3390/ijms22083867
126. Booth TC, Larkin TJ, Yuan Y, Kettunen MI, Dawson SN, Scoffings D, et al. Analysis of heterogeneity in T2-weighted MR images can differentiate pseudoprogression from progression in glioblastoma. PloS One (2017) 12:e0176528. doi: 10.1371/journal.pone.0176528
127. Hu X, Wong KK, Young GS, Guo L, Wong ST. Support vector machine multiparametric MRI identification of pseudoprogression from tumor recurrence in patients with resected glioblastoma. J Magnet Resonance Imaging (2011) 33:296–305. doi: 10.1002/jmri.22432
128. Lee J, Wang N, Turk S, Mohammed S, Lobo R, Kim J, et al. Discriminating pseudoprogression and true progression in diffuse infiltrating glioma using multi-parametric MRI data through deep learning. Sci Rep (2020) 10:20331. doi: 10.1038/s41598-020-77389-0
129. Jang B-S, Park AJ, Jeon SH, Kim IH, Lim DH, Park S-H, et al. Machine learning model to predict pseudoprogression versus progression in glioblastoma using MRI: A multi-institutional study (KROG 18-07). Cancers (2020) 12:2706. doi: 10.3390/cancers12092706
130. Akbari H, Rathore S, Bakas S, Nasrallah MP, Shukla G, Mamourian E, et al. Histopathology-validated machine learning radiographic biomarker for noninvasive discrimination between true progression and pseudo-progression in glioblastoma. Cancer (2020) 126:2625–36. doi: 10.1002/cncr.32790
131. Ismail M, Hill V, Statsevych V, Huang R, Prasanna P, Correa R, et al. Shape features of the lesion habitat to differentiate brain tumor progression from pseudoprogression on routine multiparametric MRI: A multisite study. Am J Neuroradiol (2018) 39:2187–93. doi: 10.3174/ajnr.A5858
132. Narang S, Kim D, Aithala S, Heimberger AB, Ahmed S, Rao D, et al. Tumor image-derived texture features are associated with CD3 T-cell infiltration status in glioblastoma. Oncotarget (2017) 8:101244–54. doi: 10.18632/oncotarget.20643
133. Melguizo-Gavilanes I, Bruner JM, Guha-Thakurta N, Hess KR, Puduvalli VK. Characterization of pseudoprogression in patients with glioblastoma: is histology the gold standard? J Neurooncol (2015) 123:141–50. doi: 10.1007/s11060-015-1774-5
134. Nie D, Lu J, Zhang H, Adeli E, Wang J, Yu Z, et al. Multi-channel 3D deep feature learning for survival time prediction of brain tumor patients using multi-modal neuroimages. Sci Rep (2019) 9:1103. doi: 10.1038/s41598-018-37387-9
135. Macyszyn L, Akbari H, Pisapia JM, Da X, Attiah M, Pigrish V, et al. Imaging patterns predict patient survival and molecular subtype in glioblastoma via machine learning techniques. Neuro-Oncology (2015) 18:417–25. doi: 10.1093/neuonc/nov127
136. Sanghani P, Ang BT, King NKK, Ren H. Overall survival prediction in glioblastoma multiforme patients from volumetric, shape and texture features using machine learning. Surg Oncol (2018) 27:709–14. doi: 10.1016/j.suronc.2018.09.002
137. Choi YS, Ahn SS, Chang JH, Kang S-G, Kim EH, Kim SH, et al. Machine learning and radiomic phenotyping of lower grade gliomas: improving survival prediction. Eur Radiol (2020) 30:3834–42. doi: 10.1007/s00330-020-06737-5
138. Beig N, Patel J, Prasanna P, Hill V, Gupta A, Correa R, et al. Radiogenomic analysis of hypoxia pathway is predictive of overall survival in glioblastoma. Sci Rep (2018) 8:7. doi: 10.1038/s41598-017-18310-0
139. Grist JT, Withey S, Bennett C, Rose HEL, MacPherson L, Oates A, et al. Combining multi-site magnetic resonance imaging with machine learning predicts survival in pediatric brain tumors. Sci Rep (2021) 11:18897. doi: 10.1038/s41598-021-96189-8
140. Shiroishi MS, Castellazzi G, Boxerman JL, D’Amore F, Essig M, Nguyen TB, et al. Principles of T2*-weighted dynamic susceptibility contrast MRI technique in brain tumor imaging. J Magnet Resonance Imaging (2015) 41:296–313. doi: 10.1002/jmri.24648
141. Verma R, Correa R, Hill VB, Statsevych V, Bera K, Beig N, et al. Tumor habitat–derived radiomic features at pretreatment MRI that are prognostic for progression-free survival in glioblastoma are associated with key morphologic attributes at histopathologic examination: A feasibility study. Radiol: Artif Intell (2020) 2:e190168. doi: 10.1148/ryai.2020190168
142. Beig N, Singh S, Bera K, Prasanna P, Singh G, Chen J, et al. Sexually dimorphic radiogenomic models identify distinct imaging and biological pathways that are prognostic of overall survival in glioblastoma. Neuro-Oncology (2021) 23:251–63. doi: 10.1093/neuonc/noaa231
143. Beig N, Bera K, Prasanna P, Antunes J, Correa R, Singh S, et al. Radiogenomic-based survival risk stratification of tumor habitat on gd-T1w MRI is associated with biological processes in glioblastoma. Clin Cancer Res (2020) 26:1866–76. doi: 10.1158/1078-0432.CCR-19-2556
144. Zinn PO, Majadan B, Sathyan P, Singh SK, Majumder S, Jolesz FA, et al. Radiogenomic mapping of Edema/Cellular invasion MRI-phenotypes in glioblastoma multiforme. PloS One (2011) 6:e25451. doi: 10.1371/journal.pone.0025451
145. Ward PS, Cross JR, Lu C, Weigert O, Abel-Wahab O, Levine RL, et al. Identification of additional IDH mutations associated with oncometabolite r(–)-2-hydroxyglutarate production. Oncogene (2012) 31:2491–8. doi: 10.1038/onc.2011.416
146. Yogananda CGB, Shah BR, Nalawade SS, Murugesan GK, Yu FF, Pinho MC, et al. MRI-Based deep-learning method for determining glioma MGMT promoter methylation status. Am J Neuroradiol (2021) 42:845–52. doi: 10.3174/ajnr.A7029
147. Louis DN, Perry A, Reifenberger G, von Deimling A, Figarella-Branger D, Cavenee WK, et al. The 2016 world health organization classification of tumors of the central nervous system: a summary. Acta Neuropathol (2016) 131:803–20. doi: 10.1007/s00401-016-1545-1
148. Wierzbicki K, Ravi K, Franson A, Bruzek A, Cantor E, Harris M, et al. Targeting and therapeutic monitoring of H3K27M-mutant glioma. Curr Oncol Rep (2020) 22:19. doi: 10.1007/s11912-020-0877-0
149. Pan C, Liu J, Tang J, Chen X, Chen F, Wu Y, et al. A machine learning-based prediction model of H3K27M mutations in brainstem gliomas using conventional MRI and clinical features. Radiother Oncol (2019) 130:172–9. doi: 10.1016/j.radonc.2018.07.011
150. Su X, Chen N, Sun H, Liu Y, Yang X, Wang W, et al. Automated machine learning based on radiomics features predicts H3 K27M mutation in midline gliomas of the brain. Neuro-Oncology (2020) 22:393–401. doi: 10.1093/neuonc/noz184
151. Zhuo Z, Qu L, Zhang P, Duan Y, Cheng D, Xu X, et al. Prediction of H3K27M-mutant brainstem glioma by amide proton transfer–weighted imaging and its derived radiomics. Eur J Nucl Med Mol Imaging (2021) 48:4426–36. doi: 10.1007/s00259-021-05455-4
152. McKinley R, Jungo A, Wiest R, Reyes M. Pooling-free fully convolutional networks with dense skip connections for semantic segmentation, with application to brain tumor segmentation. In: Crimi A, Bakas S, Kuijf H, Menze B, Reyes M, editors. Brainlesion: Glioma, multiple sclerosis, stroke and traumatic brain injuries. Cham: Springer International Publishing (2017). p. 169–77. doi: 10.1007/978-3-319-75238-9_15
153. Mlynarski P, Delingette H, Criminisi A, Ayache N. Deep learning with mixed supervision for brain tumor segmentation. JMI (2019) 6:34002. doi: 10.1117/1.JMI.6.3.034002
154. Deng L. The MNIST database of handwritten digit images for machine learning research [Best of the web]. IEEE Signal Process Mag (2012) 29:141–2. doi: 10.1109/MSP.2012.2211477
155. Yu B, Zhou L, Wang L, Fripp J, Bourgeat P. 2018 IEEE 15th international symposium on biomedical imaging (ISBI 2018) Washington DC: IEEE (2018). p. 626–30. doi: 10.1109/ISBI.2018.8363653
156. Zabir I, Paul R SM, Sarker T, Fattah SA, Shahnaz C. Automatic brain tumor detection and segmentation from multi-modal MRI images based on region growing and level set evolution, in: 2015 IEEE International WIE Conference on Electrical and Computer Engineering (WIECON-ECE), Dhaka, Bangladesh IEEE (2015). pp. 503–6. doi: 10.1109/WIECON-ECE.2015.7443979
157. Alqazzaz S, Sun X, Yang X, Nokes L. Automated brain tumor segmentation on multi-modal MR image using SegNet. Comp Visual Media (2019) 5:209–19. doi: 10.1007/s41095-019-0139-y
158. Gutman DA, Cooper LAD, Hwang SN, Holder CA, Gao J, Aurora TD, et al. MR imaging predictors of molecular profile and survival: Multi-institutional study of the TCGA glioblastoma data set. Radiology (2013) 267:560–9. doi: 10.1148/radiol.13120118
159. Kickingereder P, Bonekamp D, Nowosielski M, Kratz A, Sill M, Burth S, et al. Radiogenomics of glioblastoma: Machine learning–based classification of molecular characteristics by using multiparametric and multiregional MR imaging features. Radiology (2016) 281:907–18. doi: 10.1148/radiol.2016161382
160. Gevaert O, Mitchell LA, Achrol AS, Xu J, Echegaray S, Steinberg GK, et al. Glioblastoma multiforme: Exploratory radiogenomic analysis by using quantitative image features. Radiology (2014) 273:168–74. doi: 10.1148/radiol.14131731
Keywords: artificial intelligence, machine learning, brain tumor, immunotherapy, radiomics, tumor classification, survival prediction, radiogenomics
Citation: Zhu M, Li S, Kuang Y, Hill VB, Heimberger AB, Zhai L and Zhai S (2022) Artificial intelligence in the radiomic analysis of glioblastomas: A review, taxonomy, and perspective. Front. Oncol. 12:924245. doi: 10.3389/fonc.2022.924245
Received: 25 April 2022; Accepted: 04 July 2022;
Published: 02 August 2022.
Edited by:
James Grist, University of Oxford, United KingdomReviewed by:
Wei Yang, Southern Medical University, ChinaMujib Ullah, Stanford University, United States
Copyright © 2022 Zhu, Li, Kuang, Hill, Heimberger, Zhai and Zhai. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Lijie Zhai, TGlqaWUuWmhhaUBub3J0aHdlc3Rlcm4uZWR1; Shengjie Zhai, U2hlbmdqaWUuWmhhaUB1bmx2LmVkdQ==