Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 14 May 2021
Sec. Brain Imaging Methods
This article is part of the Research Topic Artificial Intelligence based Computer-aided Diagnosis Applications for Brain Disorders from Medical Imaging Data View all 14 articles

Detection and Grading of Gliomas Using a Novel Two-Phase Machine Learning Method Based on MRI Images

\r\nTao ChenTao Chen1Feng XiaoFeng Xiao2Zunpeng YuZunpeng Yu3Mengxue YuanMengxue Yuan3Haibo Xu*Haibo Xu2*Long Lu,,*Long Lu3,4,5*
  • 1School of Information Technology, Shangqiu Normal University, Shangqiu, China
  • 2Department of Radiology, Zhongnan Hospital of Wuhan University, Wuhan, China
  • 3School of Information Management, Wuhan University, Wuhan, China
  • 4Division of Biomedical Informatics, Cincinnati Children’s Hospital Medical Center, Cincinnati, OH, United States
  • 5Department of Pediatrics, University of Cincinnati College of Medicine, Cincinnati, OH, United States

The early detection and grading of gliomas is important for treatment decision and assessment of prognosis. Over the last decade numerous automated computer analysis tools have been proposed, which can potentially lead to more reliable and reproducible brain tumor diagnostic procedures. In this paper, we used the gradient-based features extracted from structural magnetic resonance imaging (sMRI) images to depict the subtle changes within brains of patients with gliomas. Based on the gradient features, we proposed a novel two-phase classification framework for detection and grading of gliomas. In the first phase, the probability of each local feature being related to different types (e.g., diseased or healthy for detection, benign or malignant for grading) was calculated. Then the high-level feature representing the whole MRI image was generated by concatenating the membership probability of each local feature. In the second phase, the supervised classification algorithm was used to train a classifier based on the high-level features and patient labels of the training subjects. We applied this framework on the brain imaging data collected from Zhongnan Hospital of Wuhan University for glioma detection, and the public TCIA datasets including glioblastomas (WHO IV) and low-grade gliomas (WHO II and III) data for glioma grading. The experimental results showed that the gradient-based classification framework could be a promising tool for automatic diagnosis of brain tumors.

Introduction

Gliomas are a group of primary brain tumors that arise from glial cells of the central nervous system (CNS). Traditionally, gliomas are classified by the World Health Organization (WHO) into four grades (from I to IV) depending on their histopathological features (Louis et al., 2007). However, the newest WHO classification of CNS tumors, published in 2016, combined both histopathological and genotypic features in the classification of these tumors (Louis et al., 2016). In this classification, WHO grade II and III are grouped under the low-grade glioma (LGG) category since they share common IDH mutations. Most of these tumors may also develop into the WHO grade IV glioblastoma (GBM) that has very high malignant degrees and poorer prognosis (Ohgaki and Kleihues, 2005). Compared to GBM, LGG has more optimistic outcomes with longer survival times (Brat et al., 2015). Therapeutic approaches are also different for these two groups of gliomas. Treatment for GBM normally includes surgical resection followed by radiotherapy with or without chemotherapy, while treatment for LGG is usually surgical resection followed by close observation (Woodworth et al., 2006). Hence, correctly grading of gliomas is very important for correct treatment. Conventionally, the determination of glioma grade depends on several histopathological features including mitotic activity, cytological atypia, neoangiogenesis, and tumor necrosis (Hsieh et al., 2017b). However, these features are not always easy to be recognized, and physicians may have different views about them, thus some misdiagnosis can still happen due to glioma heterogeneity or subjective judgments by physicians. Meanwhile, the surgery needs to resect some normal brain tissues, which may lead to sequelae, dysfunction or even functional loss after surgery.

With the rapid development of medical imaging technology, magnetic resonance imaging (MRI) is commonly used as a non-invasive tool to detect and determine the characteristics of brain tumors in clinic because it can provide a wide range of physiologically authoritative contrasts to recognize diverse tissues and enhances assessment of heterogeneous patterns of tissue compositions inside diffuse gliomas (Leach et al., 2005). Besides the typical sequences such as T1-weighted imaging (T1WI) and T2-weighted imaging (T2WI), other MRI techniques including diffusion-weighted imaging (DWI), MR spectroscopy (MRS), and perfusion-weighted imaging (PWI) can also be applied to discriminate between GBM and LGG (Provenzale et al., 2006; Tsougos et al., 2012; Svolos et al., 2013). Although a myriad of imaging data of gliomas has been produced every day worldwide, the detection and grade identification mainly depend on visual examination by experts, which is both time consuming and prone to errors. In recent years, artificial intelligence has made a huge impact on many aspects in human life including medicine and healthcare domain. Machine learning algorithms, the core techniques in artificial intelligence, can provide assistance for more automatic and objective diagnosis of brain tumors. Many algorithms can potentially discover the underlying subtle change patterns in patient brains by analyzing imaging data, which is sometimes difficult for humans to identify with eyes.

In MRI image analysis, feature extraction is a type of dimensionality reduction method that represents interesting parts of an image as informative features, facilitating the subsequent classification steps. For glioma detection and grading, traditional methods extracted hand-crafted image features and then trained machine learning models. Hsieh et al. (2017a) evaluated the malignancy of gliomas (GBM = 34, LGG = 73) using combination of global histogram moment features and local textural features, achieving an accuracy of 88% and an AUC of 0.89. Skogen et al. (2016) discriminated between low grade gliomas (grade II = 27) and high grade gliomas (grade III = 34 and grade IV = 34) using MRI textural features with different anatomical scales, achieving an AUC of 0.910. In recent years, deep learning methods such as convolutional neural networks (CNN) have shown the state-of-the-art performance in medical image analysis (Shen et al., 2017), including applications for tumor diagnosis. Yang et al. (2018) combined the deep learning with transfer learning for glioma grading (GBM = 61, LGG = 52) on conventional MRI images. AlexNet and GoogLeNet were fine-tuned from models that pre-trained on the ImageNet dataset, achieving the best test accuracy of 94.5%. Although these studies showed good performance on glioma grading, their experiments depended on manual selection of slices and ROIs by experienced neuroradiologists, which is time-consuming and error-prone. Zhuge et al. (2020) investigated fully automated methods for grading gliomas (GBM = 210, LGG = 105) by using deep CNN models including tumor segmentation and grade classification, achieving test accuracy of 97.1%. However, the process of distinguishing tumor boundaries from healthy cells is still a challenging task in the clinical routine. Some studies applied multistream deep learning method to obtain better performance. Ge et al. (2018) proposed a multistream deep CNN architecture for glioma grading followed by multi-modality MRI data fusion, achieving test accuracy of 90.87%. Ali et al. (2019) used the generative adversarial networks (GANs) for data augmentation and employed a multistream convolutional autoencoder (CAE) to extract multi-modality MRI features for classification of low/high grade gliomas, achieving test accuracy of 92.04%. Although deep learning methods have shown excellent performance in biomedical domains, their lack of interpretability still remains an issue, especially for clinical practice.

In this study, we propose a two-phase classification framework to distinguish patient from healthy controls or discriminate between GBM and LGG based on MRI images. Different from other studies, this framework analyzes all slices of a 3D MRI image without segmentation of brain tumors. In the first phase, we extract local features slice by slice using the Histogram of Oriented Gradients (HOG) algorithm. This algorithm helps to generate high-quality representations that depict image edge and texture. Zhu et al. (2015) proposed a multi-view learning method extracting both ROI features and HOG features from each MRI image for Alzheimer’s Disease diagnosis. Their method can help enhance disease status identification performance. Ghiassian et al. (2016) described a HOG-based learning algorithm that can produce effective classifiers for ADHD and autism. They applied the algorithm on two large public datasets and achieved good performance on both datasets. Since the histopathological characteristics of the two different grades of glioma influence the pixel intensity and spatial distribution within MRI images, we think that the HOG method may also help in the differentiation between GBM and LGG. We hypothesize that each local feature may be related to certain type, e.g., normal tissue or tumor tissue, with some membership probability. Then we combine the membership probability of each local feature into one high-level feature vector that is fed into the classifier trained in the second phase. By using this two-phase classification framework, we can identify whether tumors occur in the brain or predict the tumor grade.

Materials and Methods

Data Acquisition

In this study, we achieved two tasks in glioma diagnosis including glioma detection and glioma grading. We collected structural MRI (sMRI) data from patients with glioma and healthy controls in Zhongnan Hospital of Wuhan University. Because there is no grading information about these sMRI data, we used them in this study only for glioma detection task. We also downloaded the sMRI data from the Cancer Imaging Archive (TCIA) public repository. The data from TCIA only include images of different glioma grades, thus we used them in the glioma grading task. For convenience, we named these two datasets DS-Detect and DS-Grade, respectively, according to their different classification task.

The DS-Detect dataset contains 99 subjects including 62 patients with glioma and 37 healthy controls. Imaging was performed on a SIEMENS MAGNETOM Trio Tim 3.0T MRI Scanner. Whole brain coverage was obtained with 23 contiguous 6 mm axial slices (TR = 7,000 ms, TE = 94 ms, TI = 2,210 ms, FA = 130, matrix size = 464 × 512). The DS-Grade dataset includes 134 subjects among which 76 are diagnosed as GBM (grade IV), and 58 as LGG (grade II and III). Both datasets include three sMRI modalities: T1-weighted, T2-weighted, and T2-FLAIR. We chose T2-FLAIR modality since T2-FLAIR images are of higher-contrast and the high signal of tissue indicates the possible tumor growth.

Image Preprocessing

As the first step of image preprocessing, the MRIcron tool was used to convert the original DICOM scans of an individual into a single NifTI image file. To solve the problem of non-standardized MRI intensity values among intra-patient and inter-patient acquisitions, we used the bias correction and Z-score normalization method, respectively. The scan bias correction algorithm implemented in SPM12 was used to minimizes MRI intensity inhomogeneity within a tissue region. To remove inter-patient intensity variability, we performed a Z-score normalization for each image, which normalize an image by simply subtracting the mean and dividing by the standard deviation of the whole brain, followed by clipping of the intensity value at [−4, 4] and a transformation to [0, 1]. To support comparison in a similar position at similar sizes, MRI image need to be spatial normalized. We used the spatial normalization procedure provided by SPM12 toolbox to register all MRI images to standard MNI space.

Feature Extraction

Dalal and Triggs (2005) first applied HOG to pedestrian detection in static images and achieved higher performance than other local feature descriptors such as SIFT and HAAR (Dalal and Triggs, 2005). Since then HOG has been widely used in computer vision applications such as person and object detection or recognition (Qiang et al., 2006; Li et al., 2008; Yuan et al., 2009; Khan et al., 2010; Overett and Petersson, 2011; Kobayashi, 2013; Simo-Serra et al., 2013). The fundamental concept of HOG is that local object appearance and shape within an image can be described by the distribution of intensity gradients or edge directions. Here, we gave a brief introduction of HOG calculation process.

First, as illustrated in Figure 1A, the whole MRI image was divided into a dense grid of uniformed spaced regions that is called cells in HOG’s terminology. The cell could be in any shape, but rectangular shape (R-HOG) and circular shape (C-HOG) are the most widely used ones. Due to better performance in the pedestrian detection experiments, we chose the R-HOG in this study. Secondly, the gradient of each pixel within the cell was calculated including magnitude and direction. Figure 1B shows the gradient calculation result of one 4 × 4 cell. Each arrow in the figure represents the corresponding pixel gradient – the arrow direction means gradient direction and the arrow length means gradient magnitude. Then a process called orientation binning was used to generate the cell histogram. The histogram is essentially a vector of N (e.g., 8) channels (bins) that are evenly spread over 0 to 360 degrees if the gradient is considered to be “signed.” Figure 1C shows a partition scheme including 8 channels. Each arrow represents the center direction of a channel. The channel 0 is illustrated as the shaded area, and so on. Thus, which channel is selected depends on the calculated pixel gradient direction. Furthermore, each pixel within the cell casts a weighted vote for its channel based on the gradient magnitude. Finally, a cell histogram was generated by counting the weighted number of pixels distributed in different direction channels as shown in Figure 1D. And the concatenation of all the cell histograms represents the feature descriptor of the MRI image. The histogram may be contrast-normalized to improve accuracy and reduce effect of changes of illumination and shadowing (Dalal and Triggs, 2005).

FIGURE 1
www.frontiersin.org

Figure 1. HOG feature extracted from the MRI image. (A) The MRI image divided into cells. (B) The gradient calculation result of one cell. (C) The partition scheme including 8 channels. (D) The cell histogram generated by counting the weighted number of pixels distributed in different direction channels.

Two-Phase Classification Framework

In traditional HOG application like pedestrian detection, the HOG feature vector of each cell is concatenated into one high-dimensional vector representing the 2D image. However, since the preprocessed MRI image is a 3D NifTI format, we need to extract the HOG feature from each slice then concatenate all the HOG features to form a representation of the whole 3D image. The dimension of the final HOG feature vector thus is quite high, which may cause the “curse of dimensionality” problem.

To address this, we proposed a two-phase classification framework that could transform the low-level gradient features into high-level semantic features. In the first classification phase, we calculated the HOG feature of every cell in an MRI image from the first slice to the last slice. Then instead of directly concatenating these local HOG features into one high-dimensional feature vector, we analyzed these local features independently and then integrated the analyzing results for further process. We have proposed this “local to whole” approach in a previous study that used SIFT features to diagnose the neurological diseases (Chen et al., 2014). Specifically, we first transformed each HOG feature into one real number, which indicates the probability of each cell relating to one specific type, and then these real numbers were concatenated to form a compact representation of the whole MRI image. This transformation from original HOG feature to cell type feature can reduce the dimensionality of feature space, thus alleviating the impact of overfitting problem. In the second classification phase, the new representation of brain images in the training set were used to train a classifier to detect gliomas from brains or distinguish between different grades. We take the glioma grading as example to illustrate the overall two-phase classification framework.

Some studies about brain tumors depend on experienced neuroradiologists to select the most representative slices from the image scan, and delineate the tumor contour manually as ROI for feature analysis (Skogen et al., 2016; Hsieh et al., 2017a). However, the tumor location can be anywhere in the brain, thus such ROI process may be labor-intensive and time-consuming, which is not appropriate for clinical practice. In this study, we designed an automatic pipeline to analyze the image features without manual or semi-manual ROI delineation. Firstly, the whole 3D sMRI image was sliced into a series of 2D images along the scan orientation. Then for every slice, the feature descriptor of each cell was extracted using the HOG algorithm. This feature extraction process is fully automatic and can cover the entire brain. For the sake of convenient illustration, only the slice with the largest axial cross-section of the tumor was used as an example in Figure 2. Suppose a 3D sMRI image includes m slices and each slice is divided into n cells, then a total of m × n cells are required to be analyzed for one brain.

FIGURE 2
www.frontiersin.org

Figure 2. The gradient-based two-phase classification framework.

After we applied the HOG algorithm to the sMRI image, each cell was represented by its local HOG feature. We analyzed each feature independently and supposed it to be related to different types. Because in the MRI preprocessing phase the MRI images were already aligned to the standard brain template, those HOG features that were extracted from the same location of different brains could be compared based on their gradient histogram, and then their respective feature type be identified as the Figure 2 illustrated.

Since the exact label of the whole brain rather than the type of each local brain region is known, we applied unsupervised clustering methods to discriminate the type of each HOG feature. In machine learning, a clustering method is used to determine a classification of n subjects into K discrete classes (K < < n) for a given dataset X = {x1, x2, x3, …, xn}, with each subject xi, i = 1, 2, 3, …,n characterized by m features, xi = (xi1, xi2, xi3, …, xim). The main objective of clustering is to discover and describe the underlying structure in the data. There are two kinds of clustering methods including “hard clustering” that exclusively assigns each subject to a single cluster, and “soft clustering” that is flexible and allows each subject to be assigned to more than one cluster. K-means, one of the most widely used hard clustering methods, partitions the subjects into k clusters given the data points and k number of centroids in an automated fashion. As a variation of the K-means algorithm, the fuzzy C-means algorithm is based on fuzzy logic principles and assigns each subject a possibility in each cluster center from 0 to 100 percent. In this study, we used the fuzzy C-means algorithm to calculate the possibility of each HOG feature being related to diseased/healthy status or GBM/LGG status. The reason we used fuzzy clustering method is that the boundary between these different types of image features is usually ambiguous, e.g., some cells may be located in both malignant and benign brain regions. So, if only using traditional hard-thresholding clustering method like K-means to classify the feature without considering the ambiguity, the classification result can not reflect the actual grouping complexity of the brain features thus leading to reduced classification performance. In Figure 2, the decimal numbers like 0.4, 0.6 are the probability of local HOG feature being classified into GBM-related cluster (or LGG-related cluster). For the two datasets, we have tried different K values, and found K = 2 is the best value. Furthermore, the clustering process also generated centers of the two clusters that can be used to predict to which cluster the local features of new unknown subject belongs based on the nearest centroid classification method.

The local feature clustering result only represents the status of individual brain regions, while the status combination of all brain regions is more significant for describing the underlying pattern of the whole brain. Thus for every subject, we concatenated the fuzzy labels of each feature from the first cell to the last cell across all the MRI slices as shown in Figure 2. Instead of directly concatenating original HOG feature of each cell, we used the clustering method to transform each HOG feature to another high-level feature that indicates the status of the corresponding brain region (e.g., 60% for malignant). This feature transformation can reduce data dimensionality and alleviate the problem of overfitting. As a new representation of each subject’s brain image, the compact probability vector was used as the input feature to a Support Vector Machine (SVM) discriminative classifier. In machine learning, SVM is a supervised learning algorithm that outputs an optimal hyperplane which separates the data samples into two classes. And it has been proved highly successful in solving a wide range of pattern recognition and computer vision problems such as text detection and image classification, also in the classification of MRI data (Magnin et al., 2009; Zacharaki et al., 2009; Ecker et al., 2010). If the data is not linearly separable in the original feature space, SVM can efficiently performs a non-linear classification using a so-called kernel function, implicitly mapping their inputs into high-dimensional feature spaces to achieve separability.

The above steps described the whole training process of the two-phase classification framework. This training process produced two classifiers, the nearest centroid for local brain regions and SVM for the whole brain image. Then we can apply these classifiers to unknown brain samples in the test dataset. Like the training process, the test brain sample was also divided into local regions and the HOG feature was extracted for each brain region. These local features were first identified to be related to different types using the nearest centroid classifier trained in the first phase. Then the classification results of all the individual features were combined and fed to the SVM classifier trained on the second phase to predict the final result of the test brain sample, e.g., whether the brain contains gliomas or what grade the glioma is.

Results

To evaluate the performance of our two-phase classification framework and solve the problem of the imbalanced datasets, the stratified 10-fold cross-validation (CV) method was applied to train the models. The stratified method can preserve the proportion of positive to negative samples in each fold to match the original distribution in the whole dataset. Furthermore, the variance of the model will decrease by performing several random runs, in each of which all samples are first randomly shuffled and then split into a pair of train and test sets. Since we extracted HOG features slice by slice and combined their clustered results into one feature vector representing the whole 3D MRI image, our random data shuffling for CV is subject-separated.

In this study, we evaluated the performance of our proposed two-phase classification model using the following measurements: accuracy (ACC), sensitivity (SEN), specificity (SPE), and area under curve (AUC). These measurements can be calculated from the classification confusion matrix. Here, the accuracy is defined as the ratio of correctly classified subjects over all subjects. The sensitivity is the ratio of correctly classified subjects with glioma over all subjects with glioma, and the specificity is the ratio of correctly classified subjects without glioma over all subjects without glioma. The AUC refers the area under the receiver operating characteristic (ROC) curve. The larger AUC value means better model performance. For our unbalanced dataset, the AUC metric is especially useful to assess the overall performance of the model. In addition, the cell size is a parameter that will affect the performance of the model. In the experiment, we assigned the cell size with value from 10 to 20 and calculated the above measurements, respectively. In this study, we accomplished two classification tasks including glioma detection and glioma grading using our proposed gradient-based two-phase classification framework. Figure 3 shows the performance of these two classification tasks.

FIGURE 3
www.frontiersin.org

Figure 3. Performance of glioma detection and glioma grading.

The measurements from Figure 3 were calculated after 10 random runs of stratified 10-fold cross-validation for each cell size. The figures showed that the performance for glioma detection outperformed that for glioma grading. The reason lies in that glioma detection is just to identify the occupying effect in the patient brain, while glioma grading intends to distinguish different morphological pattern between high grade and lower grade, thus the detection task seems easier than the grading task.

During the second phase in SVM training, we obtained a weight vector that indicates the direction along which the two classes of subjects differ most. This vector can be used to identify and localize the most discriminant HOG features that account for case-control separation. By sorting the weights in descending order or setting a threshold value, we identified the brain regions that most likely related to tumors. Figure 4 shows the thresholding results for a patient with glioma.

FIGURE 4
www.frontiersin.org

Figure 4. Discriminant features related to brain tumors.

The baseline method for HOG is that extracting the local HOG features and then directly concatenating them into one feature vector as representation of the whole image. In the experiment, we compared the performance between the baseline method and our proposed method, which transformed regional HOG features into high-level features, for each glioma diagnostic task. We also used the stratified 10-fold cross-validation strategy for the baseline method. Tables 1, 2 shows the performance of the two methods. The cell size in each table means the optimal parameter to obtain the best model performance.

TABLE 1
www.frontiersin.org

Table 1. Comparison of the glioma detection performance evaluated by 10-fold cross-validation between the baseline concatenating HOG method and our proposed transformed HOG method.

TABLE 2
www.frontiersin.org

Table 2. Comparison of the glioma grading performance evaluated by 10-fold cross-validation between the baseline concatenating HOG method and our proposed transformed HOG method.

For summarizing the performance of our proposed classification method, we showed the confusion matrix for glioma detection task and glioma grading task in Tables 3, 4, respectively.

TABLE 3
www.frontiersin.org

Table 3. Confusion matrix for glioma detection task.

TABLE 4
www.frontiersin.org

Table 4. Confusion matrix for glioma grading task.

Discussion

As an advanced diagnostic imaging technology, brain MRI can provide more objective and reliable evidence for tumor detection and grade evaluation with invasive procedure. In the present study, a two-phase classification framework based on HOG features was developed for supporting clinical diagnosis of brain tumors such as gliomas. We applied the framework to two different tasks including identifying patients with gliomas from healthy controls and differentiating between GBM (WHO grade IV) and LGG (WHO grade II and III). In clinic, the glioma detection task is the first step for the subsequent grade differentiation. The performance for glioma detection task achieved an accuracy of 86.3%, a sensitivity of 89.4%, a specificity of 80.5%, and an AUC of 0.921. The results may be good enough to provide diagnostic suggestions to physicians. By comparison, the glioma grading performance achieved an accuracy of 76.3%, a sensitivity of 83.7%, a specificity of 68.7%, and an AUC of 0.806. The reason may be that the differences between healthy brains and diseased brains are more significant than those between GBM and LGG. Furthermore, the heterogeneous composition of aggressive cellular tissues may also cause misdiagnoses. Although machine learning techniques have been widely used in prior tumor grading studies, it is striking that most of these previous studies used the state-of-the-art radiomics or deep learning methods (Skogen et al., 2016; Hsieh et al., 2017a; Yang et al., 2018; Zhuge et al., 2020). Some of these approaches obtained grade classification accuracy over 90%. However, these methods are more dependent on stable and reproducible segmentation of the ROI, and a large amount of high-dimensional feature extraction and evaluation. Although our two-phase classification framework did not achieve much high performance on tumor grading, we directly extracted gradient features from the MRI images without much computational complexity. And our method did not depend on tumor segmentation by physicians or algorithms. Thus, our automated diagnostic tool may be more appropriate for clinical usage.

The first contribution of our work is that we have applied the computer vision techniques, e.g., HOG descriptor, to the analysis of MRI medical images. HOG and other feature descriptors such as SIFT are commonly used in object detection task such as human face identification. Compared to SIFT that only captures some salient key feature points, HOG describes the gradient change for each pixel (voxel for 3D HOG), so HOG features are good at depicting small or subtle changes within brain. Since the gliomas can cause local occupying effect, some structural changes will occur in the brain, which can lead to abnormal intensity-based gradient changes on the MRI images of patients with gliomas compared to images of healthy brains. For different grades of gliomas, the gradient change pattern is also supposed to be different thus can be used as discriminating features to estimate the tumor grade.

The second contribution of our work is to propose the two-phase classification framework. In our framework, we calculated HOG feature on each local brain region that is called a “cell” in the algorithm. Instead of directly concatenating each HOG feature into one feature vector as what it is traditionally used in pedestrian detection, we analyzed each HOG feature independently using machine learning methods. Specifically, we applied the fuzzy clustering method on HOG features from the same position on an MRI image, which transformed each HOG feature into a membership probability related to diseased status. Then we can obtain a high-level semantic feature by concatenating the clustered result of each HOG feature representing the distribution pattern of diseased brain regions for the whole MRI image. On the one hand, this transformation between different feature space can indeed reduce the dimensionality of the final feature representation in order to avoid overfitting, which is especially necessary for 3D MRI images including many 2D slices and relatively small number of annotated medical images; on the other, it enables us to identify the tumor-related features that contribute most to the final classification result. We have adopted such region-independent analysis approach in our previous studies on diagnosis of Alzheimer’s disease, Parkinson’s disease, bipolar disorder and autism using MRI images (Chen et al., 2014, 2020). And the results showed that this approach analyzing local features independently could facilitate the identification of potential tumor regions in neurological diseases.

In addition to the brain tumor classification framework, we also developed a computer-aided diagnosis platform for brain tumor integrating different parts of the framework. Currently, this platform needs to manually transfer the original DICOM image data to the platform, which is not efficient and timely. Next, we plan to develop a network interface between the platform and the Picture Archiving and Communication Systems (PACS) of Zhongnan Hospital of Wuhan University. This interface can ensure an automatic imaging data transfer from PACS to the platform, which may improve the applicability of the platform in clinical settings.

In addition to the strengths discussed earlier, our study has several limitations. First, we used only T2-Flair sequences that can depict peritumoral edema with clear signals, while may be weak in demonstrating other tumor characteristics such as necrosis and angiogenesis module. In the future study, we plan to investigate the possible complementary power from other MRI sequences under routine preoperative protocol. Second, the DS-Grade dataset includes both grade II and III gliomas with three different histological cell types including astrocytoma, oligodendroglioma, oligoastrocytoma. Each of the glioma subtypes is considered to contain different MRI signature. And the heterogeneity within the DS-Grade dataset may account for the poorer model performance for differentiating GBM and LGG. So, it is necessary to collect sufficient data consisting of various glioma subtypes for model validation in the next study. Third, our study did not consider demographic information of the subjects (e.g., patient age), which may provide additional discriminatory value.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

The studies involving human participants were reviewed and approved by the Institutional Ethics Board of Zhongnan Hospital of Wuhan University. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.

Author Contributions

TC and LL designed the study. TC implemented the algorithm. MY preprocessed the imaging data. ZY, FX, and HX gave critical suggestions. TC, MY, and LL drafted the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This research was funded by National Natural Science Foundation of China (61772375, 61936013, and 71921002), the National Social Science Fund of China (18ZDA325), National Key R&D Program of China (2019YFC0120003), Natural Science Foundation of Hubei Province of China (2019CFA025), Independent Research Project of School of Information Management Wuhan University (413100032), and the Key R&D and Promotion Projects of Henan Province of China (212102210521).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

Ali, M. B., Gu, I. H., and Jakola, A. S. (2019). “Multi-stream convolutional autoencoder and 2D generative adversarial network for glioma classification,” in Computer Analysis of Images and Patterns. CAIP 2019. Lecture Notes in Computer Science, Vol. 11678, eds M. Vento and G. Percannella (Cham: Springer), 234–245. doi: 10.1007/978-3-030-29888-3_19

CrossRef Full Text | Google Scholar

Brat, D. J., Verhaak, R. G., Aldape, K. D., Yung, W. K., Salama, S. R., Cooper, L. A., et al. (2015). Comprehensive, integrative genomic analysis of diffuse lower-grade gliomas. N. Engl. J. Med. 372, 2481–2498. doi: 10.1056/NEJMoa1402121

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, T., Chen, Y., Yuan, M., Gerstein, M., Li, T., Liang, H., et al. (2020). The development of a practical artificial intelligence tool for diagnosing and evaluating autism spectrum disorder: multicenter study. JMIR Med. Inform. 8:e15767. doi: 10.2196/15767

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, Y., Storrs, J., Tan, L., Mazlack, L. J., Lee, J. H., and Lu, L. J. (2014). Detecting brain structural changes as biomarker from magnetic resonance images using a local feature based SVM approach. J. Neurosci. Methods 221, 22–31. doi: 10.1016/j.jneumeth.2013.09.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Dalal, N., and Triggs, B. (2005). “Histograms of oriented gradients for human detection,” in Porceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), Vol. 881, (San Diego, CA), 886–893.

Google Scholar

Ecker, C., Rocha-Rego, V., Johnston, P., Mourao-Miranda, J., Marquand, A., Daly, E. M., et al. (2010). Investigating the predictive value of whole-brain structural MR scans in autism: a pattern classification approach. Neuroimage 49, 44–56. doi: 10.1016/j.neuroimage.2009.08.024

PubMed Abstract | CrossRef Full Text | Google Scholar

Ge, C., Gu, I. Y. H., Jakola, A. S., and Yang, J. (2018). “Deep learning and multi-sensor fusion for Glioma classification using multistream 2D convolutional networks,” in Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), (Honolulu, HI), 5894–5897.

Google Scholar

Ghiassian, S., Greiner, R., Jin, P., and Brown, M. R. (2016). Using functional or structural magnetic resonance images and personal characteristic data to identify ADHD and Autism. PLoS One 11:e0166934. doi: 10.1371/journal.pone.0166934

PubMed Abstract | CrossRef Full Text | Google Scholar

Hsieh, K. L., Lo, C. M., and Hsiao, C. J. (2017a). Computer-aided grading of gliomas based on local and global MRI features. Comput. Methods Programs Biomed. 139, 31–38. doi: 10.1016/j.cmpb.2016.10.021

PubMed Abstract | CrossRef Full Text | Google Scholar

Hsieh, K. L., Tsai, R. J., Teng, Y. C., and Lo, C. M. (2017b). Effect of a computer-aided diagnosis system on radiologists’ performance in grading gliomas with MRI. PLoS One 12:e0171342. doi: 10.1371/journal.pone.0171342

PubMed Abstract | CrossRef Full Text | Google Scholar

Khan, S. M., Cheng, H., Matthies, D., and Sawhney, H. (2010). “3D model based vehicle classification in aerial imagery,” in Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, (San Francisco, CA), 1681–1687.

Google Scholar

Kobayashi, T. (2013). “BFO meets HOG: feature extraction based on histograms of oriented p.d.f. gradients for image classification,” in Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, (Portland, OR), 747–754.

Google Scholar

Leach, M. O., Brindle, K. M., Evelhoch, J. L., Griffiths, J. R., Horsman, M. R., Jackson, A., et al. (2005). The assessment of antiangiogenic and antivascular therapies in early-stage clinical trials using magnetic resonance imaging: issues and recommendations. Br. J. Cancer 92, 1599–1610. doi: 10.1038/sj.bjc.6602550

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, M., Zhang, Z., Huang, K., and Tan, T. (2008). “Estimating the number of people in crowded scenes by MID based foreground segmentation and head-shoulder detection,” in Proceedings of the 2008 19th International Conference on Pattern Recognition, (Tampa, FL), 1–4. doi: 10.1117/1.jei.27.4.043028

CrossRef Full Text | Google Scholar

Louis, D. N., Ohgaki, H., Wiestler, O. D., Cavenee, W. K., Burger, P. C., Jouvet, A., et al. (2007). The 2007 WHO classification of tumours of the central nervous system. Acta Neuropathol. 114, 97–109. doi: 10.1007/s00401-007-0243-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Louis, D. N., Perry, A., Reifenberger, G., von Deimling, A., Figarella-Branger, D., Cavenee, W. K., et al. (2016). The 2016 world health organization classification of tumors of the central nervous system: a summary. Acta Neuropathol. 131, 803–820. doi: 10.1007/s00401-016-1545-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Magnin, B., Mesrob, L., Kinkingnehun, S., Pelegrini-Issac, M., Colliot, O., Sarazin, M., et al. (2009). Support vector machine-based classification of Alzheimer’s disease from whole-brain anatomical MRI. Neuroradiology 51, 73–83. doi: 10.1007/s00234-008-0463-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Ohgaki, H., and Kleihues, P. (2005). Population-based studies on incidence, survival rates, and genetic alterations in astrocytic and oligodendroglial gliomas. J. Neuropathol. Exp. Neurol. 64, 479–489. doi: 10.1093/jnen/64.6.479

PubMed Abstract | CrossRef Full Text | Google Scholar

Overett, G., and Petersson, L. (2011). “Large scale sign detection using HOG feature variants,” in Proceedings of the 2011 IEEE Intelligent Vehicles Symposium (IV), (Baden-Baden), 326–331.

Google Scholar

Provenzale, J. M., Mukundan, S., and Barboriak, D. P. (2006). Diffusion-weighted and perfusion MR imaging for brain tumor characterization and assessment of treatment response. Radiology 239, 632–649. doi: 10.1148/radiol.2393042031

PubMed Abstract | CrossRef Full Text | Google Scholar

Qiang, Z., Mei-Chen, Y., Kwang-Ting, C., and Avidan, S. (2006). “Fast human detection using a cascade of histograms of oriented gradients,” in Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), (New York, NY), 1491–1498.

Google Scholar

Shen, D., Wu, G., and Suk, H.-I. (2017). Deep learning in medical image analysis. Ann. Rev. Biomed. Eng. 19, 221–248. doi: 10.1146/annurev-bioeng-071516-044442

PubMed Abstract | CrossRef Full Text | Google Scholar

Simo-Serra, E., Quattoni, A., Torras, C., and Moreno-Noguer, F. (2013). “A joint model for 2D and 3D pose estimation from a single image,” in Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, (Portland, OR), 3634–3641.

Google Scholar

Skogen, K., Schulz, A., Dormagen, J. B., Ganeshan, B., Helseth, E., and Server, A. (2016). Diagnostic performance of texture analysis on MRI in grading cerebral gliomas. Eur. J. Radiol. 85, 824–829. doi: 10.1016/j.ejrad.2016.01.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Svolos, P., Tsolaki, E., Kapsalaki, E., Theodorou, K., Fountas, K., Fezoulidis, I., et al. (2013). Investigating brain tumor differentiation with diffusion and perfusion metrics at 3T MRI using pattern recognition techniques. Magn. Reson Imaging 31, 1567–1577. doi: 10.1016/j.mri.2013.06.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Tsougos, I., Svolos, P., Kousi, E., Fountas, K., Theodorou, K., Fezoulidis, I., et al. (2012). Differentiation of glioblastoma multiforme from metastatic brain tumor using proton magnetic resonance spectroscopy, diffusion and perfusion metrics at 3 T. Cancer Imaging 12, 423–436. doi: 10.1102/1470-7330.2012.0038

PubMed Abstract | CrossRef Full Text | Google Scholar

Woodworth, G. F., McGirt, M. J., Samdani, A., Garonzik, I., Olivi, A., and Weingart, J. D. (2006). Frameless image-guided stereotactic brain biopsy procedure: diagnostic yield, surgical morbidity, and comparison with the frame-based technique. J. Neurosurg. 104, 233–237. doi: 10.3171/jns.2006.104.2.233

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, Y., Yan, L. F., Zhang, X., Han, Y., Nan, H. Y., Hu, Y. C., et al. (2018). Glioma grading on conventional MR images: a deep learning study with transfer learning. Front. Neurosci. 12:804. doi: 10.3389/fnins.2018.00804

PubMed Abstract | CrossRef Full Text | Google Scholar

Yuan, X., Li-Feng, L., Cui-Hua, L., and Yan-Yun, Q. (2009). “Unifying visual saliency with HOG feature learning for traffic sign detection,” in Proceedings of the 2009 IEEE Intelligent Vehicles Symposium, (Xi’an), 24–29.

Google Scholar

Zacharaki, E. I., Wang, S., Chawla, S., Soo Yoo, D., Wolf, R., Melhem, E. R., et al. (2009). Classification of brain tumor type and grade using MRI texture and shape in a machine learning scheme. Magn. Reson Med 62, 1609–1618. doi: 10.1002/mrm.22147

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhu, X., Suk, H.-I., Zhu, Y., Thung, K.-H., Wu, G., and Shen, D. (2015). “Multi-view classification for identification of alzheimer’s disease,” in Machine Learning in Medical Imaging, eds L. Zhou, L. Wang, Q. Wang, and Y. Shi (New York, NY: Springer International Publishing), 255–262.

Google Scholar

Zhuge, Y., Ning, H., Mathen, P., Cheng, J. Y., Krauze, A. V., Camphausen, K., et al. (2020). Automated glioma grading on conventional MRI images using deep convolutional neural networks. Med. Phys. 47, 3044–3053. doi: 10.1002/mp.14168

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: glioma, detection, grading, gradient, classification, MRI

Citation: Chen T, Xiao F, Yu Z, Yuan M, Xu H and Lu L (2021) Detection and Grading of Gliomas Using a Novel Two-Phase Machine Learning Method Based on MRI Images. Front. Neurosci. 15:650629. doi: 10.3389/fnins.2021.650629

Received: 07 January 2021; Accepted: 16 April 2021;
Published: 14 May 2021.

Edited by:

Ahmed Soliman, University of Louisville, United States

Reviewed by:

Yi Zhang, Zhejiang University, China
Irene Gu, Chalmers University of Technology, Sweden

Copyright © 2021 Chen, Xiao, Yu, Yuan, Xu and Lu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Haibo Xu, eHVoYWlibzExMjBAaG90bWFpbC5jb20=; Long Lu, YmlvaW5mb0BnbWFpbC5jb20=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.