Skip to main content

METHODS article

Front. Cell Dev. Biol., 05 March 2021
Sec. Molecular and Cellular Pathology
This article is part of the Research Topic Omics Data Integration towards Mining of Phenotype Specific Biomarkers in Cancers and Diseases View all 67 articles

An Accurate Tool for Uncovering Cancer Subtypes by Fast Kernel Learning Method to Integrate Multiple Profile Data

\r\nHongyu Zhang&#x;Hongyu Zhang1†Limin Jiang&#x;Limin Jiang1†Jijun Tang,Jijun Tang1,2Yijie Ding*Yijie Ding3*
  • 1School of Computer Science and Technology, College of Intelligence and Computing, Tianjin University, Tianjin, China
  • 2Department of Computer Science and Engineering, University of South Carolina, Columbia, SC, United States
  • 3School of Electronic and Information Engineering, Suzhou University of Science and Technology, Suzhou, China

In recent years, cancer has become a severe threat to human health. If we can accurately identify the subtypes of cancer, it will be of great significance to the research of anti-cancer drugs, the development of personalized treatment methods, and finally conquer cancer. In this paper, we obtain three feature representation datasets (gene expression profile, isoform expression and DNA methylation data) on lung cancer and renal cancer from the Broad GDAC, which collects the standardized data extracted from The Cancer Genome Atlas (TCGA). Since the feature dimension is too large, Principal Component Analysis (PCA) is used to reduce the feature vector, thus eliminating the redundant features and speeding up the operation speed of the classification model. By multiple kernel learning (MKL), we use Kernel target alignment (KTA), fast kernel learning (FKL), Hilbert-Schmidt Independence Criterion (HSIC), Mean to calculate the weight of kernel fusion. Finally, we put the combined kernel function into the support vector machine (SVM) and get excellent results. Among them, in the classification of renal cell carcinoma subtypes, the maximum accuracy can reach 0.978 by using the method of MKL (HSIC calculation weight), while in the classification of lung cancer subtypes, the accuracy can even reach 0.990 with the same method (FKL calculation weight).

Introduction

Cancer is one of the most severe diseases endangering human life and health in the world. Among them, lung cancer and kidney cancer, which are the top ten killers of cancer, are the leading causes of cancer death. Lung cancer includes small cell lung cancer (SCLC) and non-small cell lung cancer (NSCLC). Lung adenocarcinoma (LUAD) and Lung squamous cell carcinoma (LUSC), two subtypes of NSCLC, accounting for about 85% of lung cancer (Herbst et al., 2018). Among the common types of renal cell carcinoma (RCC), Kidney renal clear cell carcinoma (KIRC) (75–80%), Kidney renal papillary cell carcinoma (KIRP) (10–15%), and Kidney Chromophobe (KICH) (5%) account for the vast majority. Correct diagnosis of cancer subtypes is helpful to find potential therapeutic targets and new drug development, so that reduce the mortality of cancer. At present, it is challenging to classify subtypes by traditional pathological analysis. The mature sequencing technology makes it possible for people to obtain a large number of gene expression profiles. According to the above gene expression data, early diagnosis of cancer can be made based on gene expression profiles even if some tissues of organisms have not changed significantly. Moreover, it also provides an excellent help for the classification of cancer subtypes.

In the classification of NSCLC subtypes, some studies have shown that the characteristics of mRNA expression or gene histology contribute to the conventional histopathological classification (Jun, 2010; Girard et al., 2016). In addition, we can also consider the relationship between genes, not just individual genes (Su et al., 2019). However, in the subtype classification of renal cell carcinoma, MiRNA signature obtained using quantitative reverse transcription-polymerase chain reaction (QRT-PCR) analysis has been proved to be effective in the classification of RCC subtypes (Youssef et al., 2011). Furthermore, using ensemble classification methods can get better results than a single machine learning algorithm (Park et al., 2018). All methods only focus on one feature of cancer or the correlation of one feature to classify cancer subtypes, while ignoring the influence of other characteristics.

At present, computational methods (Zeng et al., 2018; Qi et al., 2019) have been widely applied to biological problems. It is mainly divided into two directions: one is the traditional machine learning method, which includes mostly Logistic Regression (LR), K-Nearest Neighbor (KNN), RF (Random Forest), SVM, etc.; and the other is multiple kernel learning (Ding et al., 2020b; Guo et al., 2020; Liu et al., 2020; Zou et al., 2020). It maps different feature components of heterogeneous data with different kernel functions so that the data can be better expressed in the new feature space, and the classification performance is significantly improved. Multiple kernel learning has been widely used in the field of computational biology, for example, protein function identification (Ding et al., 2020a), drug-side effect association (Ding et al., 2018; Yijie et al., 2019), drug-target interactions (Ding et al., 2019), etc.

In this paper, we use the classical machine learning algorithm SVM to classify cancer subtypes based on gene expression, isoform expression and methylated expression. Because gene expression profiles are generally obtained by gene chip sequencing in biological systems. However, there are a huge number of genes in the cells of organisms, so the microarray data we get will also present the characteristics of small samples, high latitude, high noise, uneven distribution and so on. Therefore, it is necessary to use the PCA method to extract effective data from massive cancer features. It is worth mentioning that PCA is used to reduce the dimension of features data used in this experiment, and the dimension of features can differ by thousands of times. In addition, we use cross-validation method to make the algorithm more robust and get more accurate and reliable results. Finally, we construct the kernel function by features, and then apply multiple kernel learning (weight is calculated by HSIC, FKL, KTA, mean methods) to combine the kernel functions of multiple features into SVM, to get more excellent results. The flowchart is shown in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. The flowchart of our proposed method.

In the classification of lung cancer and kidney cancer subtypes, we have achieved the excellent results (the accuracy of kidney cancer: 0.978, the accuracy of lung cancer: 0.990). Identifying more accurate cancer subtypes will not only help to provide more appropriate individualized treatment suggestions for diseases in real life, but also promote the discovery of potential therapeutic targets and the development of new drugs, so as to improve the survival rate of patients.

In the section of “Materials and Methods,” we mainly introduced the data source, subtype types and data set size of lung cancer and renal cancer. Next, the dimension reduction method PCA and machine learning algorithm SVM used in the experiment are described in detail. Then, it is introduced the feature kernel construction method applied in SVM and the kernel fusion method applied in multi-kernel learning. The next section result shows the method of establishing model and evaluation criteria used in the experiment. After that, we draw the TSNE visualization graph of the reduced features and show the experimental results of different features in random forest, different classifiers, single-kernel SVM and multi-kernel SVM. In the last section of the paper, the conclusion summarizes the specific process of the experiment and the prospect of the paper.

Materials and Methods

Datasets

We obtain the cancer subtype data from the experiment in Broad GDAC (Xiao and Yang, 2016), which collects and analyzes the standardized data extracted from The Cancer Genome Atlas (TCGA) (Tomczak et al., 2015). We extract two cancer datasets, including renal cancer (RCC) and non-small cell lung cancer (NSCLC). RCC has three subtypes: Kidney Chromophobe (KICH), Kidney renal clear cell carcinoma (KIRC), Kidney renal papillary cell carcinoma (KIRP), where 113 samples in KICH, 537 in KIRC and 323 in KIRP. NSCLC has two subtypes: Lung adenocarcinoma (LUAD) and Lung squamous cell carcinoma (LUSC), where 585 samples in LUAD and 504 in LUSC. In addition, we remove the redundant data from the features (retain the cancer samples with sample number of 01–09), and get the sample cases with three dimensions information (gene expression, isoform expression and methylated expression) at the same time. We added tag information to each case. Finally, the lung cancer data set consists of two background subtypes, LUSC and LUAD, with a total sample number of 824; the renal cancer data set consists of KICH, KIRC and KIRP, with a total number of 658 samples.

Feature Dimension Reduction

We use three different feature sets, such as gene expression, isoform expression and methylation expression. However, the feature dimensions are too large that the smallest dimension has reached more than 20,000, so it is necessary to reduce the dimension of features. We mainly use Principal Component Analysis (PCA) (Jolliffe, 2002) to reduce characteristic dimensions of gene expression, isoform expression and methylated expression. We find that the dimension of gene and meth feature for kidney and lung is the best at about 40. Compared with isoform feature, the dimension of kidney feature is reduced to 20, and the dimension of lung feature is reduced to 30. The dimension of feature reduction can be shown in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2. Feature dimension reduction.

Support Vector Machine

Support Vector Machine (SVM) was proposed by Cortes and Vapnik (1995), by calculating the maximum margin hyperplane, be mainly applied to classification and regression problems. SVM is one of the kernel learning methods. It can be used to solve non-linear problems by mapping the low-dimensional feature to high-dimensional space.

In our experiment, we mainly use Pearson Correlation Coefficient, Cosine Similarity and Tanimoto Similarity to make kernel function, and build multi-kernel learning method to calculate the weights of kernel function through HSIC, KTA, FKL, and Mean models, so as to classify the subtype data of renal cancer and lung cancer.

Kernel Construction

Tanimoto Similarity

Tanimoto similarity (Rogers and Tanimoto, 1960) is mainly used to calculate the similarity between individuals measured by symbols or boolean values. The larger the coefficient value is, the higher the sample similarity is. Now it has been widely used in biological data analysis. If the value is a binary vector, then the Tanimoto similarity coefficient is equal to the Jaccard distance. The tanimoto similarity can be computed as Eq. 1:

E j ( A , B ) = A B | A | 2 + | B | 2 - A B A B = i A i B i | A | 2 = i A i 2 (1)

where A and B are two vectors, AB representing vector product, |A|2 and |B|2 are norms of the vector.

Pearson Correlation Coefficient

Pearson correlation coefficient (Williams, 1996) reflects the degree of correlation between the two variables, with the value between [−1, 1]. As the correlation coefficient approaches the value of -1 or 1, the linear relationship increases. If the coefficient is equal to 0, there is no linear relationship. Pearson correlation coefficient is expressed in the mathematical formula (2) as the covariance of two variables divided by the standard deviation of two variables as follows:

ρ a , b = c o v ( a , b ) σ a σ b (2)

where cov (a,b) is the covariance of variables a and b, σaσb represents the standard deviation of variables a and b, respectively.

Cosine Similarity

Cosine similarity is used to measure the difference between two individuals. Compared with the first two kernel function making methods, cosine similarity pays more attention to the difference between vector positions in direction than in distance or length. In essence, it is to calculate the cosine value of the angle between two vectors, which can be expressed as Eq. 3:

s ( M , N ) = cos θ = m n | m | | n | (3)

Multiple Kernel Learning

Compared with a single kernel function, multi-kernel learning (Gönen and Alpaydın, 2011) must be more flexible and developable. The mapping space of multiple kernel function is composed of the feature space for each single kernel function. Obviously, the combination space can map the different feature components of heterogeneous data through a suitable kernel fusion model, which has more accurate and reasonable expression, so as to improve the classification accuracy. We mainly use the multi-kernel linear combination method, which is essentially a linear combination of all kernel functions. GivenKi is the i-th kernel matrix,βi is the i-th weight of the matrix, the multi-kernel fusion function can be described as Eq. 4:

K = i β i K i (4)
i β i = 1 β i 0

Hilbert-Schmidt Independence Criterion

Hilbert-Schmidt Independence Criterion (HSIC) is to measure the distribution difference between two variables (Gretton et al., 2005; Wang et al., 2016), which is similar to covariance. The construction method depends on the covariance operator in Hilbert space as Eq. 5:

H S I C ( Z ( a ) Z ( b ) ) = ( n - 1 ) - 2 t r ( K a H K b H ) (5)

where Z(a) and Z(b) are two different data sets, n is the number of samples, tr(x) is the trace of matrix X,Ka,Kb,H∈ℜm×m,Ka,Kb is gram matrix of data set, Hij=δij-1m is a matrix with a mean value of 0.

Fast Kernel Learning

Referring to the description of fast kernel learning (FKL) (Shen et al., 2019a), in multi-kernel learning, we think that the target similarity matrix K should be close to the label similarity matrix Y, whereY=yyT. To prevent overfitting, we usually add a regularization term|α|2. Therefore, the solution of multi-kernel fusion weights can be sorted into a quadratic programming problem as Eq. 6:

min α , k | K - Y | F 2 + λ | α | 2 (6)
s . t .    w = 1 J α w = 1

Where F is Frobenius norm, λ is an equilibrium coefficient, and J is the number of kernel functions.

The formula can be further derived. Since the Frobenius norm of a matrix is equal to the trace of the product of the matrix and its transposed matrix, ||A||F2=tr(AAT). The formula can be simplified as Eq. 7:

min α α T ( A + λ I ) α - 2 b T α
s . t . w J α w = 1 (7)
α w 0 , w = 1 , , J

Here I is the identity matrix of the same size as A.

Kernel Target Alignment

Kernel target alignment (KTA) (Shen et al., 2019b) is mainly used to calculate the weight of kernel in multi-core learning. If is used to represent the score of KTA, the greater the value ofwα, the greater the correlation between the two kernels, which makes a higher contribution to the composite kernel withFtrain aligned kernel matrix, and vice versa. The alignment between Kα andKideal is called kernel target alignment (KTA). The ideal kernel matrix (Kideal) is calculated as Eq. 8:

K ideal = F train F train T (8)

Fraction formula (9) for calculating KTA:

w α = K α , K ideal F | K α | F | K ideal | F (9)

Where 〈X,Y〉F represents Frobenius inner product Trace (⋅).|X|F stands for Frobenius norm.

Results

Cross Validation

Cross validation can effectively avoid overfitting and improve the generalization ability of the model. Main idea is to divide the dataset into N subsets, randomly select N-1 subsets as the training set, the rest as the prediction set, to get the performance evaluation index of the classifier. This process continues until all subsets are predicted and only once. The final model evaluation results are obtained by combining N evaluation results. In our experiment, we mainly use the twofold cross validation to train and evaluate our model.

Evaluation Metrics

We evaluate the classifier based on sensitivity (SN), specificity (SP), accuracy (ACC), Mathew’s correlation coefficient (MCC) as Eqs 10a–d.

S N = T P T P + T N (10a)
S P = T N T N + F P (10b)
A C C = T P + T N T P + T N + F P + F N (10c)
M C C = T P × T N - F P × F N ( T P + F P ) × ( T P + F N ) × ( T N + F P ) × ( T N + F N ) (10d)

where TP, FP denote true positive, false positive; TN, FN denote true negative, false negative.

In addition, we draw the receiver operating characteristic (ROC) curves to better describe the data. We also get the area under curve (AUC) values by calculating the area of the ROC curve.

Analysis of Feature Dimension Reduction

Three feature sets, such as gene expression profile (gene), isoform expression (isoform), DNA methylation data (meth), construct kernel functions by cosine similarity (cosine), pearson correlation-based similarity (pearson), tanimoto similarity coefficient (tanimoto), and then cross-validation in SVM. It is necessary to reduce the dimension of the feature on the premise of slightly losing accuracy.

Tsne Feature Visualization

Tsne is a non-linear dimensionality reduction method, which can map the high-dimensional feature data to the low-dimensional space, so that it can be visualized in the graph. Feature visualizations on RCC and NSCLC are shown in Figures 3, 4.

FIGURE 3
www.frontiersin.org

Figure 3. Feature visualizations on RCC. (A): feature set of gene expression profile; (B): feature set of isoform expression; (C): feature set of DNA methylation.

FIGURE 4
www.frontiersin.org

Figure 4. Feature visualizations on NSCLC. (A): feature set of gene expression profile; (B): feature set of isoform expression; (C): feature set of DNA methylation.

Performance of Various Classifiers

Three dimensionally reduced features are put into random forest to get the results of Table 1. The best result of the kidney is gene feature with ACC of 0.951 and MCC of 0.917, and the best result of lung is meth feature with ACC of 0.973 and MCC of 0.946. All reduction features are feed into four classifiers, such as SVM, RF, KNN, LR, as shown in Table 2. Moreover, we also plot ROC curve to better describe the data, as shown in Figures 5A,B.

TABLE 1
www.frontiersin.org

Table 1. Results of different characteristics in random forest after dimensionality reduction.

TABLE 2
www.frontiersin.org

Table 2. Results of different classifiers after feature dimension reduction.

FIGURE 5
www.frontiersin.org

Figure 5. ROC curve of different classifiers and difference methods of kernel fusion on RCC and NSCLC. (A) The ROCs of different classifiers on Kidney cancer dataset; (B) The ROCs of different classifiers on Lung cancer dataset; (C) The ROCs of different MKL algorithms on Kidney cancer dataset; (D) The ROCs of different MKL algorithms on Lung cancer dataset.

Evaluation of Kernel Construction

After dimensionality reduction, three feature sets are constructed by different kernel construction methods as cosine, pearson and tanimoto, and put into SVM classifier for cross- validation. Results are shown in Table 3. It is not difficult to find that the dimensionality reduction features can be used to classify cancer subtypes.

TABLE 3
www.frontiersin.org

Table 3. Results of different kernel functions with different features in SVM.

Evaluation of Multiple Kernel Fusion

The characteristic gene + tanimoto, isoform + pearson, meth + cosine kernels of Kidney with the best results are fused by HSIC, KTA, FKL, and Mean weighted methods, and it is found that HSIC fusion method has better effect. The characteristic gene + cosine, isoform + cosine and meth + tanimoto of Lung with the best results are fused by HSIC, KTA, FKL, and Mean weighted methods, and it is found that FKL fusion method is the best. All results are shown in Table 4. Also, we plot ROC curves for four different methods of kernel fusion shown in Figures 5C,D.

TABLE 4
www.frontiersin.org

Table 4. Results of different kernel fusion methods in SVM.

Conclusion

In this paper, we obtained the data of two cancer subtypes (lung cancer and renal cancer) from Broad GDAC Firehouse, which collections and analyses the standardized data extracted from TCGA. We use Principal Component Analysis (PCA) method to reduce the dimension of features. The features are constructed into kernel functions by using cosine, pearson, tanimoto and other similarity measurement methods. Then the multiple kernel learning (MKL) method (KTA, FKL, HSIC, mean to calculate the weight of kernel fusion) is used to combine multiple kernel functions into a combined kernel function. Finally, the calculated kernel function is put into SVM to predict cancer subtypes. In addition, compared our model with some commonly used machine learning algorithms, such as random forest, linear regression, LR, and so on, our model has achieved good results. Our method also has some limitations. For example, we also calculated the feature data before dimension reduction by the above method. The results show that the characteristic dimension of the data after dimension reduction is reduced by several thousand times compared with that before dimension reduction. Still, the accuracy is also reduced by 3–4 percentage points. How to reduce the feature dimension while minimizing the gap with the classification result before dimension reduction is a problem we will consider in the future.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.

Author Contributions

YD, LJ, and HZ conceived and designed the experiments. HZ and LJ performed the experiments and analyzed the data. HZ and YD wrote the manuscript. YD and JT supervised the experiments and reviewed the manuscript. All authors have participated in study discussion and manuscript preparation and read and approved the final manuscript.

Funding

This work was supported by a grant from the National Natural Science Foundation of China (NSFC 61772362, 61902271, and 61972280) and National Key R&D Program of China (2018YFC0910405 and 2017YFC0908400).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

Cortes, C., and Vapnik, V. (1995). Support vector network. Mach. Learn. 20, 273–297. doi: 10.1007/BF00994018

CrossRef Full Text | Google Scholar

Ding, Y., Tang, J., and Guo, F. (2018). Identification of drug-side effect association via multiple information integration with centered kernel alignment. Neurocomputing 325, 211–224. doi: 10.1016/j.neucom.2018.10.028

CrossRef Full Text | Google Scholar

Ding, Y., Tang, J., and Guo, F. (2019). Identification of drug–target interactions via fuzzy bipartite local model. Neural Comput. Appl. 32, 1–17. doi: 10.1007/s00521-019-04569-z

CrossRef Full Text | Google Scholar

Ding, Y., Tang, J., and Guo, F. (2020a). Human protein subcellular localization identification via fuzzy model on kernelized neighborhood representation. Appl. Soft Comput. 96:106596. doi: 10.1016/j.asoc.2020.106596

CrossRef Full Text | Google Scholar

Ding, Y., Tang, J., and Guo, F. (2020b). Identification of drug–target interactions via dual laplacian regularized least squares with multiple kernel fusion. Knowl. Based Syst. 204:106254. doi: 10.1016/j.knosys.2020.106254

CrossRef Full Text | Google Scholar

Girard, L., Rodriguez-Canales, J., Behrens, C., Thompson, D. M., and Gazdar, A. F. (2016). An expression signature as an aid to the histologic classification of non-small cell lung cancer. Clin. Cancer Res. 22, 4880–4889. doi: 10.1158/1078-0432.ccr-15-2900

PubMed Abstract | CrossRef Full Text | Google Scholar

Gönen, M., and Alpaydın, E. (2011). Multiple kernel learning algorithms. J. Mach. Learn. Res. 12, 2211–2268.

Google Scholar

Gretton, A., Bousquet, O., Smola, A., and Schlkopf, B. (2005). “Measuring statistical dependence with hilbert-schmidt norms,” in Algorithmic Learning Theory, 16th International Conference, ALT 2005, Singapore, October 8-11, 2005, Proceedings, eds S. Jain, H. U. Simon, and E. Tomita (Berlin: Springer).

Google Scholar

Guo, X., Zhou, W., Yu, Y., Ding, Y., Tang, J., Guo, F., et al. (2020). A novel triple matrix factorization method for detecting drug-side effect association based on kernel target alignment. Biomed Res. Int. 2020:4675395.

Google Scholar

Herbst, R. S., Morgensztern, D., and Boshoff, C. (2018). The biology and management of non-small cell lung cancer. Nature 553, 446–454. doi: 10.1038/nature25183

PubMed Abstract | CrossRef Full Text | Google Scholar

Jolliffe, I. T. (2002). Principal component analysis. J. Mark. Res. 87:513.

Google Scholar

Jun, H. (2010). Gene expression-based classification of non-small cell lung carcinomas and survival prediction. PLoS One 5:e10312. doi: 10.1371/journal.pone.0010312

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, H., Ren, G., Chen, H., Liu, Q., Yang, Y., and Zhao, Q. (2020). Predicting lncRNA–miRNA interactions based on logistic matrix factorization with neighborhood regularized. Knowl. Based Syst. 191:105261. doi: 10.1016/j.knosys.2019.105261

CrossRef Full Text | Google Scholar

Park, K. H., Ishag, M. I. M., Ryu, K. S., Li, M., and Ryu, K. H. (2018). “Efficient ensemble methods for classification on clear cell renal cell carcinoma clinical dataset,” in Proceedings of the Asian Conference on Intelligent Information & Database Systems (Cham: Springer).

Google Scholar

Qi, Z., Yingjuan, Y., Guofei, R., Erxia, G., and Chunlong, F. (2019). Integrating bipartite network projection and KATZ measure to identify novel CircRNA-disease associations. IEEE Trans. Nanobioscience 18:578–584. doi: 10.1109/tnb.2019.2922214

PubMed Abstract | CrossRef Full Text | Google Scholar

Rogers, D. J., and Tanimoto, T. T. (1960). A computer program for classifying plants. Science 132, 1115–1118. doi: 10.1126/science.132.3434.1115

PubMed Abstract | CrossRef Full Text | Google Scholar

Shen, C., Ding, Y., Tang, J., and Guo, F. (2019a). Multivariate information fusion with fast kernel learning to kernel ridge regression in predicting LncRNA-protein interactions. Front. Genet. 9:716. doi: 10.3389/fgene.2018.00716

PubMed Abstract | CrossRef Full Text | Google Scholar

Shen, C., Ding, Y., Tang, J., Jiang, L., and Guo, F. (2019b). LPI-KTASLP: Prediction of LncRNA-Protein Interaction by Semi-Supervised Link Learning With Multivariate Information. Piscataway, NJ: IEEE Access. doi: 10.1109/ACCESS.2019.2894225

CrossRef Full Text | Google Scholar

Su, R., Zhang, J., Liu, X., and Wei, L. (2019). Identification of expression signatures for non-small-cell lung carcinoma subtype classification. Bioinformatics 36, 339–346.

Google Scholar

Tomczak, K., Czerwinska, P., and Wiznerowicz, M. (2015). The Cancer Genome Atlas (TCGA): an immeasurable source of knowledge. Contemp. Oncol. (Pozn) 19, A68–A77. doi: 10.5114/wo.2014.47136

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, T., Li, W., and He, X. (2016). “Kernel learning with hilbert-schmidt independence criterion,” in Proceedings of the Chinese Conference on Pattern Recognition (Singapore: Springer).

Google Scholar

Williams, S. (1996). Pearson’s correlation coefficient. N.Z. Med. J. 109:38.

Google Scholar

Xiao, S., and Yang, M. (2016). Discovery of a novel target for cancer: PRR14. Cell Death Dis. 7:e2502. doi: 10.1038/cddis.2016.401

PubMed Abstract | CrossRef Full Text | Google Scholar

Yijie, D., Jijun, T., and Fei, G. (2019). Identification of drug-side effect association via semisupervised model and multiple kernel learning. IEEE J. Biomed. Health Inform. 23:2619–2632. doi: 10.1109/JBHI.2018.2883834

PubMed Abstract | CrossRef Full Text | Google Scholar

Youssef, Y. M., White, N. M. A., Grigull, J. R., Krizova, A., Samy, C., Mejia-Guerrero, S., et al. (2011). Accurate molecular classification of kidney cancer subtypes using microRNA signature. Eur. Urol. 59, 721–730.

Google Scholar

Zeng, X., Liu, L., Lü, L., and Zou, Q. (2018). Prediction of potential disease-associated microRNAs using structural perturbation method. Bioinformatics 34:2425–2432. doi: 10.1093/bioinformatics/bty112

PubMed Abstract | CrossRef Full Text | Google Scholar

Zou, Y., Wu, H., Guo, X., Peng, L., and Guo, F. (2020). MK-FSVM-SVDD: a multiple kernel-based fuzzy SVM model for predicting DNA-binding proteins via support vector data description. Curr. Bioinform. 15. doi: 10.2174/1574893615999200607173829 [Epub ahead of print].

CrossRef Full Text | Google Scholar

Keywords: cancer subtypes classification, SVM, multiple kernel learning, gene expression profile, isoform expression, DNA methylation

Citation: Zhang H, Jiang L, Tang J and Ding Y (2021) An Accurate Tool for Uncovering Cancer Subtypes by Fast Kernel Learning Method to Integrate Multiple Profile Data. Front. Cell Dev. Biol. 9:615747. doi: 10.3389/fcell.2021.615747

Received: 09 October 2020; Accepted: 16 February 2021;
Published: 05 March 2021.

Edited by:

Lei Deng, Central South University, China

Reviewed by:

Qi Zhao, University of Science and Technology Liaoning, China
Xiangxiang Zeng, Hunan University, China

Copyright © 2021 Zhang, Jiang, Tang and Ding. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yijie Ding, wuxi_dyj@163.com

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.