- 1The First Hospital of Hunan University of Chinese Medicine, Changsha, Hunan, China
- 2Big Data Innovation and Entrepreneurship Education Center of Hunan Province, Changsha University, Changsha, China
- 3College of Information Science and Engineering, Hunan Women’s University, Changsha, Hunan, China
Essential proteins play important roles in the development and survival of organisms whose mutations are proven to be the drivers of common internal diseases having higher prevalence rates. Due to high costs of traditional biological experiments, an improved Transfer Neural Network (TNN) was designed to extract raw features from multiple biological information of proteins first, and then, based on the newly-constructed Transfer Neural Network, a novel computational model called TNNM was designed to infer essential proteins in this paper. Different from traditional Markov chain, since Transfer Neural Network adopted the gradient descent algorithm to automatically obtain the transition probability matrix, the prediction accuracy of TNNM was greatly improved. Moreover, additional antecedent memory coefficient and bias term were introduced in Transfer Neural Network, which further enhanced both the robustness and the non-linear expression ability of TNNM as well. Finally, in order to evaluate the identification performance of TNNM, intensive experiments have been executed based on two well-known public databases separately, and experimental results show that TNNM can achieve better performance than representative state-of-the-art prediction models in terms of both predictive accuracies and decline rate of accuracies. Therefore, TNNM may play an important role in key protein prediction in the future.
1 Introduction
Essential proteins refer to proteins that removal of them will lead to cell death or infertility (Acencio and Lemke, 2009). Identification of essential proteins can help us understand the basic requirements for maintaining life forms. However, considering that it is very costly and time-consuming to identify essential proteins by adopting traditional biological experimental techniques such as gene knock-out (Maliackal et al., 2005), RNA interference (Jeong et al., 2001) and conditional knockout (Hahn and Kern, 2005), more and more computational models have been proposed to infer essential proteins in recent years based on the topological characteristics of Protein-Protein Interaction (PPI) networks, in which, proteins are the vertices of PPI networks, while the interactions between proteins constitute the edges. Researches demonstrated that the topological position of a protein in the protein network has a strong correlation with whether the protein is critical (Jeong et al., 2001; Maslov and Sneppen, 2002; Pržulj et al., 2004). Based on the topological characteristics of proteins in PPI networks, a series of essential protein recognition methods have been proposed. For instance, based on the centrality of PPI networks (Stephenson and Zelen, 1989; Jeong et al., 2001), different computational models including CC (Closeness Centrality) (Stefan and Stadler, 2003), DC (Degree Centrality) (Hahn and Kern, 2005), BC (Between Centrality) (Wang et al., 2012), SC (Graph Centrality) (Maliackal et al., 2005) and NC (Neighbor Centrality) (Wang et al., 2012) have been proposed in succession. In addition, Li M et al. designed a recognition model named LAC (Li et al., 2011) to detect essential proteins (Li et al., 2015) based on the local average connectivity of protein nodes in the PPI network. Qi Yi et al. (Qi and Luo, 2016) introduced a prediction model based on the local interaction density (LID) of protein nodes in the PPI network to infer essential proteins. Chen B et al. (Chen and Wu, 2013) proposed an essential protein recognition method based on multiple topological features of the PPI network. In all these above methods, only topological characteristics of the PPI network were considered to identify essential proteins, however, since there is a large amount of noise data in PPI networks, then the predictive accuracy of these methods is not very satisfactory.
In order to break through the inherent limitations of existing PPI data, in the past few years, people proposed novel models by combining the topological characteristics of PPI networks with biological information of proteins. For example, M Li et al. and Xiwei Tang et al. put forward prediction models called Pec (Li et al., 2012) and WDC (Tang et al., 2014) by integrating PPI network and gene expression data of proteins respectively. W Peng et al. designed a prediction model (Peng et al., 2012) by integrating protein homology information with PPI networks, and a prediction model (Peng et al., 2015) through combining protein domain information with PPI network, simultaneously. X Zhang et al. (Zhang et al., 2013) introduced a recognition method called CoEWC by merging topological characteristics of the PPI network with the co expression characteristics of proteins. BH Zhao et al. designed a prediction model named POEM (Zhao et al., 2014) by combining gene expression data of proteins with the topological characteristics of PPI networks. J Luo et al. proposed a identification method based on local interaction density of PPI networks and biological characteristics of protein complexes (Jiawei et al., 2015). Seketoulie Keretsu et al. presented a protein complex recognition model (Li et al., 2015) based on clustering weighted edges and gene expression profile of proteins. M Li et al. designed two essential protein recognition methods by combining PPI networks with subcellular location information and complex centrality of proteins respectively (Keretsu and Sarmah, 2016; Li et al., 2017; Chen et al., 2020). J Luo et al. introduced a prediction method ECC (edge clustering coefficient) based on the complex co expression data of proteins and PPI networks (Luo and Wu, 2015). Bihai Zhao et al. proposed a model based on multiple biological networks (Zhao et al., 2020a) and a model based on diffusion distance network (Zhao et al., 2020b) to predict essential proteins respectively. S. Li et al. designed an iterative method called CVIM (Li et al., 2020) based on topological and functional characteristics of proteins to predict key proteins. Lei X et al. proposed a necessary protein prediction method AFSOEP (Lei et al., 2018) to infer protein complexes through AFSO (Artificial Fish Swarm Optimization). BH Zhao et al. designed an iterative method to identify potential essential proteins (Zhao et al., 2019) based on heterogeneous PPI networks. Dai W et al. proposed a method to discover key genes based on protein-protein interaction network embedding (Dai et al., 2020). Fengyu Zhang et al. predicted the key gene (Zhang et al., 2019) by fusing the dynamic PPI network. Chen Z et al. proposed an essential protein prediction model NPRI based on heterogeneous network, and established heterogeneous protein domain network (Chen et al., 2020) according to initial PPI network, protein domain network and gene expression data.
All these above methods show that the identification accuracy of models can be significantly improved by combining biological information of proteins with topological features of PPI networks. However, through analyzing results achieved by these existing methods, it is not difficult to find that the prediction accuracies of these algorithms decline fast with the increasing of predicted essential proteins. Hence, inspired by recognition models based on the Markov chain and the Transfer algorithm, we designed a new neural network called TNN in this manuscript, based on which, a novel model named TNNM was proposed to predict essential proteins. TNN can be divided into three parts, namely, probability transfer matrix, antecedent output and bias term. In addition, in order to evaluate the performance of TNNM, we compared it with existing representative models such as IC (Stephenson and Zelen, 1989), DC (Hahn and Kern, 2005), SC (Maliackal et al., 2005), NC (Wang et al., 2012), PeC (Li et al., 2012), ION (Peng et al., 2012), CoEWC (Zhang et al., 2013), POEM (Zhao et al., 2014), CVIM (Li et al., 2020), NPRI (Chen et al., 2020) and RWHN (Zhao et al., 2014) separately. Experimental results show that TNNM is far superior to these traditional models in terms of both predictive accuracy and decline rate of accuracy.
The rest of this paper is organized as follows: The experimental data and specific steps are organized in Section 2. In Section 3, the influence of parameters and comparison with other methods are shown. Section 4 describes the shortcomings of the model and future improvement goals. Finally, a summary is made in Part 5.
2 Method and materials
The flow chart of TNNM is shown in Figure 1. Through observing Figure 1, it is easy to see that TNNM consists of the following three major parts. Firstly, based on prior knowledge, topological features and biological features of each protein will be extracted from PPI networks, gene expression data, subcellular localization and ortholog data of proteins separately. And then, the Transfer Neural Network (TNN) will be designed. Finally, through adopting TNN, the prediction model TNNM will be constructed to infer essential proteins based on these newly extracted features.
2.1 Experimental data
In order to evaluate the prediction performance of TNNM, during experiments, we first downloaded known PPIs from different benchmark databases such as DIP 2010 (Xenarios et al., 2002) and Gavin (Gavin et al., 2006) respectively. After preprocessing, a dataset containing 5093 proteins and 24,743 known PPIs was finally obtained from the DIP2010 database, and a dataset containing 1855 proteins and 7669 known PPIs was obtained from the Gavin database. In addition, based on databases including MIPS (Mewes et al., 2006), SGD (Cherry et al., 1998), DEG (Zhang and Lin, 2009) and SGDP (StanfordMedicine, 2012), a benchmark dataset containing 1285 essential proteins was constructed, based on which, 1167 and 714 essential proteins were screened from the DIP2010 and Gavin databases respectively. Moreover, based on the dataset provided by Tu BP et al. (Tu et al., 2005), a dataset containing the gene expression data of 6776 proteins was obtained, which consists of the gene expression level data of proteins in the continuous metabolic cycle. Simultaneously, the homologous information of proteins was downloaded from the Inparanoid database (seventh edition), including paired comparison (Gabriel et al., 2010) between 100 whole genomes, and the number of times that proteins have homologous information in the reference organism. Finally, we downloaded the dataset containing subcellular localization information of proteins from the COMPART-MENTS database (Binder et al., 2014) (2014 version), and retained only 11 types of subcellular localization data closely related to essential proteins, such as cytoplasm, cytoskeleton, Golgi apparatus, cytoplasm, vacuoles, mitochondria, endosomes, plasma, nucleus, peroxisomes and extracellular enzymes, etc.
Based on above newly-downloaded datasets, firstly, we constructed an original
2.2 Topological feature extraction
Let the undirected graph
For any given protein
Based on above formula (1), we define that the degree of contact between
Here,
Here, |N(p)∩N(q)| represents the number of elements in N(p)∩N(q).
It is reasonable to consume that the smaller the total distance between a protein and all other proteins, the more important the protein will be. Hence, let l (p, q) denote the length of the shortest path from protein p to the protein q in G, if there is no path between p and q in G, then we define the length of the shortest path between p and q is a constant number N (>1). Therefore, we can calculate the importance of p related to the total distance as follows:
Moreover, it is also reasonable to assume that the more important a protein p is, the more proteins that have the shortest path through p. This indicator reflects the carrying capacity of p between other nodes in G. it is obvious that the larger the value, the greater the impact of p in the network, which also means that p will be more important. Hence, we can calculate the importance of p related to the carrying capacity as follows:
Here, kqq’ represents the number of shortest paths between q and q′ in G, and kqq′p) denotes the number of shortest paths between q and q′ in G, which pass through p.
2.3 Biological feature extraction
Let ge (p, t) represent the gene expression value of the protein p at the time point t,
Based on above formula (6), we can calculate the importance of p related to the Pearson correlation coefficient as follows:
It is reasonable to consume that essential proteins tend to be connected rather than independent. Therefore, we can believe that proteins closely related to essential proteins are more likely to be essential proteins. Thus, we can obtain another importance indicator of p as follows:
Where
Here, Sub(p) represents the set of subcellular locations of the protein p, |Sub(p)∩Sub(q)| denotes the number of elements in Sub(p)∩Sub(q), and |Sub(p)∪Sub(q)| is the number of elements in Sub(p)∪Sub(q).
Moreover, based on the reasonable assumption that the evolution of essential proteins is more conservative than that of non-essential proteins, and considering that the homologous information of proteins can objectively reflect the degree of evolutionary conservatism of proteins, let
2.4 Construction of the TNN
A Markov chain is a stochastic process, whose characteristic can be summarized as “the future depends on the past only through the present”, that is, the probability distribution of the next state can only be determined by the current state, and the events before it in the time series are independent of it. In a Markov chain, let
Due to strong predictive ability, Markov chains have been widely used in natural language processing, multivariate factor analysis, time series prediction and other fields. Inspired by the idea of Markov chains, in this manuscript, we designed a novel Transfer Neural Network called TNN, whose destination is being able to learn inherent feature representations from input data just like a Markov chain. In TNN, we introduced three main parameters such as the transition probability matrix W, the antecedent memory coefficient α with value between 0 and 1, and a bias term b. Let
In the training process, TNN will adopt the gradient descent algorithm to optimize all parameters including W, α and b in above Eq. 12, and can automatically find a set of optimal values for all these parameters. Thereafter, in the comparative experiments, through a series of complex calculations performed by itself and previous layers in TNNM based on these optimized parameters, TNN is able to assign larger weight values to more important features of proteins, and extract the most important features of proteins from the input data of TNNM, thus achieving satisfactory feature enhancement.
2.5 Construction of TNNM
Firstly, as illustrated in Figure 1, let
Secondly, considering that
Thirdly, in the ReLU layer of TNNM, we adopt the following activation function:
Here,
Moreover, in order to solve the problem of over fitting and reduce the training time of TNNM, we introduced two Dropout layers before and after the TNN layer. When each round of samples is inputted into TNNM for training, a probability p will be set in the Dropout layer so that each neuron will participate in training with the probability 1-p, that is, each neuron has a probability p of death. During experiments, we will set 0.7 to p in this manuscript.
Next, in the TNN layer, it is obvious that its input data is a
Finally, in order to estimate the criticality of proteins, we will set the input and output dimensions of the last Linear layer in TNNM as six and 1 separately, that is, TNNM will output 0 or 1 as its final predicted score.
Especially, in each Linear layer of TNNM, we will adopt the following Linear function:
Here,
2.6 Identification algorithm based on TNNM
Based on above description, we can present the identification algorithm based on TNNM as follows:
Step1: Based on the datasets of known PPIs downloaded from well-known public databases, constructing the original
Step2: According to Eqs. 2 and 4, 5, extracting three kinds of important topological features for proteins from
Step3: According to Eqs. 7, 8 and 10, extracting three kinds of important biological features for proteins separately.
Step4: According to methods proposed in section 2.4 and section 2.5, constructing the TNN based identification model TNNM first, and then, obtaining the predicted criticality scores for proteins through taking the matrix
3 Experimental results and analysis
During experiments, we will first divide the dataset of downloaded known PPI data into K subsets of proteins with the same size and proportion according to the proportion of essential proteins and non-essential proteins. And then, the K-fold cross validation will be adopted to evaluate the prediction performance of TNNM in this section.
3.1 Value selection of the parameter K
According to known results (Jung, 2017), the parameter K shall satisfy K ≈ log(N) and N/K > 3 days, where d is the number of extracted features. Hence, we can obtain the possible values of K as the following Table 1.
3.2 Comparison with representative methods
In this section, TNNM will be compared with 11 advanced competitive methods based on the DIP2010 database. Figure 2 shows the comparison results of the numbers of real essential proteins identified by TNNM and 11 recognition methods based on the DIP2010 database. During experiments, proteins will be sorted first in descending order according to predicted scores calculated by each competing methods, such as DC, IC, NC, SC, Pec, POME, CoEWC, ION, CVIM, NPRI, RWHN and TNNM. And then, we will select the top 1%, 5%, 10%, 15%, 20% and 25% proteins as candidate essential proteins. Finally, by comparing with the downloaded dataset of known essential proteins, the number of real essential proteins in the candidate essential proteins identified by each method will be calculated, and used to compare and evaluate the recognition ability of essential proteins of different methods.
FIGURE 2. Comparison between TNNM and 11 competitive methods based on the DIP 2010. (A) The number of key proteins in the top 1% candidate proteins. (B) Number of key proteins in the top 5% candidate proteins. (C) Number of key proteins in the top 10% candidate proteins. (D) Number of key proteins in the top 15% candidate proteins. (E) Number of key proteins in the top 20% candidate proteins. (F) Number of key proteins in the top 25% candidate proteins. In above figures, the number in parentheses represents the number of proteins in each interval.
From observing Figure 2, it is easy to know that TNNM outperforms all these competitive state-of-the-art prediction methods significantly based on the experimental results on DIP2010 database. And especially, among the top 1%, top 5%, and top 10% candidate key proteins, TNNM can achieve recognition accuracies of 96.07%, 99.21%, and 97.45% separately, which are all higher than 97%. Besides, among the top 15% and 20% candidate key proteins, the recognition accuracy rates of TNNM are all higher than 94%. Even for the top 25% candidate proteins, TNNM can maintain the accuracy rate above 85%.
3.3 Evaluation based on the folding knife curve
In this section, the Jackknife method (Holman et al., 2009) will be used, based on the top 1000 candidate essential proteins predicted on the DIP2010 database by TNNM and 11 competitive methods, to compare their performance in identifying essential proteins. Comparison results are shown in Figure 3.
FIGURE 3. Comparison results between TNNM and 11 competitive methods based on the jackknife method and the DIP2010 database. (A) Shows comparison results between BC, CC, DC, IC, NC, SC and TNNM, and (B) shows comparison results between Pec, CoEWC, POEM, ION, CVIM and TNNM. The X-axis represents the number of potential essential proteins predicted by each method, while the Y-axis represents the cumulative count of real essential proteins.
From Figure 3A and Figure 3B, it can be seen that with the increasing of the number of predicted proteins, the gap in term of essential protein recognition performance between TNNM and these competitive methods will grow wider and wider, which means that the prediction performance of TNNM is much better than that of these 11 competitive methods.
3.4 Comparison between TNN and representative neural networks
In order to verify the contribution of TNN to TNNM, we will compare TNN with six commonly used neural networks in this section based on the DIP2010 database, and comparison results is illustrated in Figure 4. During experiments, in TNNM, the TNN layer will be replaced by competitive neural networks such as Linear, CNN, RNN, GRU, LSTM and Transformer in turn. And then, the top 1%, 5%, 10%, 15%, 20% and 25% predicted proteins will be compared with downloaded dataset of known essential proteins. Finally, the number of real essential proteins in the candidate essential proteins identified by each method will be calculated, and used to compare and evaluate the recognition ability of essential proteins of different methods.
FIGURE 4. Performance comparison of TNNM by adopting TNN and six representative neural networks based on the DIP2010 database. (A) The number of essential proteins in the top 1% candidate proteins. (B) The number of essential proteins in the top 5% candidate proteins. (C) The number of essential proteins in the top 10% candidate proteins. (D) The number of essential proteins in the top 15% candidate proteins. (E) The number of essential proteins in the top 20% candidate proteins. (F) The number of essential proteins in the top 25% candidate proteins. In above figures, the number in parentheses represents the number of proteins in each interval.
From Figure 4, it is easy to see that if the TNN in TNNM is replaced by Linear, CNN, RNN, GRU, LSTM or Transformer, the prediction performance of TNNM will turn to be poorer, which reflects that TNN plays a positive role in the prediction performance of TNNM.
3.5 Recognition performance based on the gavin database
To prove the universal applicability of TNNM, in this section, we further compared TNNM with 11 competitive recognition methods based on the Gavin database, and illustrated comparison results in Figure 5.
FIGURE 5. Performance comparison between TNNM and 11 competitive algorithms based on the Gavin database. (A) The number of essential proteins in the top 1% candidate proteins. (B) The number of essential proteins in the top 5% candidate proteins. (C) The number of essential proteins in the top 10% candidate proteins. (D) The number of essential proteins in the top 15% candidate proteins. (E) The number of essential proteins in the top 20% candidate proteins. (F) The number of essential proteins in the top 25% candidate proteins. In above figures, the number in parentheses represents the number of proteins in each interval.
From Figure 5, it is obvious that the recognition performance of TNNM is significantly superior to all these 11 competing methods. Especially, among the top 1%, top 5%, and top 10% candidate essential proteins, TNNM can achieve accuracies of 94.73%, 98.92%, and 96.77% respectively, which are all higher than 94%. Besides, among the top 15% and 20% candidate essential proteins, the recognition accuracies of TNNM are higher than 86% as well. Even in the top 25% candidate proteins, TNNM can also maintain its accuracy rate above 85%. Hence, we can say that TNNM has much better universal applicability than all these competitive methods.
4 Conclusion
In this manuscript, a novel prediction model named TNNM was designed to identify essential proteins, and through intensive experiments, we demonstrated that TNNM outperformed various advanced algorithms in terms of both prediction accuracies and decline rate of accuracies. The major contributions of TNNM include: 1) we designed a new Transfer Neural Network (TNN), which can extract raw features from multiple biological information of proteins efficiently. 2) we introduced a TNN layer into the prediction model TNNM, which can not only improve the prediction accuracy of TNNM, but also enhance both the robustness and the non-linear expression ability of TNNM. Intensive experiments have demonstrated that TNNM can achieve satisfactory prediction accuracy in different databases, and simultaneously, TNN plays an irreplaceable positive role in TNNM as well.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding authors.
Author contributions
SC and CH produced the main ideas, and did the modeling, computation and analysis and also wrote the manuscript. LW and SZ provided supervision and effective scientific advice and related ideas, research design guidance, and added value to the article through editing and contributing completions. All authors contributed to the article and approved the submitted version.
Funding
This work was partly sponsored by the National Natural Science Foundation of China (No.62272064, No.61873221) and the Key project of Changsha Science and technology Plan (No. KQ2203001).
Acknowledgments
The authors thank the anonymous referees for suggestions that helped improve the paper substantially.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fgene.2022.1087294/full#supplementary-material
References
Acencio, M. L., and Lemke, N. (2009). Towards the prediction of essential genes by integration of network topology, cellular localization and biological process information. BMC Bioinforma. 10 (1), 290–307. doi:10.1186/1471-2105-10-290
Binder, J. X., Pletscher-Frankild, S., Tsafou, K., Stolte, C., O'Donoghue, S. I., Schneider, R., et al. (2014). Compartments: Unification and visualization of protein subcellular localization evidence. Database 2014, bau012. doi:10.1093/database/bau012
Chen, B., and Wu, F. X. (2013). Identifying protein complexes based on multiple topological structures in PPI networks. IEEE Trans Nanobio- Sci. 12 (3), 165–172. doi:10.1109/TNB.2013.2264097
Chen, Z., Meng, Z., Liu, C., Wang, X., Kuang, L., Pei, T., et al. (2020). A novel model for predicting essential proteins based on heterogeneous protein-domain network. IEEE Access 8 (99), 8946–8958. doi:10.1109/access.2020.2964571
Cherry, J. M., Adler, C., Ball, C., Chervitz, S. A., Dwight, S. S., Hester, E. T., et al. (1998). SGD: Saccharomyces genome database. Nucleic Acids Res. 26, 73–79. doi:10.1093/nar/26.1.73
Dai, W., Chang, Q., Peng, W., Zhong, J., and Li, Y. (2020). Network embedding the protein-pro tein interaction network for human essential genes identification. Genes 11 (2), 153. doi:10.3390/genes11020153
Gabriel, O., Thomas, S., Kristoffer, F., Kostler, T., Messina, D. N., Roopra, S., et al. (2010). InParanoid 7: New algorithms and tools for eukaryotic orthology analysis. Nucleic Acids Res. 38, D196–D203. doi:10.1093/nar/gkp931
Gavin, A. C., Aloy, P., Grandi, P., Krause, R., Boesche, M., Marzioch, M., et al. (2006). Proteome survey reveals modularity of the yeast cell machinery. Nature 440 (7084), 631–636. doi:10.1038/nature04532
Hahn, M. W., and Kern, A. D. (2005). Comparative genomics of centrality and essentiality in three eukaryotic protein-interaction networks. Mol. Biol. Evol. 22 (4), 803–806. doi:10.1093/molbev/msi072
Holman, A. G., Davis, P. J., Foster, J. M., Carlow, C. K. S., and Kumar, S. (2009). Computational prediction of essential genes in an unculturable endosymbiotic bacterium, Wolbachia of Brugia malayi. BMC Microbiol. 9 (1), 243. doi:10.1186/1471-2180-9-243
Jeong, H. M., Mason, S. P., Al, B., and Oltvai, Z. N. (2001). Lethality and centrality in protein networks. Nature 411 (6833), 41–42. doi:10.1038/35075138
Jiawei, L., Yi, Q., and Peter, C. (2015). Identification of essential proteins based on a new combination of local interaction density and protein complexes. Plos One 10 (6), e0131418. doi:10.1371/journal.pone.0131418
Jung, Y. (2017). Multiple predicting K -fold cross-validation for model selection. J. Nonparametric Statistics 30, 197–215. doi:10.1080/10485252.2017.1404598
Keretsu, S., and Sarmah, R. (2016). Weighted edge based clustering to identify protein complexes in protein protein inter- action networks incorporating gene expression profile. Comput. Biol. Chem. 65, 69–79. doi:10.1016/j.compbiolchem.2016.10.001
Lei, X., Yang, X., and Wu, F. (2018). Artificial fish swarm optimization based method to identify essential proteins. IEEE/ACM Trans. Comput. Biol. Bioinform. 17, 495–505. doi:10.1109/TCBB.2018.2865567
Li, M., Lu, Y., Niu, Z., and Wu, F. X. (2017). United complex centrality for identification of essential proteins from PPI networks. IEEE/ACM Trans. Comput. Biol. Bioinform. 14 (2), 370–380. doi:10.1109/TCBB.2015.2394487
Li, M., Lu, Y., Wang, J., Wu, F. X., and Pan, Y. (2015). A topology potential-based method for identifying essential proteins from PPI networks. IEEE/ACM Trans. Comput. Biol. Bioinform. 12 (2), 372–383. doi:10.1109/TCBB.2014.2361350
Li, M., Wang, J. X., Chen, X., Wang, H., and Pan, Y. (2011). A local average connectivity-based method for identifying essential proteins from the network level. Comput. Biol. Chem. 35 (3), 143–150. doi:10.1016/j.compbiolchem.2011.04.002
Li, M., Zhang, H., Wang, J. X., and Pan, Y. (2012). A new essential protein discovery method based on the integration of protein-protein interaction and gene expression data. BMC Syst. Biol. 6 (1), 15. doi:10.1186/1752-0509-6-15
Li, S., Chen, Z., He, X., Zhang, Z., Pei, T., Tan, Y., et al. (2020). An iteration method for identifying yeast essential proteins from weighted PPI network based on topological and functional features of proteins. IEEE Access 8, 90792–90804. doi:10.1109/access.2020.2993860
Luo, J., and Wu, J. (2015). A new algorithm for essential proteins identification based on the integration of protein complex co-expression information and edge clustering coefficient. Int. J. Data Min. Bioinform 12 (3), 257–274. doi:10.1504/ijdmb.2015.069654
Maliackal, P. J., Amy, B., Donald, E. I., and Sui, H. (2005). High-Betweenness proteins in the yeast protein interaction network. J. Biomed. Biotechnol. 2005 (2), 96–103. doi:10.1155/JBB.2005.96
Maslov, S., and Sneppen, K., (2002). Specificity and stability in topology of protein networks. Science 296, 910–913. doi:10.1126/science.1065103
Mewes, H. W., Frishman, D., Mayer, K. F. X., Munsterkotter, M., Noubibou, O., Pagel, P., et al. (2006). MIPS: Analysis and annotation of proteins from whole genomes in 2005. Nucleic Acids Res. 34, D169–D172. doi:10.1093/nar/gkj148
Peng, W., Wang, J. X., Cheng, Y., Lu, Y., Wu, F., and Pan, Y. (2015). UDoNC: An algorithm for identifying essential proteins based on protein domains and protein-protein interaction networks. IEEE/ACM Trans Comput. Biol. Bioinforma. 12 (2), 276–288. doi:10.1109/TCBB.2014.2338317
Peng, W., Wang, J. X., Wang, W., Liu, Q., Wu, F. X., and Pan, Y. (2012). Iteration method for predicting essential proteins based on orthology and protein-protein interaction networks. BMC Syst. Biol. 6 (1), 87. doi:10.1186/1752-0509-6-87
Pržulj, N., Wigle, D. A., and Jurisica, I. (2004). Functional topology in a network of protein interactions. Bioinformatics 20 (3), 340–348. doi:10.1093/bioinformatics/btg415
Qi, Y., and Luo, J. (2016). Prediction of essential proteins based on local interaction density. IEEE/ACM Trans Computa- tional Biol. Bioinforma. 13 (6), 1170–1182. doi:10.1109/TCBB.2015.2509989
StanfordMedicine (2012). Saccharomyces genome deletion project. Available at: http://yeastdeletion.stanford.edu/(Accessed June 20, 2012).
Stefan, W., and Stadler, P. F. (2003). Centers of complex networks. J. Theor. Biol. 223 (1), 45–53. doi:10.1016/s0022-5193(03)00071-7
Stephenson, K., and Zelen, M. (1989). Rethinking centrality: Methods and examples. Soc. Netw. 11, 1–37. doi:10.1016/0378-8733(89)90016-6
Tang, X., Wang, J., Zhong, J., and Pan, Y. (2014). Predicting essential proteins based on weighted degree centrality. IEEE/ACM Trans Comput. Biol. Bioinforma. 11 (2), 407–418. doi:10.1109/TCBB.2013.2295318
Tu, B. P., Kudlicki, A., Rowicka, M., and McKnight, S. L. (2005). Logic of the yeast metabolic cycle: Temporal compartmentalization of cellular processes. Science 310 (5751), 1152–1158. doi:10.1126/science.1120499
Wang, J. X., Li, M., Wang, H., and Pan, Y. (2012). Identification of essential proteins based on edge clustering coefficient. IEEE/ACM Trans. Comput. Biol. Bioinform. 9 (4), 1070–1080. doi:10.1109/TCBB.2011.147
Xenarios, I., Salwinski, L., Duan, X. J., Higney, P., Kim, S. M., and Eisenberg, D. (2002). DIP, the database of interacting proteins: A research tool for studying cellular networks of protein interactions. Nucleic Acids Res. 30 (1), 303–305. doi:10.1093/nar/30.1.303
Zhang, F., Peng, W., Yang, Y., Dai, W., and Song, J. (2019). A novel method for identifying essential genes by fusing dynamic protein–protein interactive networks. Genes 10 (1), 31. doi:10.3390/genes10010031
Zhang, R., and Lin, Y. (2009). DEG 5.0, a database of essential genes in both prokaryotes and eukaryotes. Nucleic Acids Res. 37, D455–D458. doi:10.1093/nar/gkn858
Zhang, X., Xu, J., and Xiao, W. (2013). A new method for the discovery of essential proteins. PLoS One 8 (3), e58763. doi:10.1371/journal.pone.0058763
Zhao, B., Han, X., Liu, X., Luo, Y., Hu, S., Zhang, Z., et al. (2020). A novel method to predict essential proteins based on diffusion distance networks. IEEE ACCESS 8, 29385–29394. doi:10.1109/access.2020.2972922
Zhao, B., Hu, S., Liu, X., Xiong, H., Han, X., Zhang, Z., et al. (2020). A novel computational approach for identifying essential proteins from multiplex biological networks. Front. Genet. 11, 343. doi:10.3389/fgene.2020.00343
Zhao, B., Zhao, Y., Zhang, X., Zhang, Z., Zhang, F., and Wang, L. (2019). An iteration method for identifying yeast essential proteins from heterogeneous network. BMC Bioinforma. 20, 355. doi:10.1186/s12859-019-2930-2
Keywords: essential protein, prediction model, transfer neural network, biological information, internal disease, protein-protein interaction network
Citation: Chen S, Huang C, Wang L and Zhou S (2023) A disease-related essential protein prediction model based on the transfer neural network. Front. Genet. 13:1087294. doi: 10.3389/fgene.2022.1087294
Received: 02 November 2022; Accepted: 14 December 2022;
Published: 04 January 2023.
Edited by:
Marco S Nobile, Ca’ Foscari University of Venice, ItalyReviewed by:
Yuhua Yao, Hainan Normal University, ChinaChaoyang Pang, Sichuan Normal University, China
Copyright © 2023 Chen, Huang, Wang and Zhou. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Chiguo Huang, MjM5NDE4MTU1N0BxcS5jb20=; Lei Wang, d2FuZ2xlaUB4dHUuZWR1LmNu; Shunxian Zhou, enhzX2hkQGhudS5lZHUuY24=