Skip to main content

ORIGINAL RESEARCH article

Front. Genet., 04 January 2023
Sec. Computational Genomics

A disease-related essential protein prediction model based on the transfer neural network

Sisi ChenSisi Chen1Chiguo Huang
Chiguo Huang2*Lei Wang,
Lei Wang1,2*Shunxian Zhou,,
Shunxian Zhou1,3,2*
  • 1The First Hospital of Hunan University of Chinese Medicine, Changsha, Hunan, China
  • 2Big Data Innovation and Entrepreneurship Education Center of Hunan Province, Changsha University, Changsha, China
  • 3College of Information Science and Engineering, Hunan Women’s University, Changsha, Hunan, China

Essential proteins play important roles in the development and survival of organisms whose mutations are proven to be the drivers of common internal diseases having higher prevalence rates. Due to high costs of traditional biological experiments, an improved Transfer Neural Network (TNN) was designed to extract raw features from multiple biological information of proteins first, and then, based on the newly-constructed Transfer Neural Network, a novel computational model called TNNM was designed to infer essential proteins in this paper. Different from traditional Markov chain, since Transfer Neural Network adopted the gradient descent algorithm to automatically obtain the transition probability matrix, the prediction accuracy of TNNM was greatly improved. Moreover, additional antecedent memory coefficient and bias term were introduced in Transfer Neural Network, which further enhanced both the robustness and the non-linear expression ability of TNNM as well. Finally, in order to evaluate the identification performance of TNNM, intensive experiments have been executed based on two well-known public databases separately, and experimental results show that TNNM can achieve better performance than representative state-of-the-art prediction models in terms of both predictive accuracies and decline rate of accuracies. Therefore, TNNM may play an important role in key protein prediction in the future.

1 Introduction

Essential proteins refer to proteins that removal of them will lead to cell death or infertility (Acencio and Lemke, 2009). Identification of essential proteins can help us understand the basic requirements for maintaining life forms. However, considering that it is very costly and time-consuming to identify essential proteins by adopting traditional biological experimental techniques such as gene knock-out (Maliackal et al., 2005), RNA interference (Jeong et al., 2001) and conditional knockout (Hahn and Kern, 2005), more and more computational models have been proposed to infer essential proteins in recent years based on the topological characteristics of Protein-Protein Interaction (PPI) networks, in which, proteins are the vertices of PPI networks, while the interactions between proteins constitute the edges. Researches demonstrated that the topological position of a protein in the protein network has a strong correlation with whether the protein is critical (Jeong et al., 2001; Maslov and Sneppen, 2002; Pržulj et al., 2004). Based on the topological characteristics of proteins in PPI networks, a series of essential protein recognition methods have been proposed. For instance, based on the centrality of PPI networks (Stephenson and Zelen, 1989; Jeong et al., 2001), different computational models including CC (Closeness Centrality) (Stefan and Stadler, 2003), DC (Degree Centrality) (Hahn and Kern, 2005), BC (Between Centrality) (Wang et al., 2012), SC (Graph Centrality) (Maliackal et al., 2005) and NC (Neighbor Centrality) (Wang et al., 2012) have been proposed in succession. In addition, Li M et al. designed a recognition model named LAC (Li et al., 2011) to detect essential proteins (Li et al., 2015) based on the local average connectivity of protein nodes in the PPI network. Qi Yi et al. (Qi and Luo, 2016) introduced a prediction model based on the local interaction density (LID) of protein nodes in the PPI network to infer essential proteins. Chen B et al. (Chen and Wu, 2013) proposed an essential protein recognition method based on multiple topological features of the PPI network. In all these above methods, only topological characteristics of the PPI network were considered to identify essential proteins, however, since there is a large amount of noise data in PPI networks, then the predictive accuracy of these methods is not very satisfactory.

In order to break through the inherent limitations of existing PPI data, in the past few years, people proposed novel models by combining the topological characteristics of PPI networks with biological information of proteins. For example, M Li et al. and Xiwei Tang et al. put forward prediction models called Pec (Li et al., 2012) and WDC (Tang et al., 2014) by integrating PPI network and gene expression data of proteins respectively. W Peng et al. designed a prediction model (Peng et al., 2012) by integrating protein homology information with PPI networks, and a prediction model (Peng et al., 2015) through combining protein domain information with PPI network, simultaneously. X Zhang et al. (Zhang et al., 2013) introduced a recognition method called CoEWC by merging topological characteristics of the PPI network with the co expression characteristics of proteins. BH Zhao et al. designed a prediction model named POEM (Zhao et al., 2014) by combining gene expression data of proteins with the topological characteristics of PPI networks. J Luo et al. proposed a identification method based on local interaction density of PPI networks and biological characteristics of protein complexes (Jiawei et al., 2015). Seketoulie Keretsu et al. presented a protein complex recognition model (Li et al., 2015) based on clustering weighted edges and gene expression profile of proteins. M Li et al. designed two essential protein recognition methods by combining PPI networks with subcellular location information and complex centrality of proteins respectively (Keretsu and Sarmah, 2016; Li et al., 2017; Chen et al., 2020). J Luo et al. introduced a prediction method ECC (edge clustering coefficient) based on the complex co expression data of proteins and PPI networks (Luo and Wu, 2015). Bihai Zhao et al. proposed a model based on multiple biological networks (Zhao et al., 2020a) and a model based on diffusion distance network (Zhao et al., 2020b) to predict essential proteins respectively. S. Li et al. designed an iterative method called CVIM (Li et al., 2020) based on topological and functional characteristics of proteins to predict key proteins. Lei X et al. proposed a necessary protein prediction method AFSOEP (Lei et al., 2018) to infer protein complexes through AFSO (Artificial Fish Swarm Optimization). BH Zhao et al. designed an iterative method to identify potential essential proteins (Zhao et al., 2019) based on heterogeneous PPI networks. Dai W et al. proposed a method to discover key genes based on protein-protein interaction network embedding (Dai et al., 2020). Fengyu Zhang et al. predicted the key gene (Zhang et al., 2019) by fusing the dynamic PPI network. Chen Z et al. proposed an essential protein prediction model NPRI based on heterogeneous network, and established heterogeneous protein domain network (Chen et al., 2020) according to initial PPI network, protein domain network and gene expression data.

All these above methods show that the identification accuracy of models can be significantly improved by combining biological information of proteins with topological features of PPI networks. However, through analyzing results achieved by these existing methods, it is not difficult to find that the prediction accuracies of these algorithms decline fast with the increasing of predicted essential proteins. Hence, inspired by recognition models based on the Markov chain and the Transfer algorithm, we designed a new neural network called TNN in this manuscript, based on which, a novel model named TNNM was proposed to predict essential proteins. TNN can be divided into three parts, namely, probability transfer matrix, antecedent output and bias term. In addition, in order to evaluate the performance of TNNM, we compared it with existing representative models such as IC (Stephenson and Zelen, 1989), DC (Hahn and Kern, 2005), SC (Maliackal et al., 2005), NC (Wang et al., 2012), PeC (Li et al., 2012), ION (Peng et al., 2012), CoEWC (Zhang et al., 2013), POEM (Zhao et al., 2014), CVIM (Li et al., 2020), NPRI (Chen et al., 2020) and RWHN (Zhao et al., 2014) separately. Experimental results show that TNNM is far superior to these traditional models in terms of both predictive accuracy and decline rate of accuracy.

The rest of this paper is organized as follows: The experimental data and specific steps are organized in Section 2. In Section 3, the influence of parameters and comparison with other methods are shown. Section 4 describes the shortcomings of the model and future improvement goals. Finally, a summary is made in Part 5.

2 Method and materials

The flow chart of TNNM is shown in Figure 1. Through observing Figure 1, it is easy to see that TNNM consists of the following three major parts. Firstly, based on prior knowledge, topological features and biological features of each protein will be extracted from PPI networks, gene expression data, subcellular localization and ortholog data of proteins separately. And then, the Transfer Neural Network (TNN) will be designed. Finally, through adopting TNN, the prediction model TNNM will be constructed to infer essential proteins based on these newly extracted features.

FIGURE 1
www.frontiersin.org

FIGURE 1. Flowchart of TNNM.

2.1 Experimental data

In order to evaluate the prediction performance of TNNM, during experiments, we first downloaded known PPIs from different benchmark databases such as DIP 2010 (Xenarios et al., 2002) and Gavin (Gavin et al., 2006) respectively. After preprocessing, a dataset containing 5093 proteins and 24,743 known PPIs was finally obtained from the DIP2010 database, and a dataset containing 1855 proteins and 7669 known PPIs was obtained from the Gavin database. In addition, based on databases including MIPS (Mewes et al., 2006), SGD (Cherry et al., 1998), DEG (Zhang and Lin, 2009) and SGDP (StanfordMedicine, 2012), a benchmark dataset containing 1285 essential proteins was constructed, based on which, 1167 and 714 essential proteins were screened from the DIP2010 and Gavin databases respectively. Moreover, based on the dataset provided by Tu BP et al. (Tu et al., 2005), a dataset containing the gene expression data of 6776 proteins was obtained, which consists of the gene expression level data of proteins in the continuous metabolic cycle. Simultaneously, the homologous information of proteins was downloaded from the Inparanoid database (seventh edition), including paired comparison (Gabriel et al., 2010) between 100 whole genomes, and the number of times that proteins have homologous information in the reference organism. Finally, we downloaded the dataset containing subcellular localization information of proteins from the COMPART-MENTS database (Binder et al., 2014) (2014 version), and retained only 11 types of subcellular localization data closely related to essential proteins, such as cytoplasm, cytoskeleton, Golgi apparatus, cytoplasm, vacuoles, mitochondria, endosomes, plasma, nucleus, peroxisomes and extracellular enzymes, etc.

Based on above newly-downloaded datasets, firstly, we constructed an original PPI network. And then, through combining with the existing complex network topological features including degree centrality, closeness centrality, node betweenness centrality and edge betweenness centrality, some new important protein topological features are extracted from the PPI network, including the degree of contact between the protein node and the neighborhood nodes, the importance of the protein node relative to the total distance, and the importance of the protein node relative to the carrying capacity. Simultaneously, we would further extract some biological features for proteins, including the importance of protein node relative to the Pearson correlation coefficient, the importance of protein node relative to the subcellular locations, and the importance of protein node relative to the homologous information, from multiple biological information existed in above newly-downloaded datasets.

2.2 Topological feature extraction

Let the undirected graph G=V,E represent the original PPI network formed by a dataset downloaded from any given base database, V=p1,p2,,pN denote the set of different proteins in the downloaded dataset, then, for any two given proteins pi and pj in V, we define that there is an edge epi,pj between pi and pj, if and only if there is a known interaction between them. And for convenience, we define that E=epi,pj|pi,pjV represents the set of edges in G. Hence, we can obtain the adjacency matrix A=aijN×N corresponding to G as follows: if there is epi,pjE, then there is aij=1, otherwise there is aij=0.

For any given protein pV in G, let NP be the set of neighboring nodes of p, then we have:

Np=q|qV,ep,qE(1)

Based on above formula (1), we define that the degree of contact between p and its neighboring nodes as follows:

TF1p=qNgpTrisp,qNp(2)

Here, Np represents the number of elements in Np, and Trisp,q denotes the number of common neighbors of p and q, which can be calculated as follows:

Trisp,q=NpNgqminNp,Nq,pNq,qNp0,otherwise(3)

Here, |N(p)∩N(q)| represents the number of elements in N(p)∩N(q).

It is reasonable to consume that the smaller the total distance between a protein and all other proteins, the more important the protein will be. Hence, let l (p, q) denote the length of the shortest path from protein p to the protein q in G, if there is no path between p and q in G, then we define the length of the shortest path between p and q is a constant number N (>1). Therefore, we can calculate the importance of p related to the total distance as follows:

TF2p=N1qVlp,q(4)

Moreover, it is also reasonable to assume that the more important a protein p is, the more proteins that have the shortest path through p. This indicator reflects the carrying capacity of p between other nodes in G. it is obvious that the larger the value, the greater the impact of p in the network, which also means that p will be more important. Hence, we can calculate the importance of p related to the carrying capacity as follows:

TF3p=pqqVkqqpkqq(5)

Here, kqq’ represents the number of shortest paths between q and q′ in G, and kqq′p) denotes the number of shortest paths between q and q′ in G, which pass through p.

2.3 Biological feature extraction

Let ge (p, t) represent the gene expression value of the protein p at the time point t, gep denote the average expression level of p at all n time points, and σp be the standard variance of the gene expression level of p at all n time points, then we can calculate the Pearson correlation coefficient between p and q as follows:

PCCp,q=1n1t=1ngep,tgepσpgeq,tgeqσq(6)

Based on above formula (6), we can calculate the importance of p related to the Pearson correlation coefficient as follows:

BF1p=qNgpPCCp,q(7)

It is reasonable to consume that essential proteins tend to be connected rather than independent. Therefore, we can believe that proteins closely related to essential proteins are more likely to be essential proteins. Thus, we can obtain another importance indicator of p as follows:

BF2p=qNgpBsubp,qNgp(8)

Where Bsubp,q can be obtained as follows:

Bsubp,q=SubpSubqSubpSubq+1(9)

Here, Sub(p) represents the set of subcellular locations of the protein p, |Sub(p)∩Sub(q)| denotes the number of elements in Sub(p)∩Sub(q), and |Sub(p)∪Sub(q)| is the number of elements in Sub(p)∪Sub(q).

Moreover, based on the reasonable assumption that the evolution of essential proteins is more conservative than that of non-essential proteins, and considering that the homologous information of proteins can objectively reflect the degree of evolutionary conservatism of proteins, let Osp denote the value of homologous score of p, then it is obvious that the higher the value of Osp, the more conservative the evolution of p will be, i.e., the more important the protein p will be. Thus calculate the importance of p related to the homologous information as follows:

BF3p=OspOsqqVmax(10)

2.4 Construction of the TNN

A Markov chain is a stochastic process, whose characteristic can be summarized as “the future depends on the past only through the present”, that is, the probability distribution of the next state can only be determined by the current state, and the events before it in the time series are independent of it. In a Markov chain, let Tn denote the state space at time step n, and Q represent the transition probability matrix, then there is:

Tn+1=QTn(11)

Due to strong predictive ability, Markov chains have been widely used in natural language processing, multivariate factor analysis, time series prediction and other fields. Inspired by the idea of Markov chains, in this manuscript, we designed a novel Transfer Neural Network called TNN, whose destination is being able to learn inherent feature representations from input data just like a Markov chain. In TNN, we introduced three main parameters such as the transition probability matrix W, the antecedent memory coefficient α with value between 0 and 1, and a bias term b. Let Xi denote the input data of the ith layer in TNN, then similar to the principle of Markov chains, we define its output Xi+1 as follows:

Xi+1=α*W*Xi+1α*Xi+b(12)

In the training process, TNN will adopt the gradient descent algorithm to optimize all parameters including W, α and b in above Eq. 12, and can automatically find a set of optimal values for all these parameters. Thereafter, in the comparative experiments, through a series of complex calculations performed by itself and previous layers in TNNM based on these optimized parameters, TNN is able to assign larger weight values to more important features of proteins, and extract the most important features of proteins from the input data of TNNM, thus achieving satisfactory feature enhancement.

2.5 Construction of TNNM

Firstly, as illustrated in Figure 1, let X0=X0p1,X0p2,,X0pNT denote the input data of the input layer in TNNM, then for any given protein piV, there is:

X0pi=<TF1piTF2piTF3piBF1piBF2piBF3pi>(13)

Secondly, considering that X0 is a N×6 dimensional matrix, during experiments, we set the input and output dimensions of the first Linear layer in TNNM as six and 8 separately.

Thirdly, in the ReLU layer of TNNM, we adopt the following activation function:

Xjki=Xjki1:ifXjki1>00:otherwise(14)

Here, Xjki denotes the element in the jth row and kth column of Xi. And Xi and Xi1 represent the input and output data of the ReLU layer respectively.

Moreover, in order to solve the problem of over fitting and reduce the training time of TNNM, we introduced two Dropout layers before and after the TNN layer. When each round of samples is inputted into TNNM for training, a probability p will be set in the Dropout layer so that each neuron will participate in training with the probability 1-p, that is, each neuron has a probability p of death. During experiments, we will set 0.7 to p in this manuscript.

Next, in the TNN layer, it is obvious that its input data is a N× 8 dimensional matrix, and for each protein, an 8-dimensional feature vector will be extracted by TNN as its output. Hence, in the second Linear layer of TNNM, we will set its input and output dimensions as eight and six respectively.

Finally, in order to estimate the criticality of proteins, we will set the input and output dimensions of the last Linear layer in TNNM as six and 1 separately, that is, TNNM will output 0 or 1 as its final predicted score.

Especially, in each Linear layer of TNNM, we will adopt the following Linear function:

Xi+1=XiW+b(15)

Here, W is a matrix with m rows and n columns, where m and n denote the dimensions of input and output data of the Linear layer respectively. For instance, it is obvious that in these three Linear layers of TNNM, the dimensions of matrix W will be 6 × 8, 8 × 6 and 6 × 1 respectively. And additionally, Xi+1 and Xi represent the input and output data of the Linear layer respectively.

2.6 Identification algorithm based on TNNM

Based on above description, we can present the identification algorithm based on TNNM as follows:

Step1: Based on the datasets of known PPIs downloaded from well-known public databases, constructing the original PPI network G and the corresponding adjacency matrix A.

Step2: According to Eqs. 2 and 4, 5, extracting three kinds of important topological features for proteins from G respectively.

Step3: According to Eqs. 7, 8 and 10, extracting three kinds of important biological features for proteins separately.

Step4: According to methods proposed in section 2.4 and section 2.5, constructing the TNN based identification model TNNM first, and then, obtaining the predicted criticality scores for proteins through taking the matrix X0 computed by Eq. 13 as the input data of TNNM.

3 Experimental results and analysis

During experiments, we will first divide the dataset of downloaded known PPI data into K subsets of proteins with the same size and proportion according to the proportion of essential proteins and non-essential proteins. And then, the K-fold cross validation will be adopted to evaluate the prediction performance of TNNM in this section.

3.1 Value selection of the parameter K

According to known results (Jung, 2017), the parameter K shall satisfy K ≈ log(N) and N/K > 3 days, where d is the number of extracted features. Hence, we can obtain the possible values of K as the following Table 1.

TABLE 1
www.frontiersin.org

TABLE 1. Values of K in different database.

3.2 Comparison with representative methods

In this section, TNNM will be compared with 11 advanced competitive methods based on the DIP2010 database. Figure 2 shows the comparison results of the numbers of real essential proteins identified by TNNM and 11 recognition methods based on the DIP2010 database. During experiments, proteins will be sorted first in descending order according to predicted scores calculated by each competing methods, such as DC, IC, NC, SC, Pec, POME, CoEWC, ION, CVIM, NPRI, RWHN and TNNM. And then, we will select the top 1%, 5%, 10%, 15%, 20% and 25% proteins as candidate essential proteins. Finally, by comparing with the downloaded dataset of known essential proteins, the number of real essential proteins in the candidate essential proteins identified by each method will be calculated, and used to compare and evaluate the recognition ability of essential proteins of different methods.

FIGURE 2
www.frontiersin.org

FIGURE 2. Comparison between TNNM and 11 competitive methods based on the DIP 2010. (A) The number of key proteins in the top 1% candidate proteins. (B) Number of key proteins in the top 5% candidate proteins. (C) Number of key proteins in the top 10% candidate proteins. (D) Number of key proteins in the top 15% candidate proteins. (E) Number of key proteins in the top 20% candidate proteins. (F) Number of key proteins in the top 25% candidate proteins. In above figures, the number in parentheses represents the number of proteins in each interval.

From observing Figure 2, it is easy to know that TNNM outperforms all these competitive state-of-the-art prediction methods significantly based on the experimental results on DIP2010 database. And especially, among the top 1%, top 5%, and top 10% candidate key proteins, TNNM can achieve recognition accuracies of 96.07%, 99.21%, and 97.45% separately, which are all higher than 97%. Besides, among the top 15% and 20% candidate key proteins, the recognition accuracy rates of TNNM are all higher than 94%. Even for the top 25% candidate proteins, TNNM can maintain the accuracy rate above 85%.

3.3 Evaluation based on the folding knife curve

In this section, the Jackknife method (Holman et al., 2009) will be used, based on the top 1000 candidate essential proteins predicted on the DIP2010 database by TNNM and 11 competitive methods, to compare their performance in identifying essential proteins. Comparison results are shown in Figure 3.

FIGURE 3
www.frontiersin.org

FIGURE 3. Comparison results between TNNM and 11 competitive methods based on the jackknife method and the DIP2010 database. (A) Shows comparison results between BC, CC, DC, IC, NC, SC and TNNM, and (B) shows comparison results between Pec, CoEWC, POEM, ION, CVIM and TNNM. The X-axis represents the number of potential essential proteins predicted by each method, while the Y-axis represents the cumulative count of real essential proteins.

From Figure 3A and Figure 3B, it can be seen that with the increasing of the number of predicted proteins, the gap in term of essential protein recognition performance between TNNM and these competitive methods will grow wider and wider, which means that the prediction performance of TNNM is much better than that of these 11 competitive methods.

3.4 Comparison between TNN and representative neural networks

In order to verify the contribution of TNN to TNNM, we will compare TNN with six commonly used neural networks in this section based on the DIP2010 database, and comparison results is illustrated in Figure 4. During experiments, in TNNM, the TNN layer will be replaced by competitive neural networks such as Linear, CNN, RNN, GRU, LSTM and Transformer in turn. And then, the top 1%, 5%, 10%, 15%, 20% and 25% predicted proteins will be compared with downloaded dataset of known essential proteins. Finally, the number of real essential proteins in the candidate essential proteins identified by each method will be calculated, and used to compare and evaluate the recognition ability of essential proteins of different methods.

FIGURE 4
www.frontiersin.org

FIGURE 4. Performance comparison of TNNM by adopting TNN and six representative neural networks based on the DIP2010 database. (A) The number of essential proteins in the top 1% candidate proteins. (B) The number of essential proteins in the top 5% candidate proteins. (C) The number of essential proteins in the top 10% candidate proteins. (D) The number of essential proteins in the top 15% candidate proteins. (E) The number of essential proteins in the top 20% candidate proteins. (F) The number of essential proteins in the top 25% candidate proteins. In above figures, the number in parentheses represents the number of proteins in each interval.

From Figure 4, it is easy to see that if the TNN in TNNM is replaced by Linear, CNN, RNN, GRU, LSTM or Transformer, the prediction performance of TNNM will turn to be poorer, which reflects that TNN plays a positive role in the prediction performance of TNNM.

3.5 Recognition performance based on the gavin database

To prove the universal applicability of TNNM, in this section, we further compared TNNM with 11 competitive recognition methods based on the Gavin database, and illustrated comparison results in Figure 5.

FIGURE 5
www.frontiersin.org

FIGURE 5. Performance comparison between TNNM and 11 competitive algorithms based on the Gavin database. (A) The number of essential proteins in the top 1% candidate proteins. (B) The number of essential proteins in the top 5% candidate proteins. (C) The number of essential proteins in the top 10% candidate proteins. (D) The number of essential proteins in the top 15% candidate proteins. (E) The number of essential proteins in the top 20% candidate proteins. (F) The number of essential proteins in the top 25% candidate proteins. In above figures, the number in parentheses represents the number of proteins in each interval.

From Figure 5, it is obvious that the recognition performance of TNNM is significantly superior to all these 11 competing methods. Especially, among the top 1%, top 5%, and top 10% candidate essential proteins, TNNM can achieve accuracies of 94.73%, 98.92%, and 96.77% respectively, which are all higher than 94%. Besides, among the top 15% and 20% candidate essential proteins, the recognition accuracies of TNNM are higher than 86% as well. Even in the top 25% candidate proteins, TNNM can also maintain its accuracy rate above 85%. Hence, we can say that TNNM has much better universal applicability than all these competitive methods.

4 Conclusion

In this manuscript, a novel prediction model named TNNM was designed to identify essential proteins, and through intensive experiments, we demonstrated that TNNM outperformed various advanced algorithms in terms of both prediction accuracies and decline rate of accuracies. The major contributions of TNNM include: 1) we designed a new Transfer Neural Network (TNN), which can extract raw features from multiple biological information of proteins efficiently. 2) we introduced a TNN layer into the prediction model TNNM, which can not only improve the prediction accuracy of TNNM, but also enhance both the robustness and the non-linear expression ability of TNNM. Intensive experiments have demonstrated that TNNM can achieve satisfactory prediction accuracy in different databases, and simultaneously, TNN plays an irreplaceable positive role in TNNM as well.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding authors.

Author contributions

SC and CH produced the main ideas, and did the modeling, computation and analysis and also wrote the manuscript. LW and SZ provided supervision and effective scientific advice and related ideas, research design guidance, and added value to the article through editing and contributing completions. All authors contributed to the article and approved the submitted version.

Funding

This work was partly sponsored by the National Natural Science Foundation of China (No.62272064, No.61873221) and the Key project of Changsha Science and technology Plan (No. KQ2203001).

Acknowledgments

The authors thank the anonymous referees for suggestions that helped improve the paper substantially.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fgene.2022.1087294/full#supplementary-material

References

Acencio, M. L., and Lemke, N. (2009). Towards the prediction of essential genes by integration of network topology, cellular localization and biological process information. BMC Bioinforma. 10 (1), 290–307. doi:10.1186/1471-2105-10-290

PubMed Abstract | CrossRef Full Text | Google Scholar

Binder, J. X., Pletscher-Frankild, S., Tsafou, K., Stolte, C., O'Donoghue, S. I., Schneider, R., et al. (2014). Compartments: Unification and visualization of protein subcellular localization evidence. Database 2014, bau012. doi:10.1093/database/bau012

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, B., and Wu, F. X. (2013). Identifying protein complexes based on multiple topological structures in PPI networks. IEEE Trans Nanobio- Sci. 12 (3), 165–172. doi:10.1109/TNB.2013.2264097

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, Z., Meng, Z., Liu, C., Wang, X., Kuang, L., Pei, T., et al. (2020). A novel model for predicting essential proteins based on heterogeneous protein-domain network. IEEE Access 8 (99), 8946–8958. doi:10.1109/access.2020.2964571

CrossRef Full Text | Google Scholar

Cherry, J. M., Adler, C., Ball, C., Chervitz, S. A., Dwight, S. S., Hester, E. T., et al. (1998). SGD: Saccharomyces genome database. Nucleic Acids Res. 26, 73–79. doi:10.1093/nar/26.1.73

PubMed Abstract | CrossRef Full Text | Google Scholar

Dai, W., Chang, Q., Peng, W., Zhong, J., and Li, Y. (2020). Network embedding the protein-pro tein interaction network for human essential genes identification. Genes 11 (2), 153. doi:10.3390/genes11020153

PubMed Abstract | CrossRef Full Text | Google Scholar

Gabriel, O., Thomas, S., Kristoffer, F., Kostler, T., Messina, D. N., Roopra, S., et al. (2010). InParanoid 7: New algorithms and tools for eukaryotic orthology analysis. Nucleic Acids Res. 38, D196–D203. doi:10.1093/nar/gkp931

PubMed Abstract | CrossRef Full Text | Google Scholar

Gavin, A. C., Aloy, P., Grandi, P., Krause, R., Boesche, M., Marzioch, M., et al. (2006). Proteome survey reveals modularity of the yeast cell machinery. Nature 440 (7084), 631–636. doi:10.1038/nature04532

PubMed Abstract | CrossRef Full Text | Google Scholar

Hahn, M. W., and Kern, A. D. (2005). Comparative genomics of centrality and essentiality in three eukaryotic protein-interaction networks. Mol. Biol. Evol. 22 (4), 803–806. doi:10.1093/molbev/msi072

PubMed Abstract | CrossRef Full Text | Google Scholar

Holman, A. G., Davis, P. J., Foster, J. M., Carlow, C. K. S., and Kumar, S. (2009). Computational prediction of essential genes in an unculturable endosymbiotic bacterium, Wolbachia of Brugia malayi. BMC Microbiol. 9 (1), 243. doi:10.1186/1471-2180-9-243

PubMed Abstract | CrossRef Full Text | Google Scholar

Jeong, H. M., Mason, S. P., Al, B., and Oltvai, Z. N. (2001). Lethality and centrality in protein networks. Nature 411 (6833), 41–42. doi:10.1038/35075138

PubMed Abstract | CrossRef Full Text | Google Scholar

Jiawei, L., Yi, Q., and Peter, C. (2015). Identification of essential proteins based on a new combination of local interaction density and protein complexes. Plos One 10 (6), e0131418. doi:10.1371/journal.pone.0131418

PubMed Abstract | CrossRef Full Text | Google Scholar

Jung, Y. (2017). Multiple predicting K -fold cross-validation for model selection. J. Nonparametric Statistics 30, 197–215. doi:10.1080/10485252.2017.1404598

CrossRef Full Text | Google Scholar

Keretsu, S., and Sarmah, R. (2016). Weighted edge based clustering to identify protein complexes in protein protein inter- action networks incorporating gene expression profile. Comput. Biol. Chem. 65, 69–79. doi:10.1016/j.compbiolchem.2016.10.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Lei, X., Yang, X., and Wu, F. (2018). Artificial fish swarm optimization based method to identify essential proteins. IEEE/ACM Trans. Comput. Biol. Bioinform. 17, 495–505. doi:10.1109/TCBB.2018.2865567

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, M., Lu, Y., Niu, Z., and Wu, F. X. (2017). United complex centrality for identification of essential proteins from PPI networks. IEEE/ACM Trans. Comput. Biol. Bioinform. 14 (2), 370–380. doi:10.1109/TCBB.2015.2394487

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, M., Lu, Y., Wang, J., Wu, F. X., and Pan, Y. (2015). A topology potential-based method for identifying essential proteins from PPI networks. IEEE/ACM Trans. Comput. Biol. Bioinform. 12 (2), 372–383. doi:10.1109/TCBB.2014.2361350

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, M., Wang, J. X., Chen, X., Wang, H., and Pan, Y. (2011). A local average connectivity-based method for identifying essential proteins from the network level. Comput. Biol. Chem. 35 (3), 143–150. doi:10.1016/j.compbiolchem.2011.04.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, M., Zhang, H., Wang, J. X., and Pan, Y. (2012). A new essential protein discovery method based on the integration of protein-protein interaction and gene expression data. BMC Syst. Biol. 6 (1), 15. doi:10.1186/1752-0509-6-15

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, S., Chen, Z., He, X., Zhang, Z., Pei, T., Tan, Y., et al. (2020). An iteration method for identifying yeast essential proteins from weighted PPI network based on topological and functional features of proteins. IEEE Access 8, 90792–90804. doi:10.1109/access.2020.2993860

CrossRef Full Text | Google Scholar

Luo, J., and Wu, J. (2015). A new algorithm for essential proteins identification based on the integration of protein complex co-expression information and edge clustering coefficient. Int. J. Data Min. Bioinform 12 (3), 257–274. doi:10.1504/ijdmb.2015.069654

PubMed Abstract | CrossRef Full Text | Google Scholar

Maliackal, P. J., Amy, B., Donald, E. I., and Sui, H. (2005). High-Betweenness proteins in the yeast protein interaction network. J. Biomed. Biotechnol. 2005 (2), 96–103. doi:10.1155/JBB.2005.96

PubMed Abstract | CrossRef Full Text | Google Scholar

Maslov, S., and Sneppen, K., (2002). Specificity and stability in topology of protein networks. Science 296, 910–913. doi:10.1126/science.1065103

PubMed Abstract | CrossRef Full Text | Google Scholar

Mewes, H. W., Frishman, D., Mayer, K. F. X., Munsterkotter, M., Noubibou, O., Pagel, P., et al. (2006). MIPS: Analysis and annotation of proteins from whole genomes in 2005. Nucleic Acids Res. 34, D169–D172. doi:10.1093/nar/gkj148

PubMed Abstract | CrossRef Full Text | Google Scholar

Peng, W., Wang, J. X., Cheng, Y., Lu, Y., Wu, F., and Pan, Y. (2015). UDoNC: An algorithm for identifying essential proteins based on protein domains and protein-protein interaction networks. IEEE/ACM Trans Comput. Biol. Bioinforma. 12 (2), 276–288. doi:10.1109/TCBB.2014.2338317

PubMed Abstract | CrossRef Full Text | Google Scholar

Peng, W., Wang, J. X., Wang, W., Liu, Q., Wu, F. X., and Pan, Y. (2012). Iteration method for predicting essential proteins based on orthology and protein-protein interaction networks. BMC Syst. Biol. 6 (1), 87. doi:10.1186/1752-0509-6-87

PubMed Abstract | CrossRef Full Text | Google Scholar

Pržulj, N., Wigle, D. A., and Jurisica, I. (2004). Functional topology in a network of protein interactions. Bioinformatics 20 (3), 340–348. doi:10.1093/bioinformatics/btg415

PubMed Abstract | CrossRef Full Text | Google Scholar

Qi, Y., and Luo, J. (2016). Prediction of essential proteins based on local interaction density. IEEE/ACM Trans Computa- tional Biol. Bioinforma. 13 (6), 1170–1182. doi:10.1109/TCBB.2015.2509989

PubMed Abstract | CrossRef Full Text | Google Scholar

StanfordMedicine (2012). Saccharomyces genome deletion project. Available at: http://yeastdeletion.stanford.edu/(Accessed June 20, 2012).

Google Scholar

Stefan, W., and Stadler, P. F. (2003). Centers of complex networks. J. Theor. Biol. 223 (1), 45–53. doi:10.1016/s0022-5193(03)00071-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Stephenson, K., and Zelen, M. (1989). Rethinking centrality: Methods and examples. Soc. Netw. 11, 1–37. doi:10.1016/0378-8733(89)90016-6

CrossRef Full Text | Google Scholar

Tang, X., Wang, J., Zhong, J., and Pan, Y. (2014). Predicting essential proteins based on weighted degree centrality. IEEE/ACM Trans Comput. Biol. Bioinforma. 11 (2), 407–418. doi:10.1109/TCBB.2013.2295318

PubMed Abstract | CrossRef Full Text | Google Scholar

Tu, B. P., Kudlicki, A., Rowicka, M., and McKnight, S. L. (2005). Logic of the yeast metabolic cycle: Temporal compartmentalization of cellular processes. Science 310 (5751), 1152–1158. doi:10.1126/science.1120499

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, J. X., Li, M., Wang, H., and Pan, Y. (2012). Identification of essential proteins based on edge clustering coefficient. IEEE/ACM Trans. Comput. Biol. Bioinform. 9 (4), 1070–1080. doi:10.1109/TCBB.2011.147

PubMed Abstract | CrossRef Full Text | Google Scholar

Xenarios, I., Salwinski, L., Duan, X. J., Higney, P., Kim, S. M., and Eisenberg, D. (2002). DIP, the database of interacting proteins: A research tool for studying cellular networks of protein interactions. Nucleic Acids Res. 30 (1), 303–305. doi:10.1093/nar/30.1.303

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, F., Peng, W., Yang, Y., Dai, W., and Song, J. (2019). A novel method for identifying essential genes by fusing dynamic protein–protein interactive networks. Genes 10 (1), 31. doi:10.3390/genes10010031

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, R., and Lin, Y. (2009). DEG 5.0, a database of essential genes in both prokaryotes and eukaryotes. Nucleic Acids Res. 37, D455–D458. doi:10.1093/nar/gkn858

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, X., Xu, J., and Xiao, W. (2013). A new method for the discovery of essential proteins. PLoS One 8 (3), e58763. doi:10.1371/journal.pone.0058763

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, B., Han, X., Liu, X., Luo, Y., Hu, S., Zhang, Z., et al. (2020). A novel method to predict essential proteins based on diffusion distance networks. IEEE ACCESS 8, 29385–29394. doi:10.1109/access.2020.2972922

CrossRef Full Text | Google Scholar

Zhao, B., Hu, S., Liu, X., Xiong, H., Han, X., Zhang, Z., et al. (2020). A novel computational approach for identifying essential proteins from multiplex biological networks. Front. Genet. 11, 343. doi:10.3389/fgene.2020.00343

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, B., Zhao, Y., Zhang, X., Zhang, Z., Zhang, F., and Wang, L. (2019). An iteration method for identifying yeast essential proteins from heterogeneous network. BMC Bioinforma. 20, 355. doi:10.1186/s12859-019-2930-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, B. H., Wang, J. X., Li, M., Wu, F. X., and Pan, Y. (2014). Prediction of essential proteins based on overlapping essential modules. IEEE Trans. Nanobioscience 13 (4), 415–424. doi:10.1109/TNB.2014.2337912

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: essential protein, prediction model, transfer neural network, biological information, internal disease, protein-protein interaction network

Citation: Chen S, Huang C, Wang L and Zhou S (2023) A disease-related essential protein prediction model based on the transfer neural network. Front. Genet. 13:1087294. doi: 10.3389/fgene.2022.1087294

Received: 02 November 2022; Accepted: 14 December 2022;
Published: 04 January 2023.

Edited by:

Marco S Nobile, Ca’ Foscari University of Venice, Italy

Reviewed by:

Yuhua Yao, Hainan Normal University, China
Chaoyang Pang, Sichuan Normal University, China

Copyright © 2023 Chen, Huang, Wang and Zhou. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Chiguo Huang, MjM5NDE4MTU1N0BxcS5jb20=; Lei Wang, d2FuZ2xlaUB4dHUuZWR1LmNu; Shunxian Zhou, enhzX2hkQGhudS5lZHUuY24=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.